Jan 16 08:58:44.077873 kernel: Linux version 6.6.71-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Jan 13 19:40:50 -00 2025 Jan 16 08:58:44.077929 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=8945029ddd0f3864592f8746dde99cfcba228e0d3cb946f5938103dbe8733507 Jan 16 08:58:44.077949 kernel: BIOS-provided physical RAM map: Jan 16 08:58:44.077959 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Jan 16 08:58:44.077969 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Jan 16 08:58:44.077978 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jan 16 08:58:44.077990 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Jan 16 08:58:44.078001 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Jan 16 08:58:44.078011 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 16 08:58:44.078026 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jan 16 08:58:44.078037 kernel: NX (Execute Disable) protection: active Jan 16 08:58:44.078047 kernel: APIC: Static calls initialized Jan 16 08:58:44.078071 kernel: SMBIOS 2.8 present. Jan 16 08:58:44.078082 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Jan 16 08:58:44.078094 kernel: Hypervisor detected: KVM Jan 16 08:58:44.078111 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 16 08:58:44.078127 kernel: kvm-clock: using sched offset of 3672786839 cycles Jan 16 08:58:44.078139 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 16 08:58:44.078150 kernel: tsc: Detected 2494.140 MHz processor Jan 16 08:58:44.078161 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 16 08:58:44.078174 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 16 08:58:44.078186 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Jan 16 08:58:44.078197 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jan 16 08:58:44.078208 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 16 08:58:44.078225 kernel: ACPI: Early table checksum verification disabled Jan 16 08:58:44.078237 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Jan 16 08:58:44.078251 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 16 08:58:44.078262 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 16 08:58:44.078273 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 16 08:58:44.078284 kernel: ACPI: FACS 0x000000007FFE0000 000040 Jan 16 08:58:44.078295 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 16 08:58:44.078306 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 16 08:58:44.078318 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 16 08:58:44.078334 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 16 08:58:44.078346 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Jan 16 08:58:44.078359 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Jan 16 08:58:44.078370 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Jan 16 08:58:44.078381 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Jan 16 08:58:44.078393 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Jan 16 08:58:44.078406 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Jan 16 08:58:44.078429 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Jan 16 08:58:44.078443 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Jan 16 08:58:44.078457 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Jan 16 08:58:44.078471 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Jan 16 08:58:44.078486 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Jan 16 08:58:44.078504 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Jan 16 08:58:44.078519 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Jan 16 08:58:44.078539 kernel: Zone ranges: Jan 16 08:58:44.078554 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 16 08:58:44.078567 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Jan 16 08:58:44.078579 kernel: Normal empty Jan 16 08:58:44.078592 kernel: Movable zone start for each node Jan 16 08:58:44.078604 kernel: Early memory node ranges Jan 16 08:58:44.078616 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jan 16 08:58:44.078629 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Jan 16 08:58:44.078643 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Jan 16 08:58:44.079051 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 16 08:58:44.079065 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jan 16 08:58:44.079133 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Jan 16 08:58:44.079147 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 16 08:58:44.079160 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 16 08:58:44.079175 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 16 08:58:44.079187 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 16 08:58:44.079200 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 16 08:58:44.079213 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 16 08:58:44.079233 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 16 08:58:44.079246 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 16 08:58:44.079257 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 16 08:58:44.079269 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 16 08:58:44.079299 kernel: TSC deadline timer available Jan 16 08:58:44.079313 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Jan 16 08:58:44.079325 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 16 08:58:44.079337 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Jan 16 08:58:44.079356 kernel: Booting paravirtualized kernel on KVM Jan 16 08:58:44.079370 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 16 08:58:44.079389 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Jan 16 08:58:44.079402 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Jan 16 08:58:44.079415 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Jan 16 08:58:44.079428 kernel: pcpu-alloc: [0] 0 1 Jan 16 08:58:44.079440 kernel: kvm-guest: PV spinlocks disabled, no host support Jan 16 08:58:44.079456 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=8945029ddd0f3864592f8746dde99cfcba228e0d3cb946f5938103dbe8733507 Jan 16 08:58:44.079471 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 16 08:58:44.079484 kernel: random: crng init done Jan 16 08:58:44.079501 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 16 08:58:44.079515 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jan 16 08:58:44.079527 kernel: Fallback order for Node 0: 0 Jan 16 08:58:44.079540 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Jan 16 08:58:44.079553 kernel: Policy zone: DMA32 Jan 16 08:58:44.079567 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 16 08:58:44.079580 kernel: Memory: 1971204K/2096612K available (12288K kernel code, 2299K rwdata, 22728K rodata, 42844K init, 2348K bss, 125148K reserved, 0K cma-reserved) Jan 16 08:58:44.079593 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jan 16 08:58:44.079614 kernel: Kernel/User page tables isolation: enabled Jan 16 08:58:44.079629 kernel: ftrace: allocating 37918 entries in 149 pages Jan 16 08:58:44.079642 kernel: ftrace: allocated 149 pages with 4 groups Jan 16 08:58:44.081747 kernel: Dynamic Preempt: voluntary Jan 16 08:58:44.081765 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 16 08:58:44.081789 kernel: rcu: RCU event tracing is enabled. Jan 16 08:58:44.081803 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jan 16 08:58:44.081816 kernel: Trampoline variant of Tasks RCU enabled. Jan 16 08:58:44.081829 kernel: Rude variant of Tasks RCU enabled. Jan 16 08:58:44.081843 kernel: Tracing variant of Tasks RCU enabled. Jan 16 08:58:44.081865 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 16 08:58:44.081878 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jan 16 08:58:44.081892 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Jan 16 08:58:44.081905 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 16 08:58:44.081928 kernel: Console: colour VGA+ 80x25 Jan 16 08:58:44.081941 kernel: printk: console [tty0] enabled Jan 16 08:58:44.081953 kernel: printk: console [ttyS0] enabled Jan 16 08:58:44.081966 kernel: ACPI: Core revision 20230628 Jan 16 08:58:44.081978 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 16 08:58:44.081999 kernel: APIC: Switch to symmetric I/O mode setup Jan 16 08:58:44.082013 kernel: x2apic enabled Jan 16 08:58:44.082026 kernel: APIC: Switched APIC routing to: physical x2apic Jan 16 08:58:44.082039 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 16 08:58:44.082053 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Jan 16 08:58:44.082066 kernel: Calibrating delay loop (skipped) preset value.. 4988.28 BogoMIPS (lpj=2494140) Jan 16 08:58:44.082078 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Jan 16 08:58:44.082092 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Jan 16 08:58:44.082124 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 16 08:58:44.082138 kernel: Spectre V2 : Mitigation: Retpolines Jan 16 08:58:44.082155 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Jan 16 08:58:44.082173 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Jan 16 08:58:44.082187 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Jan 16 08:58:44.082200 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jan 16 08:58:44.082213 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jan 16 08:58:44.082226 kernel: MDS: Mitigation: Clear CPU buffers Jan 16 08:58:44.082239 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jan 16 08:58:44.082266 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 16 08:58:44.082280 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 16 08:58:44.082293 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 16 08:58:44.082306 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 16 08:58:44.082319 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Jan 16 08:58:44.082335 kernel: Freeing SMP alternatives memory: 32K Jan 16 08:58:44.082347 kernel: pid_max: default: 32768 minimum: 301 Jan 16 08:58:44.082360 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 16 08:58:44.082379 kernel: landlock: Up and running. Jan 16 08:58:44.082393 kernel: SELinux: Initializing. Jan 16 08:58:44.082406 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 16 08:58:44.082420 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 16 08:58:44.082434 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Jan 16 08:58:44.082449 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 16 08:58:44.082462 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 16 08:58:44.082476 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 16 08:58:44.082490 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Jan 16 08:58:44.082511 kernel: signal: max sigframe size: 1776 Jan 16 08:58:44.082527 kernel: rcu: Hierarchical SRCU implementation. Jan 16 08:58:44.082542 kernel: rcu: Max phase no-delay instances is 400. Jan 16 08:58:44.082558 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 16 08:58:44.082575 kernel: smp: Bringing up secondary CPUs ... Jan 16 08:58:44.082589 kernel: smpboot: x86: Booting SMP configuration: Jan 16 08:58:44.082604 kernel: .... node #0, CPUs: #1 Jan 16 08:58:44.082619 kernel: smp: Brought up 1 node, 2 CPUs Jan 16 08:58:44.082640 kernel: smpboot: Max logical packages: 1 Jan 16 08:58:44.082695 kernel: smpboot: Total of 2 processors activated (9976.56 BogoMIPS) Jan 16 08:58:44.082710 kernel: devtmpfs: initialized Jan 16 08:58:44.082725 kernel: x86/mm: Memory block size: 128MB Jan 16 08:58:44.082741 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 16 08:58:44.082757 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jan 16 08:58:44.082772 kernel: pinctrl core: initialized pinctrl subsystem Jan 16 08:58:44.082788 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 16 08:58:44.082803 kernel: audit: initializing netlink subsys (disabled) Jan 16 08:58:44.082820 kernel: audit: type=2000 audit(1737017922.250:1): state=initialized audit_enabled=0 res=1 Jan 16 08:58:44.082841 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 16 08:58:44.082855 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 16 08:58:44.082870 kernel: cpuidle: using governor menu Jan 16 08:58:44.082883 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 16 08:58:44.082897 kernel: dca service started, version 1.12.1 Jan 16 08:58:44.082910 kernel: PCI: Using configuration type 1 for base access Jan 16 08:58:44.082924 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 16 08:58:44.082938 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 16 08:58:44.082954 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 16 08:58:44.082974 kernel: ACPI: Added _OSI(Module Device) Jan 16 08:58:44.082990 kernel: ACPI: Added _OSI(Processor Device) Jan 16 08:58:44.083006 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 16 08:58:44.083022 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 16 08:58:44.083035 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 16 08:58:44.083049 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jan 16 08:58:44.083062 kernel: ACPI: Interpreter enabled Jan 16 08:58:44.083076 kernel: ACPI: PM: (supports S0 S5) Jan 16 08:58:44.083088 kernel: ACPI: Using IOAPIC for interrupt routing Jan 16 08:58:44.083107 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 16 08:58:44.083120 kernel: PCI: Using E820 reservations for host bridge windows Jan 16 08:58:44.083134 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Jan 16 08:58:44.083148 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 16 08:58:44.083526 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Jan 16 08:58:44.084954 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Jan 16 08:58:44.085159 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Jan 16 08:58:44.085191 kernel: acpiphp: Slot [3] registered Jan 16 08:58:44.085206 kernel: acpiphp: Slot [4] registered Jan 16 08:58:44.085221 kernel: acpiphp: Slot [5] registered Jan 16 08:58:44.085236 kernel: acpiphp: Slot [6] registered Jan 16 08:58:44.085251 kernel: acpiphp: Slot [7] registered Jan 16 08:58:44.085266 kernel: acpiphp: Slot [8] registered Jan 16 08:58:44.085281 kernel: acpiphp: Slot [9] registered Jan 16 08:58:44.085296 kernel: acpiphp: Slot [10] registered Jan 16 08:58:44.085311 kernel: acpiphp: Slot [11] registered Jan 16 08:58:44.085328 kernel: acpiphp: Slot [12] registered Jan 16 08:58:44.085349 kernel: acpiphp: Slot [13] registered Jan 16 08:58:44.085363 kernel: acpiphp: Slot [14] registered Jan 16 08:58:44.085378 kernel: acpiphp: Slot [15] registered Jan 16 08:58:44.085394 kernel: acpiphp: Slot [16] registered Jan 16 08:58:44.085409 kernel: acpiphp: Slot [17] registered Jan 16 08:58:44.085423 kernel: acpiphp: Slot [18] registered Jan 16 08:58:44.085436 kernel: acpiphp: Slot [19] registered Jan 16 08:58:44.085450 kernel: acpiphp: Slot [20] registered Jan 16 08:58:44.085464 kernel: acpiphp: Slot [21] registered Jan 16 08:58:44.085482 kernel: acpiphp: Slot [22] registered Jan 16 08:58:44.085497 kernel: acpiphp: Slot [23] registered Jan 16 08:58:44.085514 kernel: acpiphp: Slot [24] registered Jan 16 08:58:44.085529 kernel: acpiphp: Slot [25] registered Jan 16 08:58:44.085544 kernel: acpiphp: Slot [26] registered Jan 16 08:58:44.085558 kernel: acpiphp: Slot [27] registered Jan 16 08:58:44.085574 kernel: acpiphp: Slot [28] registered Jan 16 08:58:44.085590 kernel: acpiphp: Slot [29] registered Jan 16 08:58:44.085605 kernel: acpiphp: Slot [30] registered Jan 16 08:58:44.085620 kernel: acpiphp: Slot [31] registered Jan 16 08:58:44.085638 kernel: PCI host bridge to bus 0000:00 Jan 16 08:58:44.085883 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 16 08:58:44.086040 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 16 08:58:44.086175 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 16 08:58:44.086310 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Jan 16 08:58:44.086441 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Jan 16 08:58:44.086575 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 16 08:58:44.089013 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Jan 16 08:58:44.089252 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Jan 16 08:58:44.089438 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Jan 16 08:58:44.089593 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Jan 16 08:58:44.090949 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Jan 16 08:58:44.091161 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Jan 16 08:58:44.091337 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Jan 16 08:58:44.091499 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Jan 16 08:58:44.092817 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Jan 16 08:58:44.093044 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Jan 16 08:58:44.093240 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Jan 16 08:58:44.093415 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Jan 16 08:58:44.093589 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Jan 16 08:58:44.094907 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Jan 16 08:58:44.095106 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Jan 16 08:58:44.095274 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Jan 16 08:58:44.095443 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Jan 16 08:58:44.095612 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Jan 16 08:58:44.097939 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 16 08:58:44.098308 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Jan 16 08:58:44.098500 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Jan 16 08:58:44.098682 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Jan 16 08:58:44.098856 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Jan 16 08:58:44.099060 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Jan 16 08:58:44.099233 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Jan 16 08:58:44.099487 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Jan 16 08:58:44.100819 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Jan 16 08:58:44.101082 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Jan 16 08:58:44.101258 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Jan 16 08:58:44.101410 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Jan 16 08:58:44.101560 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Jan 16 08:58:44.102998 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Jan 16 08:58:44.103226 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Jan 16 08:58:44.103404 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Jan 16 08:58:44.103561 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Jan 16 08:58:44.105135 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Jan 16 08:58:44.105347 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Jan 16 08:58:44.105525 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Jan 16 08:58:44.107861 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Jan 16 08:58:44.108185 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Jan 16 08:58:44.108447 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Jan 16 08:58:44.108620 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Jan 16 08:58:44.108691 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 16 08:58:44.108709 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 16 08:58:44.108724 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 16 08:58:44.108740 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 16 08:58:44.108756 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Jan 16 08:58:44.108809 kernel: iommu: Default domain type: Translated Jan 16 08:58:44.108825 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 16 08:58:44.108842 kernel: PCI: Using ACPI for IRQ routing Jan 16 08:58:44.108858 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 16 08:58:44.108875 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Jan 16 08:58:44.108890 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Jan 16 08:58:44.109080 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Jan 16 08:58:44.109235 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Jan 16 08:58:44.109394 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 16 08:58:44.109414 kernel: vgaarb: loaded Jan 16 08:58:44.109431 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 16 08:58:44.109447 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 16 08:58:44.109461 kernel: clocksource: Switched to clocksource kvm-clock Jan 16 08:58:44.109477 kernel: VFS: Disk quotas dquot_6.6.0 Jan 16 08:58:44.109492 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 16 08:58:44.109506 kernel: pnp: PnP ACPI init Jan 16 08:58:44.109521 kernel: pnp: PnP ACPI: found 4 devices Jan 16 08:58:44.109541 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 16 08:58:44.109557 kernel: NET: Registered PF_INET protocol family Jan 16 08:58:44.109573 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 16 08:58:44.109586 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Jan 16 08:58:44.109603 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 16 08:58:44.109616 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Jan 16 08:58:44.109629 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Jan 16 08:58:44.109643 kernel: TCP: Hash tables configured (established 16384 bind 16384) Jan 16 08:58:44.109693 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 16 08:58:44.109712 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 16 08:58:44.109724 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 16 08:58:44.109737 kernel: NET: Registered PF_XDP protocol family Jan 16 08:58:44.109907 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 16 08:58:44.110055 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 16 08:58:44.110198 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 16 08:58:44.110333 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Jan 16 08:58:44.110540 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Jan 16 08:58:44.112904 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Jan 16 08:58:44.113140 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jan 16 08:58:44.113171 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Jan 16 08:58:44.113335 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x7a0 took 45808 usecs Jan 16 08:58:44.113356 kernel: PCI: CLS 0 bytes, default 64 Jan 16 08:58:44.113371 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jan 16 08:58:44.113386 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Jan 16 08:58:44.113400 kernel: Initialise system trusted keyrings Jan 16 08:58:44.113415 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Jan 16 08:58:44.113442 kernel: Key type asymmetric registered Jan 16 08:58:44.113456 kernel: Asymmetric key parser 'x509' registered Jan 16 08:58:44.113469 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jan 16 08:58:44.113483 kernel: io scheduler mq-deadline registered Jan 16 08:58:44.113497 kernel: io scheduler kyber registered Jan 16 08:58:44.113511 kernel: io scheduler bfq registered Jan 16 08:58:44.113525 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 16 08:58:44.113540 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Jan 16 08:58:44.113553 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Jan 16 08:58:44.113572 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Jan 16 08:58:44.113586 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 16 08:58:44.113599 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 16 08:58:44.113613 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 16 08:58:44.113627 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 16 08:58:44.113640 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 16 08:58:44.113890 kernel: rtc_cmos 00:03: RTC can wake from S4 Jan 16 08:58:44.113918 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 16 08:58:44.114081 kernel: rtc_cmos 00:03: registered as rtc0 Jan 16 08:58:44.114246 kernel: rtc_cmos 00:03: setting system clock to 2025-01-16T08:58:43 UTC (1737017923) Jan 16 08:58:44.114386 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Jan 16 08:58:44.114406 kernel: intel_pstate: CPU model not supported Jan 16 08:58:44.114421 kernel: NET: Registered PF_INET6 protocol family Jan 16 08:58:44.114436 kernel: Segment Routing with IPv6 Jan 16 08:58:44.114451 kernel: In-situ OAM (IOAM) with IPv6 Jan 16 08:58:44.114465 kernel: NET: Registered PF_PACKET protocol family Jan 16 08:58:44.114478 kernel: Key type dns_resolver registered Jan 16 08:58:44.114501 kernel: IPI shorthand broadcast: enabled Jan 16 08:58:44.114515 kernel: sched_clock: Marking stable (1345008605, 100492215)->(1488467066, -42966246) Jan 16 08:58:44.114528 kernel: registered taskstats version 1 Jan 16 08:58:44.114542 kernel: Loading compiled-in X.509 certificates Jan 16 08:58:44.114555 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.71-flatcar: e8ca4908f7ff887d90a0430272c92dde55624447' Jan 16 08:58:44.114569 kernel: Key type .fscrypt registered Jan 16 08:58:44.114582 kernel: Key type fscrypt-provisioning registered Jan 16 08:58:44.114596 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 16 08:58:44.114615 kernel: ima: Allocated hash algorithm: sha1 Jan 16 08:58:44.114629 kernel: ima: No architecture policies found Jan 16 08:58:44.116705 kernel: clk: Disabling unused clocks Jan 16 08:58:44.116761 kernel: Freeing unused kernel image (initmem) memory: 42844K Jan 16 08:58:44.116780 kernel: Write protecting the kernel read-only data: 36864k Jan 16 08:58:44.116830 kernel: Freeing unused kernel image (rodata/data gap) memory: 1848K Jan 16 08:58:44.116848 kernel: Run /init as init process Jan 16 08:58:44.116862 kernel: with arguments: Jan 16 08:58:44.116877 kernel: /init Jan 16 08:58:44.116894 kernel: with environment: Jan 16 08:58:44.116908 kernel: HOME=/ Jan 16 08:58:44.116922 kernel: TERM=linux Jan 16 08:58:44.116936 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 16 08:58:44.116958 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 16 08:58:44.116977 systemd[1]: Detected virtualization kvm. Jan 16 08:58:44.116994 systemd[1]: Detected architecture x86-64. Jan 16 08:58:44.117008 systemd[1]: Running in initrd. Jan 16 08:58:44.117028 systemd[1]: No hostname configured, using default hostname. Jan 16 08:58:44.117043 systemd[1]: Hostname set to . Jan 16 08:58:44.117062 systemd[1]: Initializing machine ID from VM UUID. Jan 16 08:58:44.117076 systemd[1]: Queued start job for default target initrd.target. Jan 16 08:58:44.117093 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 16 08:58:44.117109 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 16 08:58:44.117126 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 16 08:58:44.117142 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 16 08:58:44.117162 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 16 08:58:44.117178 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 16 08:58:44.117195 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 16 08:58:44.117211 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 16 08:58:44.117226 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 16 08:58:44.117242 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 16 08:58:44.117257 systemd[1]: Reached target paths.target - Path Units. Jan 16 08:58:44.117277 systemd[1]: Reached target slices.target - Slice Units. Jan 16 08:58:44.117292 systemd[1]: Reached target swap.target - Swaps. Jan 16 08:58:44.117311 systemd[1]: Reached target timers.target - Timer Units. Jan 16 08:58:44.117326 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 16 08:58:44.117343 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 16 08:58:44.117363 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 16 08:58:44.117378 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 16 08:58:44.117395 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 16 08:58:44.117412 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 16 08:58:44.117428 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 16 08:58:44.117443 systemd[1]: Reached target sockets.target - Socket Units. Jan 16 08:58:44.117465 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 16 08:58:44.117479 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 16 08:58:44.117495 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 16 08:58:44.117517 systemd[1]: Starting systemd-fsck-usr.service... Jan 16 08:58:44.117532 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 16 08:58:44.117546 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 16 08:58:44.117560 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 16 08:58:44.117575 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 16 08:58:44.117590 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 16 08:58:44.117700 systemd-journald[183]: Collecting audit messages is disabled. Jan 16 08:58:44.117747 systemd[1]: Finished systemd-fsck-usr.service. Jan 16 08:58:44.117765 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 16 08:58:44.117786 systemd-journald[183]: Journal started Jan 16 08:58:44.117816 systemd-journald[183]: Runtime Journal (/run/log/journal/b32100f301ca40d9bb78873a577f8845) is 4.9M, max 39.3M, 34.4M free. Jan 16 08:58:44.074218 systemd-modules-load[184]: Inserted module 'overlay' Jan 16 08:58:44.156839 systemd[1]: Started systemd-journald.service - Journal Service. Jan 16 08:58:44.156881 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 16 08:58:44.156920 kernel: Bridge firewalling registered Jan 16 08:58:44.125891 systemd-modules-load[184]: Inserted module 'br_netfilter' Jan 16 08:58:44.156601 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 16 08:58:44.157633 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 16 08:58:44.158674 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 16 08:58:44.172086 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 16 08:58:44.179733 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 16 08:58:44.183529 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 16 08:58:44.189915 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 16 08:58:44.204368 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 16 08:58:44.216558 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 16 08:58:44.220893 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 16 08:58:44.221743 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 16 08:58:44.228158 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 16 08:58:44.236082 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 16 08:58:44.260585 dracut-cmdline[216]: dracut-dracut-053 Jan 16 08:58:44.268710 dracut-cmdline[216]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=8945029ddd0f3864592f8746dde99cfcba228e0d3cb946f5938103dbe8733507 Jan 16 08:58:44.288320 systemd-resolved[218]: Positive Trust Anchors: Jan 16 08:58:44.288341 systemd-resolved[218]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 16 08:58:44.288392 systemd-resolved[218]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 16 08:58:44.293886 systemd-resolved[218]: Defaulting to hostname 'linux'. Jan 16 08:58:44.295797 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 16 08:58:44.296569 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 16 08:58:44.403736 kernel: SCSI subsystem initialized Jan 16 08:58:44.415710 kernel: Loading iSCSI transport class v2.0-870. Jan 16 08:58:44.429703 kernel: iscsi: registered transport (tcp) Jan 16 08:58:44.457781 kernel: iscsi: registered transport (qla4xxx) Jan 16 08:58:44.457914 kernel: QLogic iSCSI HBA Driver Jan 16 08:58:44.529490 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 16 08:58:44.537990 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 16 08:58:44.580881 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 16 08:58:44.581012 kernel: device-mapper: uevent: version 1.0.3 Jan 16 08:58:44.582476 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 16 08:58:44.642764 kernel: raid6: avx2x4 gen() 15357 MB/s Jan 16 08:58:44.659732 kernel: raid6: avx2x2 gen() 14434 MB/s Jan 16 08:58:44.677490 kernel: raid6: avx2x1 gen() 11453 MB/s Jan 16 08:58:44.677600 kernel: raid6: using algorithm avx2x4 gen() 15357 MB/s Jan 16 08:58:44.696704 kernel: raid6: .... xor() 4847 MB/s, rmw enabled Jan 16 08:58:44.696825 kernel: raid6: using avx2x2 recovery algorithm Jan 16 08:58:44.726716 kernel: xor: automatically using best checksumming function avx Jan 16 08:58:44.956710 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 16 08:58:44.975419 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 16 08:58:44.983106 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 16 08:58:45.021639 systemd-udevd[402]: Using default interface naming scheme 'v255'. Jan 16 08:58:45.030759 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 16 08:58:45.039899 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 16 08:58:45.075751 dracut-pre-trigger[407]: rd.md=0: removing MD RAID activation Jan 16 08:58:45.128176 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 16 08:58:45.144181 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 16 08:58:45.233400 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 16 08:58:45.242046 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 16 08:58:45.268264 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 16 08:58:45.269626 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 16 08:58:45.270642 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 16 08:58:45.271134 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 16 08:58:45.280708 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 16 08:58:45.322877 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 16 08:58:45.345687 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Jan 16 08:58:45.397117 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Jan 16 08:58:45.397358 kernel: scsi host0: Virtio SCSI HBA Jan 16 08:58:45.397592 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 16 08:58:45.397616 kernel: GPT:9289727 != 125829119 Jan 16 08:58:45.397633 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 16 08:58:45.397679 kernel: GPT:9289727 != 125829119 Jan 16 08:58:45.397697 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 16 08:58:45.397725 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 16 08:58:45.401687 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Jan 16 08:58:45.430998 kernel: cryptd: max_cpu_qlen set to 1000 Jan 16 08:58:45.431030 kernel: virtio_blk virtio5: [vdb] 952 512-byte logical blocks (487 kB/476 KiB) Jan 16 08:58:45.445088 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 16 08:58:45.446714 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 16 08:58:45.450516 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 16 08:58:45.451553 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 16 08:58:45.452367 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 16 08:58:45.454603 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 16 08:58:45.470261 kernel: ACPI: bus type USB registered Jan 16 08:58:45.471582 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 16 08:58:45.475872 kernel: AVX2 version of gcm_enc/dec engaged. Jan 16 08:58:45.475914 kernel: AES CTR mode by8 optimization enabled Jan 16 08:58:45.500473 kernel: libata version 3.00 loaded. Jan 16 08:58:45.500571 kernel: ata_piix 0000:00:01.1: version 2.13 Jan 16 08:58:45.543167 kernel: usbcore: registered new interface driver usbfs Jan 16 08:58:45.543196 kernel: scsi host1: ata_piix Jan 16 08:58:45.543431 kernel: usbcore: registered new interface driver hub Jan 16 08:58:45.543452 kernel: usbcore: registered new device driver usb Jan 16 08:58:45.543470 kernel: scsi host2: ata_piix Jan 16 08:58:45.543708 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Jan 16 08:58:45.543730 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Jan 16 08:58:45.549716 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (456) Jan 16 08:58:45.577445 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 16 08:58:45.619487 kernel: BTRFS: device fsid b8e2d3c5-4bed-4339-bed5-268c66823686 devid 1 transid 37 /dev/vda3 scanned by (udev-worker) (455) Jan 16 08:58:45.626578 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 16 08:58:45.631293 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 16 08:58:45.649884 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 16 08:58:45.656289 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Jan 16 08:58:45.656915 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 16 08:58:45.666985 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 16 08:58:45.668984 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 16 08:58:45.681752 disk-uuid[531]: Primary Header is updated. Jan 16 08:58:45.681752 disk-uuid[531]: Secondary Entries is updated. Jan 16 08:58:45.681752 disk-uuid[531]: Secondary Header is updated. Jan 16 08:58:45.700721 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 16 08:58:45.723093 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 16 08:58:45.732693 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 16 08:58:45.741691 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 16 08:58:45.744793 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Jan 16 08:58:45.767522 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Jan 16 08:58:45.767754 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Jan 16 08:58:45.767991 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Jan 16 08:58:45.768195 kernel: hub 1-0:1.0: USB hub found Jan 16 08:58:45.768380 kernel: hub 1-0:1.0: 2 ports detected Jan 16 08:58:46.739697 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 16 08:58:46.740844 disk-uuid[532]: The operation has completed successfully. Jan 16 08:58:46.812846 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 16 08:58:46.813032 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 16 08:58:46.821063 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 16 08:58:46.833517 sh[562]: Success Jan 16 08:58:46.850142 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Jan 16 08:58:46.920231 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 16 08:58:46.923895 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 16 08:58:46.927796 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 16 08:58:46.965541 kernel: BTRFS info (device dm-0): first mount of filesystem b8e2d3c5-4bed-4339-bed5-268c66823686 Jan 16 08:58:46.965622 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 16 08:58:46.965636 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 16 08:58:46.965677 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 16 08:58:46.965696 kernel: BTRFS info (device dm-0): using free space tree Jan 16 08:58:46.979188 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 16 08:58:46.980701 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 16 08:58:46.986933 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 16 08:58:46.989878 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 16 08:58:47.005688 kernel: BTRFS info (device vda6): first mount of filesystem 70d8a0b5-70da-4efb-a618-d15543718b1e Jan 16 08:58:47.005778 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 16 08:58:47.005803 kernel: BTRFS info (device vda6): using free space tree Jan 16 08:58:47.012128 kernel: BTRFS info (device vda6): auto enabling async discard Jan 16 08:58:47.027299 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 16 08:58:47.030749 kernel: BTRFS info (device vda6): last unmount of filesystem 70d8a0b5-70da-4efb-a618-d15543718b1e Jan 16 08:58:47.044883 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 16 08:58:47.050010 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 16 08:58:47.224394 ignition[650]: Ignition 2.19.0 Jan 16 08:58:47.224409 ignition[650]: Stage: fetch-offline Jan 16 08:58:47.224463 ignition[650]: no configs at "/usr/lib/ignition/base.d" Jan 16 08:58:47.224475 ignition[650]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 16 08:58:47.224616 ignition[650]: parsed url from cmdline: "" Jan 16 08:58:47.224621 ignition[650]: no config URL provided Jan 16 08:58:47.224628 ignition[650]: reading system config file "/usr/lib/ignition/user.ign" Jan 16 08:58:47.229460 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 16 08:58:47.224637 ignition[650]: no config at "/usr/lib/ignition/user.ign" Jan 16 08:58:47.224662 ignition[650]: failed to fetch config: resource requires networking Jan 16 08:58:47.225502 ignition[650]: Ignition finished successfully Jan 16 08:58:47.244288 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 16 08:58:47.254110 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 16 08:58:47.284455 systemd-networkd[754]: lo: Link UP Jan 16 08:58:47.284472 systemd-networkd[754]: lo: Gained carrier Jan 16 08:58:47.287834 systemd-networkd[754]: Enumeration completed Jan 16 08:58:47.288789 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 16 08:58:47.289090 systemd-networkd[754]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Jan 16 08:58:47.289097 systemd-networkd[754]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Jan 16 08:58:47.289416 systemd[1]: Reached target network.target - Network. Jan 16 08:58:47.291615 systemd-networkd[754]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 16 08:58:47.291621 systemd-networkd[754]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 16 08:58:47.292779 systemd-networkd[754]: eth0: Link UP Jan 16 08:58:47.292784 systemd-networkd[754]: eth0: Gained carrier Jan 16 08:58:47.292799 systemd-networkd[754]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Jan 16 08:58:47.298010 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jan 16 08:58:47.298905 systemd-networkd[754]: eth1: Link UP Jan 16 08:58:47.298911 systemd-networkd[754]: eth1: Gained carrier Jan 16 08:58:47.298936 systemd-networkd[754]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 16 08:58:47.330449 ignition[756]: Ignition 2.19.0 Jan 16 08:58:47.330478 ignition[756]: Stage: fetch Jan 16 08:58:47.330745 ignition[756]: no configs at "/usr/lib/ignition/base.d" Jan 16 08:58:47.331751 systemd-networkd[754]: eth0: DHCPv4 address 147.182.199.56/20, gateway 147.182.192.1 acquired from 169.254.169.253 Jan 16 08:58:47.330761 ignition[756]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 16 08:58:47.330889 ignition[756]: parsed url from cmdline: "" Jan 16 08:58:47.330894 ignition[756]: no config URL provided Jan 16 08:58:47.330903 ignition[756]: reading system config file "/usr/lib/ignition/user.ign" Jan 16 08:58:47.330916 ignition[756]: no config at "/usr/lib/ignition/user.ign" Jan 16 08:58:47.330946 ignition[756]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Jan 16 08:58:47.331180 ignition[756]: GET error: Get "http://169.254.169.254/metadata/v1/user-data": dial tcp 169.254.169.254:80: connect: network is unreachable Jan 16 08:58:47.336087 systemd-networkd[754]: eth1: DHCPv4 address 10.124.0.4/20 acquired from 169.254.169.253 Jan 16 08:58:47.531425 ignition[756]: GET http://169.254.169.254/metadata/v1/user-data: attempt #2 Jan 16 08:58:47.575005 ignition[756]: GET result: OK Jan 16 08:58:47.576289 ignition[756]: parsing config with SHA512: f3c54693941381d712e52d2c94e4e3212ec895edde274fa69044e413a79938bc26ebc6ed3f848f64fa7a6c24b011f87c59f8897751d35d01143e54ce0aa3ff5a Jan 16 08:58:47.582690 unknown[756]: fetched base config from "system" Jan 16 08:58:47.582701 unknown[756]: fetched base config from "system" Jan 16 08:58:47.583264 ignition[756]: fetch: fetch complete Jan 16 08:58:47.582708 unknown[756]: fetched user config from "digitalocean" Jan 16 08:58:47.583276 ignition[756]: fetch: fetch passed Jan 16 08:58:47.586791 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jan 16 08:58:47.583397 ignition[756]: Ignition finished successfully Jan 16 08:58:47.600077 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 16 08:58:47.632846 ignition[763]: Ignition 2.19.0 Jan 16 08:58:47.632862 ignition[763]: Stage: kargs Jan 16 08:58:47.633437 ignition[763]: no configs at "/usr/lib/ignition/base.d" Jan 16 08:58:47.633463 ignition[763]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 16 08:58:47.639165 ignition[763]: kargs: kargs passed Jan 16 08:58:47.639553 ignition[763]: Ignition finished successfully Jan 16 08:58:47.644406 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 16 08:58:47.652235 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 16 08:58:47.681531 ignition[769]: Ignition 2.19.0 Jan 16 08:58:47.681548 ignition[769]: Stage: disks Jan 16 08:58:47.685589 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 16 08:58:47.681879 ignition[769]: no configs at "/usr/lib/ignition/base.d" Jan 16 08:58:47.688176 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 16 08:58:47.681897 ignition[769]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 16 08:58:47.689096 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 16 08:58:47.683471 ignition[769]: disks: disks passed Jan 16 08:58:47.690120 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 16 08:58:47.683557 ignition[769]: Ignition finished successfully Jan 16 08:58:47.691096 systemd[1]: Reached target sysinit.target - System Initialization. Jan 16 08:58:47.692203 systemd[1]: Reached target basic.target - Basic System. Jan 16 08:58:47.700357 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 16 08:58:47.740693 systemd-fsck[778]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jan 16 08:58:47.747556 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 16 08:58:47.756026 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 16 08:58:47.904117 kernel: EXT4-fs (vda9): mounted filesystem 39899d4c-a8b1-4feb-9875-e812cc535888 r/w with ordered data mode. Quota mode: none. Jan 16 08:58:47.904894 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 16 08:58:47.906364 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 16 08:58:47.920198 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 16 08:58:47.924704 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 16 08:58:47.926936 systemd[1]: Starting flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent... Jan 16 08:58:47.946157 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Jan 16 08:58:47.947272 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 16 08:58:47.947327 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 16 08:58:47.961529 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (786) Jan 16 08:58:47.961567 kernel: BTRFS info (device vda6): first mount of filesystem 70d8a0b5-70da-4efb-a618-d15543718b1e Jan 16 08:58:47.961598 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 16 08:58:47.961611 kernel: BTRFS info (device vda6): using free space tree Jan 16 08:58:47.964346 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 16 08:58:47.969903 kernel: BTRFS info (device vda6): auto enabling async discard Jan 16 08:58:47.972285 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 16 08:58:47.981388 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 16 08:58:48.088596 coreos-metadata[788]: Jan 16 08:58:48.084 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 16 08:58:48.106142 coreos-metadata[788]: Jan 16 08:58:48.105 INFO Fetch successful Jan 16 08:58:48.111688 coreos-metadata[789]: Jan 16 08:58:48.111 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 16 08:58:48.116698 initrd-setup-root[818]: cut: /sysroot/etc/passwd: No such file or directory Jan 16 08:58:48.119181 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. Jan 16 08:58:48.119408 systemd[1]: Finished flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent. Jan 16 08:58:48.125579 coreos-metadata[789]: Jan 16 08:58:48.125 INFO Fetch successful Jan 16 08:58:48.130596 initrd-setup-root[826]: cut: /sysroot/etc/group: No such file or directory Jan 16 08:58:48.135171 coreos-metadata[789]: Jan 16 08:58:48.135 INFO wrote hostname ci-4081.3.0-f-8a831eefc3 to /sysroot/etc/hostname Jan 16 08:58:48.137606 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 16 08:58:48.150737 initrd-setup-root[834]: cut: /sysroot/etc/shadow: No such file or directory Jan 16 08:58:48.162728 initrd-setup-root[841]: cut: /sysroot/etc/gshadow: No such file or directory Jan 16 08:58:48.379407 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 16 08:58:48.386897 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 16 08:58:48.389948 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 16 08:58:48.417406 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 16 08:58:48.422673 kernel: BTRFS info (device vda6): last unmount of filesystem 70d8a0b5-70da-4efb-a618-d15543718b1e Jan 16 08:58:48.452724 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 16 08:58:48.473694 ignition[912]: INFO : Ignition 2.19.0 Jan 16 08:58:48.473694 ignition[912]: INFO : Stage: mount Jan 16 08:58:48.476336 ignition[912]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 16 08:58:48.476336 ignition[912]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 16 08:58:48.479387 ignition[912]: INFO : mount: mount passed Jan 16 08:58:48.479387 ignition[912]: INFO : Ignition finished successfully Jan 16 08:58:48.480047 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 16 08:58:48.489291 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 16 08:58:48.522551 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 16 08:58:48.548705 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (923) Jan 16 08:58:48.553234 kernel: BTRFS info (device vda6): first mount of filesystem 70d8a0b5-70da-4efb-a618-d15543718b1e Jan 16 08:58:48.553374 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 16 08:58:48.555376 kernel: BTRFS info (device vda6): using free space tree Jan 16 08:58:48.561832 kernel: BTRFS info (device vda6): auto enabling async discard Jan 16 08:58:48.567907 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 16 08:58:48.625402 ignition[940]: INFO : Ignition 2.19.0 Jan 16 08:58:48.625402 ignition[940]: INFO : Stage: files Jan 16 08:58:48.627064 ignition[940]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 16 08:58:48.627064 ignition[940]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 16 08:58:48.628325 ignition[940]: DEBUG : files: compiled without relabeling support, skipping Jan 16 08:58:48.630186 ignition[940]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 16 08:58:48.630186 ignition[940]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 16 08:58:48.635317 ignition[940]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 16 08:58:48.638139 ignition[940]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 16 08:58:48.638139 ignition[940]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 16 08:58:48.636239 unknown[940]: wrote ssh authorized keys file for user: core Jan 16 08:58:48.641485 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jan 16 08:58:48.641485 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jan 16 08:58:48.663014 systemd-networkd[754]: eth1: Gained IPv6LL Jan 16 08:58:48.683799 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 16 08:58:48.771738 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 16 08:58:48.774452 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 16 08:58:48.788622 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 16 08:58:48.788622 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 16 08:58:48.788622 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 16 08:58:48.788622 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 Jan 16 08:58:49.109993 systemd-networkd[754]: eth0: Gained IPv6LL Jan 16 08:58:49.241824 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 16 08:58:49.562442 ignition[940]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Jan 16 08:58:49.562442 ignition[940]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 16 08:58:49.565776 ignition[940]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 16 08:58:49.567467 ignition[940]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 16 08:58:49.567467 ignition[940]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 16 08:58:49.567467 ignition[940]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Jan 16 08:58:49.567467 ignition[940]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Jan 16 08:58:49.571387 ignition[940]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 16 08:58:49.571387 ignition[940]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 16 08:58:49.571387 ignition[940]: INFO : files: files passed Jan 16 08:58:49.571387 ignition[940]: INFO : Ignition finished successfully Jan 16 08:58:49.570415 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 16 08:58:49.581117 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 16 08:58:49.587043 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 16 08:58:49.598422 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 16 08:58:49.599336 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 16 08:58:49.625285 initrd-setup-root-after-ignition[968]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 16 08:58:49.625285 initrd-setup-root-after-ignition[968]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 16 08:58:49.628762 initrd-setup-root-after-ignition[972]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 16 08:58:49.631346 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 16 08:58:49.632887 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 16 08:58:49.638057 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 16 08:58:49.680789 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 16 08:58:49.680944 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 16 08:58:49.682430 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 16 08:58:49.683091 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 16 08:58:49.684181 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 16 08:58:49.689132 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 16 08:58:49.729170 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 16 08:58:49.737021 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 16 08:58:49.761876 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 16 08:58:49.763342 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 16 08:58:49.764154 systemd[1]: Stopped target timers.target - Timer Units. Jan 16 08:58:49.765888 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 16 08:58:49.766150 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 16 08:58:49.768254 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 16 08:58:49.769095 systemd[1]: Stopped target basic.target - Basic System. Jan 16 08:58:49.769957 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 16 08:58:49.770929 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 16 08:58:49.772166 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 16 08:58:49.773200 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 16 08:58:49.774186 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 16 08:58:49.775344 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 16 08:58:49.776364 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 16 08:58:49.777195 systemd[1]: Stopped target swap.target - Swaps. Jan 16 08:58:49.777986 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 16 08:58:49.778200 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 16 08:58:49.779308 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 16 08:58:49.780794 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 16 08:58:49.782012 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 16 08:58:49.782198 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 16 08:58:49.783311 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 16 08:58:49.783549 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 16 08:58:49.784697 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 16 08:58:49.784954 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 16 08:58:49.785837 systemd[1]: ignition-files.service: Deactivated successfully. Jan 16 08:58:49.786107 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 16 08:58:49.787425 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Jan 16 08:58:49.787703 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 16 08:58:49.795174 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 16 08:58:49.808300 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 16 08:58:49.811197 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 16 08:58:49.811556 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 16 08:58:49.813091 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 16 08:58:49.813311 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 16 08:58:49.823534 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 16 08:58:49.823715 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 16 08:58:49.846784 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 16 08:58:49.859870 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 16 08:58:49.888807 ignition[992]: INFO : Ignition 2.19.0 Jan 16 08:58:49.888807 ignition[992]: INFO : Stage: umount Jan 16 08:58:49.888807 ignition[992]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 16 08:58:49.888807 ignition[992]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 16 08:58:49.888807 ignition[992]: INFO : umount: umount passed Jan 16 08:58:49.888807 ignition[992]: INFO : Ignition finished successfully Jan 16 08:58:49.860040 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 16 08:58:49.887337 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 16 08:58:49.887499 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 16 08:58:49.891964 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 16 08:58:49.892064 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 16 08:58:49.894814 systemd[1]: ignition-fetch.service: Deactivated successfully. Jan 16 08:58:49.894898 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jan 16 08:58:49.895495 systemd[1]: Stopped target network.target - Network. Jan 16 08:58:49.896462 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 16 08:58:49.896562 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 16 08:58:49.897390 systemd[1]: Stopped target paths.target - Path Units. Jan 16 08:58:49.898093 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 16 08:58:49.898161 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 16 08:58:49.899064 systemd[1]: Stopped target slices.target - Slice Units. Jan 16 08:58:49.900462 systemd[1]: Stopped target sockets.target - Socket Units. Jan 16 08:58:49.901394 systemd[1]: iscsid.socket: Deactivated successfully. Jan 16 08:58:49.901462 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 16 08:58:49.902248 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 16 08:58:49.902310 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 16 08:58:49.903042 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 16 08:58:49.903117 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 16 08:58:49.904178 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 16 08:58:49.904263 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 16 08:58:49.905293 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 16 08:58:49.906274 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 16 08:58:49.907529 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 16 08:58:49.907702 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 16 08:58:49.908737 systemd-networkd[754]: eth1: DHCPv6 lease lost Jan 16 08:58:49.909177 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 16 08:58:49.909295 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 16 08:58:49.911734 systemd-networkd[754]: eth0: DHCPv6 lease lost Jan 16 08:58:49.913585 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 16 08:58:49.913841 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 16 08:58:49.918472 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 16 08:58:49.919168 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 16 08:58:49.934466 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 16 08:58:49.934589 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 16 08:58:49.948107 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 16 08:58:49.951315 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 16 08:58:49.951473 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 16 08:58:49.953395 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 16 08:58:49.953497 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 16 08:58:49.954101 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 16 08:58:49.954200 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 16 08:58:49.954774 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 16 08:58:49.954850 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 16 08:58:49.956032 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 16 08:58:49.989617 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 16 08:58:49.990014 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 16 08:58:49.992359 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 16 08:58:49.992460 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 16 08:58:49.993566 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 16 08:58:49.993643 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 16 08:58:49.995574 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 16 08:58:49.995728 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 16 08:58:49.998277 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 16 08:58:49.998395 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 16 08:58:49.999002 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 16 08:58:49.999091 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 16 08:58:50.012225 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 16 08:58:50.012952 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 16 08:58:50.013097 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 16 08:58:50.015387 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 16 08:58:50.015510 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 16 08:58:50.016013 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 16 08:58:50.016073 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 16 08:58:50.016447 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 16 08:58:50.016487 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 16 08:58:50.019695 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 16 08:58:50.019869 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 16 08:58:50.028628 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 16 08:58:50.029469 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 16 08:58:50.031421 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 16 08:58:50.042163 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 16 08:58:50.063393 systemd[1]: Switching root. Jan 16 08:58:50.108917 systemd-journald[183]: Journal stopped Jan 16 08:58:51.838879 systemd-journald[183]: Received SIGTERM from PID 1 (systemd). Jan 16 08:58:51.839042 kernel: SELinux: policy capability network_peer_controls=1 Jan 16 08:58:51.839125 kernel: SELinux: policy capability open_perms=1 Jan 16 08:58:51.839163 kernel: SELinux: policy capability extended_socket_class=1 Jan 16 08:58:51.839193 kernel: SELinux: policy capability always_check_network=0 Jan 16 08:58:51.839215 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 16 08:58:51.839235 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 16 08:58:51.839256 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 16 08:58:51.839274 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 16 08:58:51.839296 kernel: audit: type=1403 audit(1737017930.333:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 16 08:58:51.839322 systemd[1]: Successfully loaded SELinux policy in 52.725ms. Jan 16 08:58:51.839346 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 20.108ms. Jan 16 08:58:51.839365 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 16 08:58:51.839384 systemd[1]: Detected virtualization kvm. Jan 16 08:58:51.839408 systemd[1]: Detected architecture x86-64. Jan 16 08:58:51.839430 systemd[1]: Detected first boot. Jan 16 08:58:51.839451 systemd[1]: Hostname set to . Jan 16 08:58:51.839471 systemd[1]: Initializing machine ID from VM UUID. Jan 16 08:58:51.839491 zram_generator::config[1035]: No configuration found. Jan 16 08:58:51.839512 systemd[1]: Populated /etc with preset unit settings. Jan 16 08:58:51.839531 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 16 08:58:51.839549 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 16 08:58:51.839574 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 16 08:58:51.839595 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 16 08:58:51.839616 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 16 08:58:51.839652 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 16 08:58:51.839669 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 16 08:58:51.839716 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 16 08:58:51.839739 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 16 08:58:51.839760 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 16 08:58:51.839785 systemd[1]: Created slice user.slice - User and Session Slice. Jan 16 08:58:51.839802 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 16 08:58:51.839820 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 16 08:58:51.839845 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 16 08:58:51.839863 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 16 08:58:51.839881 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 16 08:58:51.839899 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 16 08:58:51.839932 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 16 08:58:51.839952 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 16 08:58:51.839975 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 16 08:58:51.839992 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 16 08:58:51.840009 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 16 08:58:51.840029 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 16 08:58:51.840048 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 16 08:58:51.840066 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 16 08:58:51.840088 systemd[1]: Reached target slices.target - Slice Units. Jan 16 08:58:51.840107 systemd[1]: Reached target swap.target - Swaps. Jan 16 08:58:51.840129 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 16 08:58:51.840152 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 16 08:58:51.840177 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 16 08:58:51.840196 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 16 08:58:51.840217 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 16 08:58:51.840238 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 16 08:58:51.840258 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 16 08:58:51.840281 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 16 08:58:51.840314 systemd[1]: Mounting media.mount - External Media Directory... Jan 16 08:58:51.840336 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:51.840356 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 16 08:58:51.840375 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 16 08:58:51.840397 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 16 08:58:51.840418 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 16 08:58:51.840440 systemd[1]: Reached target machines.target - Containers. Jan 16 08:58:51.840460 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 16 08:58:51.840486 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 16 08:58:51.840523 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 16 08:58:51.840544 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 16 08:58:51.840572 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 16 08:58:51.840600 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 16 08:58:51.840619 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 16 08:58:51.840638 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 16 08:58:51.847977 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 16 08:58:51.848030 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 16 08:58:51.848053 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 16 08:58:51.857904 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 16 08:58:51.857951 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 16 08:58:51.857980 systemd[1]: Stopped systemd-fsck-usr.service. Jan 16 08:58:51.858008 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 16 08:58:51.858035 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 16 08:58:51.858064 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 16 08:58:51.858092 kernel: loop: module loaded Jan 16 08:58:51.858137 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 16 08:58:51.858164 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 16 08:58:51.858191 systemd[1]: verity-setup.service: Deactivated successfully. Jan 16 08:58:51.858217 systemd[1]: Stopped verity-setup.service. Jan 16 08:58:51.858247 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:51.858269 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 16 08:58:51.858291 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 16 08:58:51.858308 systemd[1]: Mounted media.mount - External Media Directory. Jan 16 08:58:51.858326 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 16 08:58:51.858371 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 16 08:58:51.858392 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 16 08:58:51.858409 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 16 08:58:51.858434 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 16 08:58:51.858465 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 16 08:58:51.858487 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 16 08:58:51.858506 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 16 08:58:51.858526 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 16 08:58:51.858549 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 16 08:58:51.858567 kernel: fuse: init (API version 7.39) Jan 16 08:58:51.858592 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 16 08:58:51.858612 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 16 08:58:51.858633 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 16 08:58:51.858673 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 16 08:58:51.858695 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 16 08:58:51.858717 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 16 08:58:51.858740 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 16 08:58:51.858761 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 16 08:58:51.858782 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 16 08:58:51.858815 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 16 08:58:51.858841 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 16 08:58:51.858868 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 16 08:58:51.858897 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 16 08:58:51.858927 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 16 08:58:51.858956 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 16 08:58:51.858983 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 16 08:58:51.859012 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 16 08:58:51.859125 systemd-journald[1111]: Collecting audit messages is disabled. Jan 16 08:58:51.859183 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 16 08:58:51.859213 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 16 08:58:51.859245 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 16 08:58:51.859273 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 16 08:58:51.859304 systemd-journald[1111]: Journal started Jan 16 08:58:51.859369 systemd-journald[1111]: Runtime Journal (/run/log/journal/b32100f301ca40d9bb78873a577f8845) is 4.9M, max 39.3M, 34.4M free. Jan 16 08:58:51.301315 systemd[1]: Queued start job for default target multi-user.target. Jan 16 08:58:51.913771 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 16 08:58:51.913834 systemd[1]: Started systemd-journald.service - Journal Service. Jan 16 08:58:51.913869 kernel: loop0: detected capacity change from 0 to 8 Jan 16 08:58:51.326168 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 16 08:58:51.327087 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 16 08:58:51.889550 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 16 08:58:51.890494 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 16 08:58:51.937013 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 16 08:58:51.891573 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 16 08:58:51.892740 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 16 08:58:51.893842 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 16 08:58:51.894492 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 16 08:58:51.901256 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 16 08:58:51.912092 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 16 08:58:51.915323 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 16 08:58:51.945833 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 16 08:58:51.960955 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 16 08:58:51.962453 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 16 08:58:51.976134 kernel: loop1: detected capacity change from 0 to 140768 Jan 16 08:58:51.996087 systemd-journald[1111]: Time spent on flushing to /var/log/journal/b32100f301ca40d9bb78873a577f8845 is 84.159ms for 997 entries. Jan 16 08:58:51.996087 systemd-journald[1111]: System Journal (/var/log/journal/b32100f301ca40d9bb78873a577f8845) is 8.0M, max 195.6M, 187.6M free. Jan 16 08:58:52.122242 systemd-journald[1111]: Received client request to flush runtime journal. Jan 16 08:58:52.122381 kernel: ACPI: bus type drm_connector registered Jan 16 08:58:52.122435 kernel: loop2: detected capacity change from 0 to 210664 Jan 16 08:58:52.028183 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 16 08:58:52.076481 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 16 08:58:52.076823 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 16 08:58:52.079029 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 16 08:58:52.127396 systemd-tmpfiles[1135]: ACLs are not supported, ignoring. Jan 16 08:58:52.127422 systemd-tmpfiles[1135]: ACLs are not supported, ignoring. Jan 16 08:58:52.131862 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 16 08:58:52.145689 kernel: loop3: detected capacity change from 0 to 142488 Jan 16 08:58:52.157782 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 16 08:58:52.170050 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 16 08:58:52.177728 udevadm[1168]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jan 16 08:58:52.237418 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 16 08:58:52.248595 kernel: loop4: detected capacity change from 0 to 8 Jan 16 08:58:52.248051 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 16 08:58:52.253700 kernel: loop5: detected capacity change from 0 to 140768 Jan 16 08:58:52.286403 kernel: loop6: detected capacity change from 0 to 210664 Jan 16 08:58:52.326051 kernel: loop7: detected capacity change from 0 to 142488 Jan 16 08:58:52.333608 systemd-tmpfiles[1182]: ACLs are not supported, ignoring. Jan 16 08:58:52.333643 systemd-tmpfiles[1182]: ACLs are not supported, ignoring. Jan 16 08:58:52.341809 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 16 08:58:52.354731 (sd-merge)[1181]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. Jan 16 08:58:52.357913 (sd-merge)[1181]: Merged extensions into '/usr'. Jan 16 08:58:52.370807 systemd[1]: Reloading requested from client PID 1134 ('systemd-sysext') (unit systemd-sysext.service)... Jan 16 08:58:52.371038 systemd[1]: Reloading... Jan 16 08:58:52.533735 zram_generator::config[1209]: No configuration found. Jan 16 08:58:52.807686 ldconfig[1129]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 16 08:58:52.840321 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 16 08:58:52.932040 systemd[1]: Reloading finished in 557 ms. Jan 16 08:58:52.982779 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 16 08:58:52.988442 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 16 08:58:52.998212 systemd[1]: Starting ensure-sysext.service... Jan 16 08:58:53.011432 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 16 08:58:53.037860 systemd[1]: Reloading requested from client PID 1253 ('systemctl') (unit ensure-sysext.service)... Jan 16 08:58:53.037890 systemd[1]: Reloading... Jan 16 08:58:53.072179 systemd-tmpfiles[1254]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 16 08:58:53.072693 systemd-tmpfiles[1254]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 16 08:58:53.076223 systemd-tmpfiles[1254]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 16 08:58:53.076543 systemd-tmpfiles[1254]: ACLs are not supported, ignoring. Jan 16 08:58:53.076614 systemd-tmpfiles[1254]: ACLs are not supported, ignoring. Jan 16 08:58:53.090586 systemd-tmpfiles[1254]: Detected autofs mount point /boot during canonicalization of boot. Jan 16 08:58:53.090604 systemd-tmpfiles[1254]: Skipping /boot Jan 16 08:58:53.123842 systemd-tmpfiles[1254]: Detected autofs mount point /boot during canonicalization of boot. Jan 16 08:58:53.123861 systemd-tmpfiles[1254]: Skipping /boot Jan 16 08:58:53.220687 zram_generator::config[1281]: No configuration found. Jan 16 08:58:53.426212 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 16 08:58:53.525480 systemd[1]: Reloading finished in 486 ms. Jan 16 08:58:53.550015 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 16 08:58:53.556776 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 16 08:58:53.575167 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 16 08:58:53.585605 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 16 08:58:53.599396 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 16 08:58:53.611390 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 16 08:58:53.619025 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 16 08:58:53.634773 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 16 08:58:53.650987 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:53.651372 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 16 08:58:53.663289 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 16 08:58:53.674621 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 16 08:58:53.685439 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 16 08:58:53.686400 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 16 08:58:53.688787 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:53.728079 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 16 08:58:53.734550 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:53.736220 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 16 08:58:53.736640 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 16 08:58:53.738946 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:53.745640 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:53.747270 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 16 08:58:53.758114 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 16 08:58:53.760304 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 16 08:58:53.761139 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:53.783245 systemd[1]: Finished ensure-sysext.service. Jan 16 08:58:53.785488 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 16 08:58:53.786397 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 16 08:58:53.789133 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 16 08:58:53.819161 systemd-udevd[1332]: Using default interface naming scheme 'v255'. Jan 16 08:58:53.822744 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 16 08:58:53.827147 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 16 08:58:53.829550 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 16 08:58:53.829900 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 16 08:58:53.831419 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 16 08:58:53.831835 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 16 08:58:53.836013 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 16 08:58:53.844611 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 16 08:58:53.844772 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 16 08:58:53.850493 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 16 08:58:53.852941 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 16 08:58:53.853480 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 16 08:58:53.856957 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 16 08:58:53.900474 augenrules[1362]: No rules Jan 16 08:58:53.902107 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 16 08:58:53.928769 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 16 08:58:53.931869 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 16 08:58:53.944991 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 16 08:58:53.945969 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 16 08:58:54.144952 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 16 08:58:54.145811 systemd[1]: Reached target time-set.target - System Time Set. Jan 16 08:58:54.168862 systemd-resolved[1330]: Positive Trust Anchors: Jan 16 08:58:54.168881 systemd-resolved[1330]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 16 08:58:54.168920 systemd-resolved[1330]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 16 08:58:54.179101 systemd-resolved[1330]: Using system hostname 'ci-4081.3.0-f-8a831eefc3'. Jan 16 08:58:54.181922 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 16 08:58:54.182758 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 16 08:58:54.198126 systemd-networkd[1372]: lo: Link UP Jan 16 08:58:54.198140 systemd-networkd[1372]: lo: Gained carrier Jan 16 08:58:54.202558 systemd-networkd[1372]: Enumeration completed Jan 16 08:58:54.202772 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 16 08:58:54.203392 systemd[1]: Reached target network.target - Network. Jan 16 08:58:54.214951 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 16 08:58:54.247441 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Jan 16 08:58:54.248159 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:54.248429 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 16 08:58:54.257007 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 16 08:58:54.268966 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 16 08:58:54.274039 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 16 08:58:54.274871 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 16 08:58:54.274938 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 16 08:58:54.274971 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 16 08:58:54.276419 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 16 08:58:54.282297 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 16 08:58:54.282578 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 16 08:58:54.287608 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 16 08:58:54.291206 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 16 08:58:54.294207 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 16 08:58:54.338819 kernel: ISO 9660 Extensions: RRIP_1991A Jan 16 08:58:54.343857 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Jan 16 08:58:54.350789 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 16 08:58:54.351926 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 16 08:58:54.354367 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 16 08:58:54.367085 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1387) Jan 16 08:58:54.448566 systemd-networkd[1372]: eth1: Configuring with /run/systemd/network/10-86:dd:0a:5f:e5:e7.network. Jan 16 08:58:54.451451 systemd-networkd[1372]: eth1: Link UP Jan 16 08:58:54.451634 systemd-networkd[1372]: eth1: Gained carrier Jan 16 08:58:54.458855 systemd-timesyncd[1354]: Network configuration changed, trying to establish connection. Jan 16 08:58:54.461151 systemd-networkd[1372]: eth0: Configuring with /run/systemd/network/10-92:80:c4:e3:0a:23.network. Jan 16 08:58:54.462421 systemd-timesyncd[1354]: Network configuration changed, trying to establish connection. Jan 16 08:58:54.464027 systemd-networkd[1372]: eth0: Link UP Jan 16 08:58:54.464041 systemd-networkd[1372]: eth0: Gained carrier Jan 16 08:58:54.469839 systemd-timesyncd[1354]: Network configuration changed, trying to establish connection. Jan 16 08:58:54.493692 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Jan 16 08:58:54.499721 kernel: ACPI: button: Power Button [PWRF] Jan 16 08:58:54.531299 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 16 08:58:54.540905 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Jan 16 08:58:54.540100 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 16 08:58:54.564678 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Jan 16 08:58:54.585775 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 16 08:58:54.638234 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 16 08:58:54.641677 kernel: mousedev: PS/2 mouse device common for all mice Jan 16 08:58:54.660990 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Jan 16 08:58:54.661129 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Jan 16 08:58:54.666685 kernel: Console: switching to colour dummy device 80x25 Jan 16 08:58:54.668689 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Jan 16 08:58:54.668783 kernel: [drm] features: -context_init Jan 16 08:58:54.668806 kernel: [drm] number of scanouts: 1 Jan 16 08:58:54.668839 kernel: [drm] number of cap sets: 0 Jan 16 08:58:54.671727 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 Jan 16 08:58:54.680730 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Jan 16 08:58:54.680833 kernel: Console: switching to colour frame buffer device 128x48 Jan 16 08:58:54.697105 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Jan 16 08:58:54.724086 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 16 08:58:54.725776 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 16 08:58:54.739171 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 16 08:58:54.890979 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 16 08:58:54.937140 kernel: EDAC MC: Ver: 3.0.0 Jan 16 08:58:54.969627 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 16 08:58:54.983497 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 16 08:58:55.005065 lvm[1434]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 16 08:58:55.045520 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 16 08:58:55.048793 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 16 08:58:55.049983 systemd[1]: Reached target sysinit.target - System Initialization. Jan 16 08:58:55.050600 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 16 08:58:55.050834 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 16 08:58:55.051566 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 16 08:58:55.052061 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 16 08:58:55.052222 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 16 08:58:55.052328 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 16 08:58:55.052372 systemd[1]: Reached target paths.target - Path Units. Jan 16 08:58:55.052460 systemd[1]: Reached target timers.target - Timer Units. Jan 16 08:58:55.054777 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 16 08:58:55.058298 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 16 08:58:55.066309 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 16 08:58:55.072906 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 16 08:58:55.074148 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 16 08:58:55.077853 systemd[1]: Reached target sockets.target - Socket Units. Jan 16 08:58:55.079717 systemd[1]: Reached target basic.target - Basic System. Jan 16 08:58:55.081521 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 16 08:58:55.081623 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 16 08:58:55.084903 systemd[1]: Starting containerd.service - containerd container runtime... Jan 16 08:58:55.090772 lvm[1438]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 16 08:58:55.098963 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jan 16 08:58:55.111995 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 16 08:58:55.120857 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 16 08:58:55.134362 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 16 08:58:55.140241 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 16 08:58:55.145044 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 16 08:58:55.146594 jq[1442]: false Jan 16 08:58:55.154053 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 16 08:58:55.166068 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 16 08:58:55.177004 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 16 08:58:55.193097 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 16 08:58:55.201109 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 16 08:58:55.204148 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 16 08:58:55.215310 systemd[1]: Starting update-engine.service - Update Engine... Jan 16 08:58:55.234012 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 16 08:58:55.238577 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 16 08:58:55.253421 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 16 08:58:55.256090 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 16 08:58:55.312909 extend-filesystems[1443]: Found loop4 Jan 16 08:58:55.312909 extend-filesystems[1443]: Found loop5 Jan 16 08:58:55.312909 extend-filesystems[1443]: Found loop6 Jan 16 08:58:55.312909 extend-filesystems[1443]: Found loop7 Jan 16 08:58:55.312509 dbus-daemon[1441]: [system] SELinux support is enabled Jan 16 08:58:55.325878 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda1 Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda2 Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda3 Jan 16 08:58:55.371786 extend-filesystems[1443]: Found usr Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda4 Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda6 Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda7 Jan 16 08:58:55.371786 extend-filesystems[1443]: Found vda9 Jan 16 08:58:55.371786 extend-filesystems[1443]: Checking size of /dev/vda9 Jan 16 08:58:55.427115 jq[1454]: true Jan 16 08:58:55.361226 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 16 08:58:55.427710 update_engine[1452]: I20250116 08:58:55.376956 1452 main.cc:92] Flatcar Update Engine starting Jan 16 08:58:55.427710 update_engine[1452]: I20250116 08:58:55.416074 1452 update_check_scheduler.cc:74] Next update check in 2m56s Jan 16 08:58:55.361513 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 16 08:58:55.369311 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 16 08:58:55.369399 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 16 08:58:55.370318 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 16 08:58:55.370440 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Jan 16 08:58:55.370470 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 16 08:58:55.385885 (ntainerd)[1466]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 16 08:58:55.404041 systemd[1]: motdgen.service: Deactivated successfully. Jan 16 08:58:55.404440 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 16 08:58:55.415466 systemd[1]: Started update-engine.service - Update Engine. Jan 16 08:58:55.438988 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 16 08:58:55.444944 coreos-metadata[1440]: Jan 16 08:58:55.441 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 16 08:58:55.463629 coreos-metadata[1440]: Jan 16 08:58:55.460 INFO Fetch successful Jan 16 08:58:55.465984 tar[1457]: linux-amd64/helm Jan 16 08:58:55.498494 extend-filesystems[1443]: Resized partition /dev/vda9 Jan 16 08:58:55.509958 systemd-networkd[1372]: eth0: Gained IPv6LL Jan 16 08:58:55.518496 systemd-timesyncd[1354]: Network configuration changed, trying to establish connection. Jan 16 08:58:55.520841 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 16 08:58:55.534722 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (1389) Jan 16 08:58:55.539724 systemd[1]: Reached target network-online.target - Network is Online. Jan 16 08:58:55.546679 jq[1472]: true Jan 16 08:58:55.549681 extend-filesystems[1481]: resize2fs 1.47.1 (20-May-2024) Jan 16 08:58:55.555953 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 16 08:58:55.567111 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 16 08:58:55.577700 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Jan 16 08:58:55.760319 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 16 08:58:55.775110 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jan 16 08:58:55.783532 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 16 08:58:55.793301 systemd-logind[1451]: New seat seat0. Jan 16 08:58:55.800491 systemd-logind[1451]: Watching system buttons on /dev/input/event1 (Power Button) Jan 16 08:58:55.804624 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Jan 16 08:58:55.805003 systemd-logind[1451]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 16 08:58:55.806963 systemd[1]: Started systemd-logind.service - User Login Management. Jan 16 08:58:55.822066 extend-filesystems[1481]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 16 08:58:55.822066 extend-filesystems[1481]: old_desc_blocks = 1, new_desc_blocks = 8 Jan 16 08:58:55.822066 extend-filesystems[1481]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Jan 16 08:58:55.839515 extend-filesystems[1443]: Resized filesystem in /dev/vda9 Jan 16 08:58:55.839515 extend-filesystems[1443]: Found vdb Jan 16 08:58:55.836846 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 16 08:58:55.837197 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 16 08:58:55.883861 bash[1514]: Updated "/home/core/.ssh/authorized_keys" Jan 16 08:58:55.887292 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 16 08:58:55.913866 systemd[1]: Starting sshkeys.service... Jan 16 08:58:56.006824 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jan 16 08:58:56.019829 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jan 16 08:58:56.159502 coreos-metadata[1521]: Jan 16 08:58:56.158 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 16 08:58:56.194534 coreos-metadata[1521]: Jan 16 08:58:56.194 INFO Fetch successful Jan 16 08:58:56.201409 locksmithd[1476]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 16 08:58:56.230428 unknown[1521]: wrote ssh authorized keys file for user: core Jan 16 08:58:56.284505 systemd-networkd[1372]: eth1: Gained IPv6LL Jan 16 08:58:56.285004 systemd-timesyncd[1354]: Network configuration changed, trying to establish connection. Jan 16 08:58:56.302003 update-ssh-keys[1531]: Updated "/home/core/.ssh/authorized_keys" Jan 16 08:58:56.304317 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jan 16 08:58:56.309199 systemd[1]: Finished sshkeys.service. Jan 16 08:58:56.510552 containerd[1466]: time="2025-01-16T08:58:56.508528451Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Jan 16 08:58:56.594445 containerd[1466]: time="2025-01-16T08:58:56.593834000Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 16 08:58:56.600567 containerd[1466]: time="2025-01-16T08:58:56.600507631Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.71-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 16 08:58:56.601113 containerd[1466]: time="2025-01-16T08:58:56.601071732Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 16 08:58:56.601240 containerd[1466]: time="2025-01-16T08:58:56.601222181Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 16 08:58:56.601534 containerd[1466]: time="2025-01-16T08:58:56.601507191Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 16 08:58:56.601989 containerd[1466]: time="2025-01-16T08:58:56.601964730Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 16 08:58:56.602212 containerd[1466]: time="2025-01-16T08:58:56.602187728Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 16 08:58:56.602515 containerd[1466]: time="2025-01-16T08:58:56.602489648Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 16 08:58:56.604612 containerd[1466]: time="2025-01-16T08:58:56.603659793Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 16 08:58:56.604612 containerd[1466]: time="2025-01-16T08:58:56.603793022Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 16 08:58:56.604612 containerd[1466]: time="2025-01-16T08:58:56.603822790Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 16 08:58:56.604612 containerd[1466]: time="2025-01-16T08:58:56.603840059Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 16 08:58:56.604612 containerd[1466]: time="2025-01-16T08:58:56.604020679Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 16 08:58:56.604612 containerd[1466]: time="2025-01-16T08:58:56.604353859Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 16 08:58:56.605606 containerd[1466]: time="2025-01-16T08:58:56.605477329Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 16 08:58:56.606691 containerd[1466]: time="2025-01-16T08:58:56.606311668Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 16 08:58:56.606691 containerd[1466]: time="2025-01-16T08:58:56.606544758Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 16 08:58:56.606691 containerd[1466]: time="2025-01-16T08:58:56.606620838Z" level=info msg="metadata content store policy set" policy=shared Jan 16 08:58:56.618829 containerd[1466]: time="2025-01-16T08:58:56.617858906Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 16 08:58:56.618829 containerd[1466]: time="2025-01-16T08:58:56.617954371Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 16 08:58:56.618829 containerd[1466]: time="2025-01-16T08:58:56.617973063Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 16 08:58:56.618829 containerd[1466]: time="2025-01-16T08:58:56.618083676Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 16 08:58:56.618829 containerd[1466]: time="2025-01-16T08:58:56.618103169Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 16 08:58:56.618829 containerd[1466]: time="2025-01-16T08:58:56.618373287Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 16 08:58:56.618829 containerd[1466]: time="2025-01-16T08:58:56.618749287Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 16 08:58:56.619412 containerd[1466]: time="2025-01-16T08:58:56.619221836Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 16 08:58:56.619634 containerd[1466]: time="2025-01-16T08:58:56.619509911Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 16 08:58:56.619634 containerd[1466]: time="2025-01-16T08:58:56.619539159Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 16 08:58:56.620213 containerd[1466]: time="2025-01-16T08:58:56.619566103Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.620213 containerd[1466]: time="2025-01-16T08:58:56.620068924Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.620213 containerd[1466]: time="2025-01-16T08:58:56.620087528Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.623847 containerd[1466]: time="2025-01-16T08:58:56.620379607Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.623847 containerd[1466]: time="2025-01-16T08:58:56.623719676Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.623847 containerd[1466]: time="2025-01-16T08:58:56.623790438Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624383120Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624422769Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624477850Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624500560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624531139Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624551665Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624570484Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624599900Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624620509Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624668816Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624688596Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624711197Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624756180Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625364 containerd[1466]: time="2025-01-16T08:58:56.624780515Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.624798891Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.624837149Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.624875070Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.624905712Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.624924124Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625021384Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625135431Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625153779Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625178099Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625204505Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625222164Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625239564Z" level=info msg="NRI interface is disabled by configuration." Jan 16 08:58:56.625978 containerd[1466]: time="2025-01-16T08:58:56.625254258Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 16 08:58:56.628221 containerd[1466]: time="2025-01-16T08:58:56.627912005Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 16 08:58:56.631575 containerd[1466]: time="2025-01-16T08:58:56.628677063Z" level=info msg="Connect containerd service" Jan 16 08:58:56.631575 containerd[1466]: time="2025-01-16T08:58:56.628768211Z" level=info msg="using legacy CRI server" Jan 16 08:58:56.631575 containerd[1466]: time="2025-01-16T08:58:56.628778156Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 16 08:58:56.631575 containerd[1466]: time="2025-01-16T08:58:56.628971624Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 16 08:58:56.633247 containerd[1466]: time="2025-01-16T08:58:56.632603528Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 16 08:58:56.636575 containerd[1466]: time="2025-01-16T08:58:56.635766672Z" level=info msg="Start subscribing containerd event" Jan 16 08:58:56.636575 containerd[1466]: time="2025-01-16T08:58:56.636039046Z" level=info msg="Start recovering state" Jan 16 08:58:56.636575 containerd[1466]: time="2025-01-16T08:58:56.636189657Z" level=info msg="Start event monitor" Jan 16 08:58:56.636575 containerd[1466]: time="2025-01-16T08:58:56.636219563Z" level=info msg="Start snapshots syncer" Jan 16 08:58:56.636575 containerd[1466]: time="2025-01-16T08:58:56.636233689Z" level=info msg="Start cni network conf syncer for default" Jan 16 08:58:56.636575 containerd[1466]: time="2025-01-16T08:58:56.636246078Z" level=info msg="Start streaming server" Jan 16 08:58:56.641040 containerd[1466]: time="2025-01-16T08:58:56.640973859Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 16 08:58:56.644615 containerd[1466]: time="2025-01-16T08:58:56.642924198Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 16 08:58:56.653096 containerd[1466]: time="2025-01-16T08:58:56.653035906Z" level=info msg="containerd successfully booted in 0.150392s" Jan 16 08:58:56.653181 systemd[1]: Started containerd.service - containerd container runtime. Jan 16 08:58:56.682406 sshd_keygen[1478]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 16 08:58:56.777294 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 16 08:58:56.794254 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 16 08:58:56.843577 systemd[1]: issuegen.service: Deactivated successfully. Jan 16 08:58:56.844198 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 16 08:58:56.862544 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 16 08:58:56.892086 tar[1457]: linux-amd64/LICENSE Jan 16 08:58:56.892678 tar[1457]: linux-amd64/README.md Jan 16 08:58:56.914678 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 16 08:58:56.919405 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 16 08:58:56.932510 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 16 08:58:56.945077 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 16 08:58:56.949774 systemd[1]: Reached target getty.target - Login Prompts. Jan 16 08:58:57.638021 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:58:57.643881 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 16 08:58:57.648404 (kubelet)[1562]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 16 08:58:57.649261 systemd[1]: Startup finished in 1.541s (kernel) + 6.550s (initrd) + 7.366s (userspace) = 15.458s. Jan 16 08:58:58.896355 kubelet[1562]: E0116 08:58:58.896189 1562 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 16 08:58:58.900510 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 16 08:58:58.900830 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 16 08:58:58.901385 systemd[1]: kubelet.service: Consumed 1.524s CPU time. Jan 16 08:59:04.824429 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 16 08:59:04.832329 systemd[1]: Started sshd@0-147.182.199.56:22-139.178.68.195:33652.service - OpenSSH per-connection server daemon (139.178.68.195:33652). Jan 16 08:59:04.966578 sshd[1575]: Accepted publickey for core from 139.178.68.195 port 33652 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 08:59:04.973159 sshd[1575]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 08:59:05.001027 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 16 08:59:05.012462 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 16 08:59:05.017918 systemd-logind[1451]: New session 1 of user core. Jan 16 08:59:05.044315 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 16 08:59:05.055306 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 16 08:59:05.074690 (systemd)[1579]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 16 08:59:05.280168 systemd[1579]: Queued start job for default target default.target. Jan 16 08:59:05.302252 systemd[1579]: Created slice app.slice - User Application Slice. Jan 16 08:59:05.302570 systemd[1579]: Reached target paths.target - Paths. Jan 16 08:59:05.302779 systemd[1579]: Reached target timers.target - Timers. Jan 16 08:59:05.306420 systemd[1579]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 16 08:59:05.345140 systemd[1579]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 16 08:59:05.345405 systemd[1579]: Reached target sockets.target - Sockets. Jan 16 08:59:05.345433 systemd[1579]: Reached target basic.target - Basic System. Jan 16 08:59:05.345514 systemd[1579]: Reached target default.target - Main User Target. Jan 16 08:59:05.345742 systemd[1579]: Startup finished in 250ms. Jan 16 08:59:05.346148 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 16 08:59:05.353078 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 16 08:59:05.435208 systemd[1]: Started sshd@1-147.182.199.56:22-139.178.68.195:33666.service - OpenSSH per-connection server daemon (139.178.68.195:33666). Jan 16 08:59:05.514361 sshd[1590]: Accepted publickey for core from 139.178.68.195 port 33666 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 08:59:05.519208 sshd[1590]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 08:59:05.544030 systemd-logind[1451]: New session 2 of user core. Jan 16 08:59:05.553108 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 16 08:59:05.651683 sshd[1590]: pam_unix(sshd:session): session closed for user core Jan 16 08:59:05.666188 systemd[1]: sshd@1-147.182.199.56:22-139.178.68.195:33666.service: Deactivated successfully. Jan 16 08:59:05.671593 systemd[1]: session-2.scope: Deactivated successfully. Jan 16 08:59:05.675250 systemd-logind[1451]: Session 2 logged out. Waiting for processes to exit. Jan 16 08:59:05.685251 systemd[1]: Started sshd@2-147.182.199.56:22-139.178.68.195:33678.service - OpenSSH per-connection server daemon (139.178.68.195:33678). Jan 16 08:59:05.687057 systemd-logind[1451]: Removed session 2. Jan 16 08:59:05.764018 sshd[1597]: Accepted publickey for core from 139.178.68.195 port 33678 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 08:59:05.767198 sshd[1597]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 08:59:05.778393 systemd-logind[1451]: New session 3 of user core. Jan 16 08:59:05.792634 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 16 08:59:05.866786 sshd[1597]: pam_unix(sshd:session): session closed for user core Jan 16 08:59:05.886694 systemd[1]: sshd@2-147.182.199.56:22-139.178.68.195:33678.service: Deactivated successfully. Jan 16 08:59:05.891027 systemd[1]: session-3.scope: Deactivated successfully. Jan 16 08:59:05.895188 systemd-logind[1451]: Session 3 logged out. Waiting for processes to exit. Jan 16 08:59:05.913010 systemd[1]: Started sshd@3-147.182.199.56:22-139.178.68.195:33682.service - OpenSSH per-connection server daemon (139.178.68.195:33682). Jan 16 08:59:05.916308 systemd-logind[1451]: Removed session 3. Jan 16 08:59:05.978748 sshd[1604]: Accepted publickey for core from 139.178.68.195 port 33682 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 08:59:05.981974 sshd[1604]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 08:59:06.002051 systemd-logind[1451]: New session 4 of user core. Jan 16 08:59:06.025075 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 16 08:59:06.127073 sshd[1604]: pam_unix(sshd:session): session closed for user core Jan 16 08:59:06.144194 systemd[1]: sshd@3-147.182.199.56:22-139.178.68.195:33682.service: Deactivated successfully. Jan 16 08:59:06.149792 systemd[1]: session-4.scope: Deactivated successfully. Jan 16 08:59:06.157079 systemd-logind[1451]: Session 4 logged out. Waiting for processes to exit. Jan 16 08:59:06.163324 systemd[1]: Started sshd@4-147.182.199.56:22-139.178.68.195:33698.service - OpenSSH per-connection server daemon (139.178.68.195:33698). Jan 16 08:59:06.165950 systemd-logind[1451]: Removed session 4. Jan 16 08:59:06.251633 sshd[1611]: Accepted publickey for core from 139.178.68.195 port 33698 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 08:59:06.255235 sshd[1611]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 08:59:06.265601 systemd-logind[1451]: New session 5 of user core. Jan 16 08:59:06.272229 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 16 08:59:06.368381 sudo[1614]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 16 08:59:06.368963 sudo[1614]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 16 08:59:07.072595 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 16 08:59:07.089354 (dockerd)[1630]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 16 08:59:07.707600 dockerd[1630]: time="2025-01-16T08:59:07.707087728Z" level=info msg="Starting up" Jan 16 08:59:07.865588 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1699789631-merged.mount: Deactivated successfully. Jan 16 08:59:07.922054 dockerd[1630]: time="2025-01-16T08:59:07.921978254Z" level=info msg="Loading containers: start." Jan 16 08:59:08.160835 kernel: Initializing XFRM netlink socket Jan 16 08:59:08.224437 systemd-timesyncd[1354]: Network configuration changed, trying to establish connection. Jan 16 08:59:08.292461 systemd-timesyncd[1354]: Contacted time server 23.150.41.122:123 (2.flatcar.pool.ntp.org). Jan 16 08:59:08.292713 systemd-timesyncd[1354]: Initial clock synchronization to Thu 2025-01-16 08:59:08.376028 UTC. Jan 16 08:59:08.340810 systemd-networkd[1372]: docker0: Link UP Jan 16 08:59:08.361122 dockerd[1630]: time="2025-01-16T08:59:08.361056528Z" level=info msg="Loading containers: done." Jan 16 08:59:08.383448 dockerd[1630]: time="2025-01-16T08:59:08.383364521Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 16 08:59:08.383735 dockerd[1630]: time="2025-01-16T08:59:08.383548057Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Jan 16 08:59:08.383787 dockerd[1630]: time="2025-01-16T08:59:08.383740493Z" level=info msg="Daemon has completed initialization" Jan 16 08:59:08.448599 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 16 08:59:08.449296 dockerd[1630]: time="2025-01-16T08:59:08.445832470Z" level=info msg="API listen on /run/docker.sock" Jan 16 08:59:09.151476 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 16 08:59:09.160151 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 16 08:59:09.368173 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:59:09.370265 (kubelet)[1782]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 16 08:59:09.456952 kubelet[1782]: E0116 08:59:09.456757 1782 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 16 08:59:09.463167 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 16 08:59:09.463428 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 16 08:59:09.829346 containerd[1466]: time="2025-01-16T08:59:09.828837925Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.9\"" Jan 16 08:59:10.541899 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount679239190.mount: Deactivated successfully. Jan 16 08:59:12.432042 containerd[1466]: time="2025-01-16T08:59:12.431960126Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:12.434136 containerd[1466]: time="2025-01-16T08:59:12.434037506Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.9: active requests=0, bytes read=32677012" Jan 16 08:59:12.435397 containerd[1466]: time="2025-01-16T08:59:12.434972874Z" level=info msg="ImageCreate event name:\"sha256:4f53be91109c4dd4658bb0141e8af556b94293ec9fad72b2b62a617edb48e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:12.440696 containerd[1466]: time="2025-01-16T08:59:12.440522747Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:540de8f810ac963b8ed93f7393a8746d68e7e8a2c79ea58ff409ac5b9ca6a9fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:12.443222 containerd[1466]: time="2025-01-16T08:59:12.442894662Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.9\" with image id \"sha256:4f53be91109c4dd4658bb0141e8af556b94293ec9fad72b2b62a617edb48e5c4\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:540de8f810ac963b8ed93f7393a8746d68e7e8a2c79ea58ff409ac5b9ca6a9fc\", size \"32673812\" in 2.613978122s" Jan 16 08:59:12.443222 containerd[1466]: time="2025-01-16T08:59:12.442977276Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.9\" returns image reference \"sha256:4f53be91109c4dd4658bb0141e8af556b94293ec9fad72b2b62a617edb48e5c4\"" Jan 16 08:59:12.489510 containerd[1466]: time="2025-01-16T08:59:12.488996947Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.9\"" Jan 16 08:59:14.707829 containerd[1466]: time="2025-01-16T08:59:14.707315691Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:14.710177 containerd[1466]: time="2025-01-16T08:59:14.710082198Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.9: active requests=0, bytes read=29605745" Jan 16 08:59:14.712686 containerd[1466]: time="2025-01-16T08:59:14.711076330Z" level=info msg="ImageCreate event name:\"sha256:d4203c1bb2593a7429c3df3c040da333190e5d7e01f377d0255b7b813ca09568\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:14.717200 containerd[1466]: time="2025-01-16T08:59:14.716434346Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6350693c04956b13db2519e01ca12a0bbe58466e9f12ef8617f1429da6081f43\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:14.717958 containerd[1466]: time="2025-01-16T08:59:14.717844908Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.9\" with image id \"sha256:d4203c1bb2593a7429c3df3c040da333190e5d7e01f377d0255b7b813ca09568\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6350693c04956b13db2519e01ca12a0bbe58466e9f12ef8617f1429da6081f43\", size \"31052327\" in 2.22878393s" Jan 16 08:59:14.718245 containerd[1466]: time="2025-01-16T08:59:14.718209322Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.9\" returns image reference \"sha256:d4203c1bb2593a7429c3df3c040da333190e5d7e01f377d0255b7b813ca09568\"" Jan 16 08:59:14.759463 containerd[1466]: time="2025-01-16T08:59:14.759363035Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.9\"" Jan 16 08:59:14.761318 systemd-resolved[1330]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. Jan 16 08:59:16.276322 containerd[1466]: time="2025-01-16T08:59:16.276230326Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:16.277784 containerd[1466]: time="2025-01-16T08:59:16.277705251Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.9: active requests=0, bytes read=17783064" Jan 16 08:59:16.278459 containerd[1466]: time="2025-01-16T08:59:16.278116830Z" level=info msg="ImageCreate event name:\"sha256:41cce68b0c8c3c4862ff55ac17be57616cce36a04e719aee733e5c7c1a24b725\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:16.282701 containerd[1466]: time="2025-01-16T08:59:16.282485268Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:153efd6dc89e61a38ef273cf4c4cebd2bfee68082c2ee3d4fab5da94e4ae13d3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:16.284738 containerd[1466]: time="2025-01-16T08:59:16.284630827Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.9\" with image id \"sha256:41cce68b0c8c3c4862ff55ac17be57616cce36a04e719aee733e5c7c1a24b725\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:153efd6dc89e61a38ef273cf4c4cebd2bfee68082c2ee3d4fab5da94e4ae13d3\", size \"19229664\" in 1.525205177s" Jan 16 08:59:16.284738 containerd[1466]: time="2025-01-16T08:59:16.284734288Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.9\" returns image reference \"sha256:41cce68b0c8c3c4862ff55ac17be57616cce36a04e719aee733e5c7c1a24b725\"" Jan 16 08:59:16.340594 containerd[1466]: time="2025-01-16T08:59:16.340364494Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.9\"" Jan 16 08:59:17.523218 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1984113727.mount: Deactivated successfully. Jan 16 08:59:17.846173 systemd-resolved[1330]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Jan 16 08:59:18.257535 containerd[1466]: time="2025-01-16T08:59:18.256101022Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:18.257535 containerd[1466]: time="2025-01-16T08:59:18.257500174Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.9: active requests=0, bytes read=29058337" Jan 16 08:59:18.258882 containerd[1466]: time="2025-01-16T08:59:18.258753038Z" level=info msg="ImageCreate event name:\"sha256:4c369683c359609256b8907f424fc2355f1e7e3eeb7295b1fd8ffc5304f4cede\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:18.262579 containerd[1466]: time="2025-01-16T08:59:18.262490432Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:d78dc40d97ff862fd8ddb47f80a5ba3feec17bc73e58a60e963885e33faa0083\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:18.265309 containerd[1466]: time="2025-01-16T08:59:18.265211682Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.9\" with image id \"sha256:4c369683c359609256b8907f424fc2355f1e7e3eeb7295b1fd8ffc5304f4cede\", repo tag \"registry.k8s.io/kube-proxy:v1.30.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:d78dc40d97ff862fd8ddb47f80a5ba3feec17bc73e58a60e963885e33faa0083\", size \"29057356\" in 1.924790205s" Jan 16 08:59:18.265309 containerd[1466]: time="2025-01-16T08:59:18.265281183Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.9\" returns image reference \"sha256:4c369683c359609256b8907f424fc2355f1e7e3eeb7295b1fd8ffc5304f4cede\"" Jan 16 08:59:18.317348 containerd[1466]: time="2025-01-16T08:59:18.317300581Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Jan 16 08:59:18.946086 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1143808778.mount: Deactivated successfully. Jan 16 08:59:19.715111 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 16 08:59:19.728687 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 16 08:59:19.955697 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:59:19.973495 (kubelet)[1933]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 16 08:59:20.078836 kubelet[1933]: E0116 08:59:20.077841 1933 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 16 08:59:20.084578 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 16 08:59:20.085492 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 16 08:59:20.328985 containerd[1466]: time="2025-01-16T08:59:20.328584049Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:20.331433 containerd[1466]: time="2025-01-16T08:59:20.331030521Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Jan 16 08:59:20.332441 containerd[1466]: time="2025-01-16T08:59:20.332376619Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:20.340954 containerd[1466]: time="2025-01-16T08:59:20.340732608Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:20.343919 containerd[1466]: time="2025-01-16T08:59:20.343313669Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 2.025726461s" Jan 16 08:59:20.343919 containerd[1466]: time="2025-01-16T08:59:20.343382932Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Jan 16 08:59:20.382288 containerd[1466]: time="2025-01-16T08:59:20.382240364Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Jan 16 08:59:20.901942 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3779132155.mount: Deactivated successfully. Jan 16 08:59:20.908737 containerd[1466]: time="2025-01-16T08:59:20.907578071Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:20.909822 containerd[1466]: time="2025-01-16T08:59:20.909635938Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322290" Jan 16 08:59:20.912047 containerd[1466]: time="2025-01-16T08:59:20.910578135Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:20.912871 containerd[1466]: time="2025-01-16T08:59:20.912832908Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:20.913972 containerd[1466]: time="2025-01-16T08:59:20.913923206Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 531.638706ms" Jan 16 08:59:20.914122 containerd[1466]: time="2025-01-16T08:59:20.914102625Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Jan 16 08:59:20.950237 containerd[1466]: time="2025-01-16T08:59:20.950178884Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Jan 16 08:59:20.953458 systemd-resolved[1330]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. Jan 16 08:59:21.524456 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1075906226.mount: Deactivated successfully. Jan 16 08:59:23.941685 containerd[1466]: time="2025-01-16T08:59:23.940384195Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:23.941685 containerd[1466]: time="2025-01-16T08:59:23.940924150Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=57238571" Jan 16 08:59:23.947267 containerd[1466]: time="2025-01-16T08:59:23.947209157Z" level=info msg="ImageCreate event name:\"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:23.948949 containerd[1466]: time="2025-01-16T08:59:23.948896751Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"57236178\" in 2.998375789s" Jan 16 08:59:23.949143 containerd[1466]: time="2025-01-16T08:59:23.949123688Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" Jan 16 08:59:23.950442 containerd[1466]: time="2025-01-16T08:59:23.950374584Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:27.689563 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:59:27.703700 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 16 08:59:27.754069 systemd[1]: Reloading requested from client PID 2065 ('systemctl') (unit session-5.scope)... Jan 16 08:59:27.754092 systemd[1]: Reloading... Jan 16 08:59:27.935745 zram_generator::config[2107]: No configuration found. Jan 16 08:59:28.155874 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 16 08:59:28.295588 systemd[1]: Reloading finished in 540 ms. Jan 16 08:59:28.405298 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 16 08:59:28.405526 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 16 08:59:28.406610 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:59:28.425168 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 16 08:59:28.635604 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:59:28.652859 (kubelet)[2159]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 16 08:59:28.757818 kubelet[2159]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 16 08:59:28.757818 kubelet[2159]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 16 08:59:28.757818 kubelet[2159]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 16 08:59:28.759738 kubelet[2159]: I0116 08:59:28.759529 2159 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 16 08:59:29.461710 kubelet[2159]: I0116 08:59:29.461212 2159 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Jan 16 08:59:29.461710 kubelet[2159]: I0116 08:59:29.461256 2159 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 16 08:59:29.461710 kubelet[2159]: I0116 08:59:29.461523 2159 server.go:927] "Client rotation is on, will bootstrap in background" Jan 16 08:59:29.491057 kubelet[2159]: I0116 08:59:29.490975 2159 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 16 08:59:29.492647 kubelet[2159]: E0116 08:59:29.492546 2159 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://147.182.199.56:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.514737 kubelet[2159]: I0116 08:59:29.514685 2159 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 16 08:59:29.515120 kubelet[2159]: I0116 08:59:29.515068 2159 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 16 08:59:29.515405 kubelet[2159]: I0116 08:59:29.515118 2159 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.0-f-8a831eefc3","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jan 16 08:59:29.515562 kubelet[2159]: I0116 08:59:29.515425 2159 topology_manager.go:138] "Creating topology manager with none policy" Jan 16 08:59:29.515562 kubelet[2159]: I0116 08:59:29.515479 2159 container_manager_linux.go:301] "Creating device plugin manager" Jan 16 08:59:29.515793 kubelet[2159]: I0116 08:59:29.515760 2159 state_mem.go:36] "Initialized new in-memory state store" Jan 16 08:59:29.516946 kubelet[2159]: I0116 08:59:29.516913 2159 kubelet.go:400] "Attempting to sync node with API server" Jan 16 08:59:29.516946 kubelet[2159]: I0116 08:59:29.516951 2159 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 16 08:59:29.517728 kubelet[2159]: I0116 08:59:29.516989 2159 kubelet.go:312] "Adding apiserver pod source" Jan 16 08:59:29.517728 kubelet[2159]: I0116 08:59:29.517011 2159 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 16 08:59:29.526784 kubelet[2159]: I0116 08:59:29.526470 2159 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 16 08:59:29.529333 kubelet[2159]: I0116 08:59:29.528725 2159 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 16 08:59:29.529333 kubelet[2159]: W0116 08:59:29.528873 2159 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 16 08:59:29.531681 kubelet[2159]: I0116 08:59:29.531086 2159 server.go:1264] "Started kubelet" Jan 16 08:59:29.531681 kubelet[2159]: W0116 08:59:29.531384 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://147.182.199.56:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.0-f-8a831eefc3&limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.531681 kubelet[2159]: E0116 08:59:29.531478 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://147.182.199.56:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.0-f-8a831eefc3&limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.535273 kubelet[2159]: I0116 08:59:29.535197 2159 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 16 08:59:29.543874 kubelet[2159]: I0116 08:59:29.543640 2159 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 16 08:59:29.546290 kubelet[2159]: I0116 08:59:29.544901 2159 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 16 08:59:29.546290 kubelet[2159]: I0116 08:59:29.545627 2159 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 16 08:59:29.547906 kubelet[2159]: E0116 08:59:29.546744 2159 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://147.182.199.56:6443/api/v1/namespaces/default/events\": dial tcp 147.182.199.56:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.0-f-8a831eefc3.181b20a3e25dffcd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.0-f-8a831eefc3,UID:ci-4081.3.0-f-8a831eefc3,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.0-f-8a831eefc3,},FirstTimestamp:2025-01-16 08:59:29.531027405 +0000 UTC m=+0.868845907,LastTimestamp:2025-01-16 08:59:29.531027405 +0000 UTC m=+0.868845907,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.0-f-8a831eefc3,}" Jan 16 08:59:29.547906 kubelet[2159]: W0116 08:59:29.547613 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://147.182.199.56:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.547906 kubelet[2159]: E0116 08:59:29.547730 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://147.182.199.56:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.551669 kubelet[2159]: I0116 08:59:29.551588 2159 server.go:455] "Adding debug handlers to kubelet server" Jan 16 08:59:29.558559 kubelet[2159]: I0116 08:59:29.558095 2159 volume_manager.go:291] "Starting Kubelet Volume Manager" Jan 16 08:59:29.562855 kubelet[2159]: I0116 08:59:29.561612 2159 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 16 08:59:29.562855 kubelet[2159]: I0116 08:59:29.561737 2159 reconciler.go:26] "Reconciler: start to sync state" Jan 16 08:59:29.562855 kubelet[2159]: E0116 08:59:29.562266 2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.182.199.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.0-f-8a831eefc3?timeout=10s\": dial tcp 147.182.199.56:6443: connect: connection refused" interval="200ms" Jan 16 08:59:29.562855 kubelet[2159]: W0116 08:59:29.562750 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://147.182.199.56:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.562855 kubelet[2159]: E0116 08:59:29.562809 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://147.182.199.56:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.563923 kubelet[2159]: E0116 08:59:29.563894 2159 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 16 08:59:29.564326 kubelet[2159]: I0116 08:59:29.564309 2159 factory.go:221] Registration of the systemd container factory successfully Jan 16 08:59:29.564577 kubelet[2159]: I0116 08:59:29.564557 2159 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 16 08:59:29.566400 kubelet[2159]: I0116 08:59:29.566359 2159 factory.go:221] Registration of the containerd container factory successfully Jan 16 08:59:29.582707 kubelet[2159]: I0116 08:59:29.582608 2159 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 16 08:59:29.587254 kubelet[2159]: I0116 08:59:29.584807 2159 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 16 08:59:29.587254 kubelet[2159]: I0116 08:59:29.584869 2159 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 16 08:59:29.587254 kubelet[2159]: I0116 08:59:29.584904 2159 kubelet.go:2337] "Starting kubelet main sync loop" Jan 16 08:59:29.587254 kubelet[2159]: E0116 08:59:29.584990 2159 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 16 08:59:29.593274 kubelet[2159]: W0116 08:59:29.593200 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://147.182.199.56:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.593548 kubelet[2159]: E0116 08:59:29.593523 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://147.182.199.56:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:29.603346 kubelet[2159]: I0116 08:59:29.603299 2159 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 16 08:59:29.603346 kubelet[2159]: I0116 08:59:29.603325 2159 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 16 08:59:29.603346 kubelet[2159]: I0116 08:59:29.603355 2159 state_mem.go:36] "Initialized new in-memory state store" Jan 16 08:59:29.606919 kubelet[2159]: I0116 08:59:29.606871 2159 policy_none.go:49] "None policy: Start" Jan 16 08:59:29.608264 kubelet[2159]: I0116 08:59:29.608171 2159 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 16 08:59:29.608264 kubelet[2159]: I0116 08:59:29.608225 2159 state_mem.go:35] "Initializing new in-memory state store" Jan 16 08:59:29.621036 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 16 08:59:29.633457 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 16 08:59:29.641672 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 16 08:59:29.656319 kubelet[2159]: I0116 08:59:29.655626 2159 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 16 08:59:29.656524 kubelet[2159]: I0116 08:59:29.656329 2159 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 16 08:59:29.656524 kubelet[2159]: I0116 08:59:29.656515 2159 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 16 08:59:29.661424 kubelet[2159]: I0116 08:59:29.660221 2159 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.661424 kubelet[2159]: E0116 08:59:29.660831 2159 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.0-f-8a831eefc3\" not found" Jan 16 08:59:29.662660 kubelet[2159]: E0116 08:59:29.662448 2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://147.182.199.56:6443/api/v1/nodes\": dial tcp 147.182.199.56:6443: connect: connection refused" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.687065 kubelet[2159]: I0116 08:59:29.686159 2159 topology_manager.go:215] "Topology Admit Handler" podUID="aa52e7fe45d1a335d793e597a3ae7d2f" podNamespace="kube-system" podName="kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.688679 kubelet[2159]: I0116 08:59:29.688574 2159 topology_manager.go:215] "Topology Admit Handler" podUID="7bdd42b838f759a426d7389eb5da6227" podNamespace="kube-system" podName="kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.691368 kubelet[2159]: I0116 08:59:29.691291 2159 topology_manager.go:215] "Topology Admit Handler" podUID="a419fee806f483c3e88f2da31d48c9dc" podNamespace="kube-system" podName="kube-scheduler-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.701737 systemd[1]: Created slice kubepods-burstable-podaa52e7fe45d1a335d793e597a3ae7d2f.slice - libcontainer container kubepods-burstable-podaa52e7fe45d1a335d793e597a3ae7d2f.slice. Jan 16 08:59:29.725202 systemd[1]: Created slice kubepods-burstable-pod7bdd42b838f759a426d7389eb5da6227.slice - libcontainer container kubepods-burstable-pod7bdd42b838f759a426d7389eb5da6227.slice. Jan 16 08:59:29.743981 systemd[1]: Created slice kubepods-burstable-poda419fee806f483c3e88f2da31d48c9dc.slice - libcontainer container kubepods-burstable-poda419fee806f483c3e88f2da31d48c9dc.slice. Jan 16 08:59:29.763235 kubelet[2159]: E0116 08:59:29.763118 2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.182.199.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.0-f-8a831eefc3?timeout=10s\": dial tcp 147.182.199.56:6443: connect: connection refused" interval="400ms" Jan 16 08:59:29.764042 kubelet[2159]: I0116 08:59:29.763338 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aa52e7fe45d1a335d793e597a3ae7d2f-ca-certs\") pod \"kube-apiserver-ci-4081.3.0-f-8a831eefc3\" (UID: \"aa52e7fe45d1a335d793e597a3ae7d2f\") " pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.864192 kubelet[2159]: I0116 08:59:29.864133 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.864998 kubelet[2159]: I0116 08:59:29.864591 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.864998 kubelet[2159]: I0116 08:59:29.864677 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a419fee806f483c3e88f2da31d48c9dc-kubeconfig\") pod \"kube-scheduler-ci-4081.3.0-f-8a831eefc3\" (UID: \"a419fee806f483c3e88f2da31d48c9dc\") " pod="kube-system/kube-scheduler-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.864998 kubelet[2159]: I0116 08:59:29.864681 2159 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.864998 kubelet[2159]: I0116 08:59:29.864761 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aa52e7fe45d1a335d793e597a3ae7d2f-k8s-certs\") pod \"kube-apiserver-ci-4081.3.0-f-8a831eefc3\" (UID: \"aa52e7fe45d1a335d793e597a3ae7d2f\") " pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.864998 kubelet[2159]: I0116 08:59:29.864791 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aa52e7fe45d1a335d793e597a3ae7d2f-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.0-f-8a831eefc3\" (UID: \"aa52e7fe45d1a335d793e597a3ae7d2f\") " pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.865320 kubelet[2159]: I0116 08:59:29.864844 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.865320 kubelet[2159]: I0116 08:59:29.864871 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-ca-certs\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.865320 kubelet[2159]: I0116 08:59:29.864916 2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:29.865320 kubelet[2159]: E0116 08:59:29.865202 2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://147.182.199.56:6443/api/v1/nodes\": dial tcp 147.182.199.56:6443: connect: connection refused" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:30.021222 kubelet[2159]: E0116 08:59:30.021021 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:30.022418 containerd[1466]: time="2025-01-16T08:59:30.022178587Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.0-f-8a831eefc3,Uid:aa52e7fe45d1a335d793e597a3ae7d2f,Namespace:kube-system,Attempt:0,}" Jan 16 08:59:30.039230 kubelet[2159]: E0116 08:59:30.038765 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:30.040446 containerd[1466]: time="2025-01-16T08:59:30.039791407Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.0-f-8a831eefc3,Uid:7bdd42b838f759a426d7389eb5da6227,Namespace:kube-system,Attempt:0,}" Jan 16 08:59:30.049985 kubelet[2159]: E0116 08:59:30.049043 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:30.053433 containerd[1466]: time="2025-01-16T08:59:30.053198966Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.0-f-8a831eefc3,Uid:a419fee806f483c3e88f2da31d48c9dc,Namespace:kube-system,Attempt:0,}" Jan 16 08:59:30.164013 kubelet[2159]: E0116 08:59:30.163934 2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.182.199.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.0-f-8a831eefc3?timeout=10s\": dial tcp 147.182.199.56:6443: connect: connection refused" interval="800ms" Jan 16 08:59:30.273275 kubelet[2159]: I0116 08:59:30.272730 2159 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:30.274104 kubelet[2159]: E0116 08:59:30.274063 2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://147.182.199.56:6443/api/v1/nodes\": dial tcp 147.182.199.56:6443: connect: connection refused" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:30.399973 kubelet[2159]: W0116 08:59:30.399886 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://147.182.199.56:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:30.399973 kubelet[2159]: E0116 08:59:30.399945 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://147.182.199.56:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:30.590015 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3016339336.mount: Deactivated successfully. Jan 16 08:59:30.604292 containerd[1466]: time="2025-01-16T08:59:30.604222917Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 16 08:59:30.605559 containerd[1466]: time="2025-01-16T08:59:30.605506049Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 16 08:59:30.606721 containerd[1466]: time="2025-01-16T08:59:30.606417921Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 16 08:59:30.606945 containerd[1466]: time="2025-01-16T08:59:30.606904536Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Jan 16 08:59:30.607287 containerd[1466]: time="2025-01-16T08:59:30.607258491Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 16 08:59:30.609571 containerd[1466]: time="2025-01-16T08:59:30.609337854Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 16 08:59:30.609571 containerd[1466]: time="2025-01-16T08:59:30.609513389Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 16 08:59:30.619178 containerd[1466]: time="2025-01-16T08:59:30.619107080Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 16 08:59:30.620522 containerd[1466]: time="2025-01-16T08:59:30.620067690Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 580.160653ms" Jan 16 08:59:30.622053 containerd[1466]: time="2025-01-16T08:59:30.621998149Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 568.286368ms" Jan 16 08:59:30.625691 containerd[1466]: time="2025-01-16T08:59:30.625468025Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 603.15723ms" Jan 16 08:59:30.791425 kubelet[2159]: W0116 08:59:30.791290 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://147.182.199.56:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.0-f-8a831eefc3&limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:30.791425 kubelet[2159]: E0116 08:59:30.791384 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://147.182.199.56:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.0-f-8a831eefc3&limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:30.835717 kubelet[2159]: W0116 08:59:30.835417 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://147.182.199.56:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:30.835717 kubelet[2159]: E0116 08:59:30.835484 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://147.182.199.56:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:30.841262 containerd[1466]: time="2025-01-16T08:59:30.839815871Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 16 08:59:30.841262 containerd[1466]: time="2025-01-16T08:59:30.839928948Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 16 08:59:30.841262 containerd[1466]: time="2025-01-16T08:59:30.839969831Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:30.841262 containerd[1466]: time="2025-01-16T08:59:30.840176510Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:30.841947 containerd[1466]: time="2025-01-16T08:59:30.840820777Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 16 08:59:30.841947 containerd[1466]: time="2025-01-16T08:59:30.840903589Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 16 08:59:30.841947 containerd[1466]: time="2025-01-16T08:59:30.840936022Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:30.841947 containerd[1466]: time="2025-01-16T08:59:30.841064141Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:30.860422 containerd[1466]: time="2025-01-16T08:59:30.859969645Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 16 08:59:30.860422 containerd[1466]: time="2025-01-16T08:59:30.860059978Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 16 08:59:30.860422 containerd[1466]: time="2025-01-16T08:59:30.860108820Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:30.860422 containerd[1466]: time="2025-01-16T08:59:30.860273660Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:30.875949 systemd[1]: Started cri-containerd-cf23f7d621ff5874687fc9ce6e9041457739810d2bce462f05829ef448e57efd.scope - libcontainer container cf23f7d621ff5874687fc9ce6e9041457739810d2bce462f05829ef448e57efd. Jan 16 08:59:30.913949 systemd[1]: Started cri-containerd-041e7d3c3e8fe6b7097ec5d869665882592020141f72ac7f533c35171513db4e.scope - libcontainer container 041e7d3c3e8fe6b7097ec5d869665882592020141f72ac7f533c35171513db4e. Jan 16 08:59:30.941013 systemd[1]: Started cri-containerd-d7841b67b16c6b049074f34ba0071ebbbc41ba76169dbdd6b7c0633774f1f5bc.scope - libcontainer container d7841b67b16c6b049074f34ba0071ebbbc41ba76169dbdd6b7c0633774f1f5bc. Jan 16 08:59:30.968837 kubelet[2159]: E0116 08:59:30.967153 2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.182.199.56:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.0-f-8a831eefc3?timeout=10s\": dial tcp 147.182.199.56:6443: connect: connection refused" interval="1.6s" Jan 16 08:59:30.987130 containerd[1466]: time="2025-01-16T08:59:30.986872689Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.0-f-8a831eefc3,Uid:7bdd42b838f759a426d7389eb5da6227,Namespace:kube-system,Attempt:0,} returns sandbox id \"cf23f7d621ff5874687fc9ce6e9041457739810d2bce462f05829ef448e57efd\"" Jan 16 08:59:30.990811 kubelet[2159]: E0116 08:59:30.990721 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:31.005303 containerd[1466]: time="2025-01-16T08:59:31.005232332Z" level=info msg="CreateContainer within sandbox \"cf23f7d621ff5874687fc9ce6e9041457739810d2bce462f05829ef448e57efd\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 16 08:59:31.024032 containerd[1466]: time="2025-01-16T08:59:31.023943224Z" level=info msg="CreateContainer within sandbox \"cf23f7d621ff5874687fc9ce6e9041457739810d2bce462f05829ef448e57efd\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"86508142c7055a39aa253cfe2c0fb8e3e6766bdad20b46431842cbd386910ea8\"" Jan 16 08:59:31.027758 containerd[1466]: time="2025-01-16T08:59:31.027685944Z" level=info msg="StartContainer for \"86508142c7055a39aa253cfe2c0fb8e3e6766bdad20b46431842cbd386910ea8\"" Jan 16 08:59:31.073313 containerd[1466]: time="2025-01-16T08:59:31.073251602Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.0-f-8a831eefc3,Uid:aa52e7fe45d1a335d793e597a3ae7d2f,Namespace:kube-system,Attempt:0,} returns sandbox id \"d7841b67b16c6b049074f34ba0071ebbbc41ba76169dbdd6b7c0633774f1f5bc\"" Jan 16 08:59:31.075482 containerd[1466]: time="2025-01-16T08:59:31.075320671Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.0-f-8a831eefc3,Uid:a419fee806f483c3e88f2da31d48c9dc,Namespace:kube-system,Attempt:0,} returns sandbox id \"041e7d3c3e8fe6b7097ec5d869665882592020141f72ac7f533c35171513db4e\"" Jan 16 08:59:31.076825 kubelet[2159]: E0116 08:59:31.076780 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:31.077935 kubelet[2159]: I0116 08:59:31.077541 2159 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:31.077935 kubelet[2159]: E0116 08:59:31.077683 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:31.078103 kubelet[2159]: E0116 08:59:31.077982 2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://147.182.199.56:6443/api/v1/nodes\": dial tcp 147.182.199.56:6443: connect: connection refused" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:31.084561 containerd[1466]: time="2025-01-16T08:59:31.084357260Z" level=info msg="CreateContainer within sandbox \"041e7d3c3e8fe6b7097ec5d869665882592020141f72ac7f533c35171513db4e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 16 08:59:31.084814 containerd[1466]: time="2025-01-16T08:59:31.084358994Z" level=info msg="CreateContainer within sandbox \"d7841b67b16c6b049074f34ba0071ebbbc41ba76169dbdd6b7c0633774f1f5bc\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 16 08:59:31.103425 kubelet[2159]: W0116 08:59:31.101991 2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://147.182.199.56:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:31.103425 kubelet[2159]: E0116 08:59:31.102079 2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://147.182.199.56:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:31.106642 systemd[1]: Started cri-containerd-86508142c7055a39aa253cfe2c0fb8e3e6766bdad20b46431842cbd386910ea8.scope - libcontainer container 86508142c7055a39aa253cfe2c0fb8e3e6766bdad20b46431842cbd386910ea8. Jan 16 08:59:31.115165 containerd[1466]: time="2025-01-16T08:59:31.115064509Z" level=info msg="CreateContainer within sandbox \"041e7d3c3e8fe6b7097ec5d869665882592020141f72ac7f533c35171513db4e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"0c27a3e79c5208de6e407a1e7f283c639a0cf1f901bb5eb7cd5f0baadf0d21d4\"" Jan 16 08:59:31.116686 containerd[1466]: time="2025-01-16T08:59:31.116552358Z" level=info msg="StartContainer for \"0c27a3e79c5208de6e407a1e7f283c639a0cf1f901bb5eb7cd5f0baadf0d21d4\"" Jan 16 08:59:31.120969 containerd[1466]: time="2025-01-16T08:59:31.120294236Z" level=info msg="CreateContainer within sandbox \"d7841b67b16c6b049074f34ba0071ebbbc41ba76169dbdd6b7c0633774f1f5bc\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0f9722deebafd1904d285b5fc184e038370a8380a00d0df5799bba7a3b55a0f5\"" Jan 16 08:59:31.123497 containerd[1466]: time="2025-01-16T08:59:31.123445752Z" level=info msg="StartContainer for \"0f9722deebafd1904d285b5fc184e038370a8380a00d0df5799bba7a3b55a0f5\"" Jan 16 08:59:31.186230 systemd[1]: Started cri-containerd-0c27a3e79c5208de6e407a1e7f283c639a0cf1f901bb5eb7cd5f0baadf0d21d4.scope - libcontainer container 0c27a3e79c5208de6e407a1e7f283c639a0cf1f901bb5eb7cd5f0baadf0d21d4. Jan 16 08:59:31.211961 systemd[1]: Started cri-containerd-0f9722deebafd1904d285b5fc184e038370a8380a00d0df5799bba7a3b55a0f5.scope - libcontainer container 0f9722deebafd1904d285b5fc184e038370a8380a00d0df5799bba7a3b55a0f5. Jan 16 08:59:31.221752 containerd[1466]: time="2025-01-16T08:59:31.221638191Z" level=info msg="StartContainer for \"86508142c7055a39aa253cfe2c0fb8e3e6766bdad20b46431842cbd386910ea8\" returns successfully" Jan 16 08:59:31.301372 containerd[1466]: time="2025-01-16T08:59:31.301307365Z" level=info msg="StartContainer for \"0f9722deebafd1904d285b5fc184e038370a8380a00d0df5799bba7a3b55a0f5\" returns successfully" Jan 16 08:59:31.340711 containerd[1466]: time="2025-01-16T08:59:31.340637991Z" level=info msg="StartContainer for \"0c27a3e79c5208de6e407a1e7f283c639a0cf1f901bb5eb7cd5f0baadf0d21d4\" returns successfully" Jan 16 08:59:31.616092 kubelet[2159]: E0116 08:59:31.614790 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:31.625219 kubelet[2159]: E0116 08:59:31.623909 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:31.626768 kubelet[2159]: E0116 08:59:31.625856 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:31.628184 kubelet[2159]: E0116 08:59:31.628140 2159 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://147.182.199.56:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 147.182.199.56:6443: connect: connection refused Jan 16 08:59:32.631139 kubelet[2159]: E0116 08:59:32.631016 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:32.682710 kubelet[2159]: I0116 08:59:32.681734 2159 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:34.362139 kubelet[2159]: E0116 08:59:34.362050 2159 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.0-f-8a831eefc3\" not found" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:34.467825 kubelet[2159]: I0116 08:59:34.467415 2159 kubelet_node_status.go:76] "Successfully registered node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:34.524857 kubelet[2159]: I0116 08:59:34.524354 2159 apiserver.go:52] "Watching apiserver" Jan 16 08:59:34.562697 kubelet[2159]: I0116 08:59:34.562616 2159 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Jan 16 08:59:35.678436 kubelet[2159]: W0116 08:59:35.678372 2159 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 16 08:59:35.681279 kubelet[2159]: E0116 08:59:35.681065 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:36.652667 kubelet[2159]: E0116 08:59:36.652493 2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:37.104754 systemd[1]: Reloading requested from client PID 2433 ('systemctl') (unit session-5.scope)... Jan 16 08:59:37.105092 systemd[1]: Reloading... Jan 16 08:59:37.314222 zram_generator::config[2475]: No configuration found. Jan 16 08:59:37.559675 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 16 08:59:37.708997 systemd[1]: Reloading finished in 602 ms. Jan 16 08:59:37.791771 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 16 08:59:37.793069 kubelet[2159]: E0116 08:59:37.792807 2159 event.go:319] "Unable to write event (broadcaster is shut down)" event="&Event{ObjectMeta:{ci-4081.3.0-f-8a831eefc3.181b20a3e25dffcd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.0-f-8a831eefc3,UID:ci-4081.3.0-f-8a831eefc3,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.0-f-8a831eefc3,},FirstTimestamp:2025-01-16 08:59:29.531027405 +0000 UTC m=+0.868845907,LastTimestamp:2025-01-16 08:59:29.531027405 +0000 UTC m=+0.868845907,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.0-f-8a831eefc3,}" Jan 16 08:59:37.812531 systemd[1]: kubelet.service: Deactivated successfully. Jan 16 08:59:37.813195 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:59:37.813519 systemd[1]: kubelet.service: Consumed 1.463s CPU time, 113.2M memory peak, 0B memory swap peak. Jan 16 08:59:37.827273 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 16 08:59:38.123073 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 16 08:59:38.135621 (kubelet)[2523]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 16 08:59:38.294684 kubelet[2523]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 16 08:59:38.295623 kubelet[2523]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 16 08:59:38.295623 kubelet[2523]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 16 08:59:38.298445 kubelet[2523]: I0116 08:59:38.297837 2523 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 16 08:59:38.314881 kubelet[2523]: I0116 08:59:38.312615 2523 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Jan 16 08:59:38.314881 kubelet[2523]: I0116 08:59:38.312680 2523 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 16 08:59:38.315198 kubelet[2523]: I0116 08:59:38.315071 2523 server.go:927] "Client rotation is on, will bootstrap in background" Jan 16 08:59:38.331718 kubelet[2523]: I0116 08:59:38.331625 2523 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 16 08:59:38.341195 kubelet[2523]: I0116 08:59:38.340429 2523 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 16 08:59:38.389804 kubelet[2523]: I0116 08:59:38.386894 2523 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 16 08:59:38.389804 kubelet[2523]: I0116 08:59:38.387243 2523 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 16 08:59:38.389804 kubelet[2523]: I0116 08:59:38.387291 2523 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.0-f-8a831eefc3","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jan 16 08:59:38.389804 kubelet[2523]: I0116 08:59:38.387511 2523 topology_manager.go:138] "Creating topology manager with none policy" Jan 16 08:59:38.390284 kubelet[2523]: I0116 08:59:38.387524 2523 container_manager_linux.go:301] "Creating device plugin manager" Jan 16 08:59:38.390284 kubelet[2523]: I0116 08:59:38.388469 2523 state_mem.go:36] "Initialized new in-memory state store" Jan 16 08:59:38.398626 kubelet[2523]: I0116 08:59:38.398577 2523 kubelet.go:400] "Attempting to sync node with API server" Jan 16 08:59:38.400777 kubelet[2523]: I0116 08:59:38.400705 2523 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 16 08:59:38.400970 kubelet[2523]: I0116 08:59:38.400839 2523 kubelet.go:312] "Adding apiserver pod source" Jan 16 08:59:38.400970 kubelet[2523]: I0116 08:59:38.400873 2523 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 16 08:59:38.410401 kubelet[2523]: I0116 08:59:38.409211 2523 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 16 08:59:38.412088 kubelet[2523]: I0116 08:59:38.412027 2523 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 16 08:59:38.413162 kubelet[2523]: I0116 08:59:38.413110 2523 server.go:1264] "Started kubelet" Jan 16 08:59:38.423697 kubelet[2523]: I0116 08:59:38.420527 2523 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 16 08:59:38.434113 kubelet[2523]: I0116 08:59:38.434014 2523 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 16 08:59:38.439595 kubelet[2523]: I0116 08:59:38.439419 2523 server.go:455] "Adding debug handlers to kubelet server" Jan 16 08:59:38.443877 kubelet[2523]: I0116 08:59:38.440817 2523 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 16 08:59:38.443877 kubelet[2523]: I0116 08:59:38.441221 2523 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 16 08:59:38.444157 kubelet[2523]: I0116 08:59:38.444077 2523 volume_manager.go:291] "Starting Kubelet Volume Manager" Jan 16 08:59:38.444927 kubelet[2523]: I0116 08:59:38.444769 2523 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 16 08:59:38.454140 kubelet[2523]: I0116 08:59:38.445885 2523 reconciler.go:26] "Reconciler: start to sync state" Jan 16 08:59:38.461853 kubelet[2523]: I0116 08:59:38.458193 2523 factory.go:221] Registration of the systemd container factory successfully Jan 16 08:59:38.461853 kubelet[2523]: I0116 08:59:38.458402 2523 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 16 08:59:38.492595 kubelet[2523]: I0116 08:59:38.486459 2523 factory.go:221] Registration of the containerd container factory successfully Jan 16 08:59:38.506414 kubelet[2523]: E0116 08:59:38.506368 2523 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 16 08:59:38.540353 kubelet[2523]: I0116 08:59:38.540278 2523 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 16 08:59:38.553094 kubelet[2523]: I0116 08:59:38.553029 2523 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 16 08:59:38.553338 kubelet[2523]: I0116 08:59:38.553319 2523 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 16 08:59:38.553491 kubelet[2523]: I0116 08:59:38.553475 2523 kubelet.go:2337] "Starting kubelet main sync loop" Jan 16 08:59:38.555694 kubelet[2523]: E0116 08:59:38.554564 2523 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 16 08:59:38.587131 kubelet[2523]: I0116 08:59:38.587080 2523 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.632902 kubelet[2523]: I0116 08:59:38.631381 2523 kubelet_node_status.go:112] "Node was previously registered" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.633113 kubelet[2523]: I0116 08:59:38.632970 2523 kubelet_node_status.go:76] "Successfully registered node" node="ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.656591 kubelet[2523]: E0116 08:59:38.654714 2523 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 16 08:59:38.691510 kubelet[2523]: I0116 08:59:38.690782 2523 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 16 08:59:38.691510 kubelet[2523]: I0116 08:59:38.690902 2523 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 16 08:59:38.691510 kubelet[2523]: I0116 08:59:38.690945 2523 state_mem.go:36] "Initialized new in-memory state store" Jan 16 08:59:38.691510 kubelet[2523]: I0116 08:59:38.691270 2523 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 16 08:59:38.691510 kubelet[2523]: I0116 08:59:38.691291 2523 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 16 08:59:38.691510 kubelet[2523]: I0116 08:59:38.691341 2523 policy_none.go:49] "None policy: Start" Jan 16 08:59:38.697413 kubelet[2523]: I0116 08:59:38.696541 2523 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 16 08:59:38.699145 kubelet[2523]: I0116 08:59:38.697935 2523 state_mem.go:35] "Initializing new in-memory state store" Jan 16 08:59:38.699145 kubelet[2523]: I0116 08:59:38.698284 2523 state_mem.go:75] "Updated machine memory state" Jan 16 08:59:38.722680 kubelet[2523]: I0116 08:59:38.721979 2523 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 16 08:59:38.725553 kubelet[2523]: I0116 08:59:38.725474 2523 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 16 08:59:38.727013 kubelet[2523]: I0116 08:59:38.726247 2523 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 16 08:59:38.858689 kubelet[2523]: I0116 08:59:38.856688 2523 topology_manager.go:215] "Topology Admit Handler" podUID="aa52e7fe45d1a335d793e597a3ae7d2f" podNamespace="kube-system" podName="kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.858689 kubelet[2523]: I0116 08:59:38.856890 2523 topology_manager.go:215] "Topology Admit Handler" podUID="7bdd42b838f759a426d7389eb5da6227" podNamespace="kube-system" podName="kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.858689 kubelet[2523]: I0116 08:59:38.856982 2523 topology_manager.go:215] "Topology Admit Handler" podUID="a419fee806f483c3e88f2da31d48c9dc" podNamespace="kube-system" podName="kube-scheduler-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.869633 kubelet[2523]: W0116 08:59:38.869584 2523 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 16 08:59:38.880792 kubelet[2523]: W0116 08:59:38.880458 2523 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 16 08:59:38.881799 kubelet[2523]: E0116 08:59:38.881352 2523 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" already exists" pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.881945 kubelet[2523]: W0116 08:59:38.881879 2523 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 16 08:59:38.958366 kubelet[2523]: I0116 08:59:38.955956 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aa52e7fe45d1a335d793e597a3ae7d2f-ca-certs\") pod \"kube-apiserver-ci-4081.3.0-f-8a831eefc3\" (UID: \"aa52e7fe45d1a335d793e597a3ae7d2f\") " pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.958366 kubelet[2523]: I0116 08:59:38.956132 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aa52e7fe45d1a335d793e597a3ae7d2f-k8s-certs\") pod \"kube-apiserver-ci-4081.3.0-f-8a831eefc3\" (UID: \"aa52e7fe45d1a335d793e597a3ae7d2f\") " pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.958366 kubelet[2523]: I0116 08:59:38.956193 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-ca-certs\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.958366 kubelet[2523]: I0116 08:59:38.956218 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.958366 kubelet[2523]: I0116 08:59:38.956293 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.959950 kubelet[2523]: I0116 08:59:38.956445 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a419fee806f483c3e88f2da31d48c9dc-kubeconfig\") pod \"kube-scheduler-ci-4081.3.0-f-8a831eefc3\" (UID: \"a419fee806f483c3e88f2da31d48c9dc\") " pod="kube-system/kube-scheduler-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.959950 kubelet[2523]: I0116 08:59:38.957957 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aa52e7fe45d1a335d793e597a3ae7d2f-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.0-f-8a831eefc3\" (UID: \"aa52e7fe45d1a335d793e597a3ae7d2f\") " pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.959950 kubelet[2523]: I0116 08:59:38.958094 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:38.959950 kubelet[2523]: I0116 08:59:38.958124 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7bdd42b838f759a426d7389eb5da6227-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.0-f-8a831eefc3\" (UID: \"7bdd42b838f759a426d7389eb5da6227\") " pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:39.173218 kubelet[2523]: E0116 08:59:39.173143 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:39.183822 kubelet[2523]: E0116 08:59:39.182891 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:39.187867 kubelet[2523]: E0116 08:59:39.187824 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:39.426399 kubelet[2523]: I0116 08:59:39.425987 2523 apiserver.go:52] "Watching apiserver" Jan 16 08:59:39.453968 kubelet[2523]: I0116 08:59:39.453201 2523 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Jan 16 08:59:39.614037 kubelet[2523]: E0116 08:59:39.613701 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:39.627181 kubelet[2523]: W0116 08:59:39.627129 2523 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 16 08:59:39.629684 kubelet[2523]: E0116 08:59:39.627515 2523 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4081.3.0-f-8a831eefc3\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:39.631437 kubelet[2523]: E0116 08:59:39.630485 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:39.631778 kubelet[2523]: W0116 08:59:39.631640 2523 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 16 08:59:39.631778 kubelet[2523]: E0116 08:59:39.631754 2523 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081.3.0-f-8a831eefc3\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" Jan 16 08:59:39.632535 kubelet[2523]: E0116 08:59:39.632494 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:39.698031 kubelet[2523]: I0116 08:59:39.696943 2523 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.0-f-8a831eefc3" podStartSLOduration=1.696921361 podStartE2EDuration="1.696921361s" podCreationTimestamp="2025-01-16 08:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-16 08:59:39.661832128 +0000 UTC m=+1.503531049" watchObservedRunningTime="2025-01-16 08:59:39.696921361 +0000 UTC m=+1.538620269" Jan 16 08:59:39.722482 kubelet[2523]: I0116 08:59:39.722391 2523 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.0-f-8a831eefc3" podStartSLOduration=4.722368138 podStartE2EDuration="4.722368138s" podCreationTimestamp="2025-01-16 08:59:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-16 08:59:39.697108542 +0000 UTC m=+1.538807430" watchObservedRunningTime="2025-01-16 08:59:39.722368138 +0000 UTC m=+1.564067049" Jan 16 08:59:40.194118 update_engine[1452]: I20250116 08:59:40.192701 1452 update_attempter.cc:509] Updating boot flags... Jan 16 08:59:40.195403 sudo[1614]: pam_unix(sudo:session): session closed for user root Jan 16 08:59:40.205135 sshd[1611]: pam_unix(sshd:session): session closed for user core Jan 16 08:59:40.216450 systemd[1]: sshd@4-147.182.199.56:22-139.178.68.195:33698.service: Deactivated successfully. Jan 16 08:59:40.216761 systemd-logind[1451]: Session 5 logged out. Waiting for processes to exit. Jan 16 08:59:40.225031 systemd[1]: session-5.scope: Deactivated successfully. Jan 16 08:59:40.225962 systemd[1]: session-5.scope: Consumed 5.916s CPU time, 191.0M memory peak, 0B memory swap peak. Jan 16 08:59:40.237024 systemd-logind[1451]: Removed session 5. Jan 16 08:59:40.287207 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2589) Jan 16 08:59:40.353734 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 37 scanned by (udev-worker) (2588) Jan 16 08:59:40.614709 kubelet[2523]: E0116 08:59:40.614511 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:40.617341 kubelet[2523]: E0116 08:59:40.617174 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:41.616949 kubelet[2523]: E0116 08:59:41.616819 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:47.164336 kubelet[2523]: E0116 08:59:47.162671 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:47.185070 kubelet[2523]: I0116 08:59:47.184982 2523 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.0-f-8a831eefc3" podStartSLOduration=9.184936999 podStartE2EDuration="9.184936999s" podCreationTimestamp="2025-01-16 08:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-16 08:59:39.722626465 +0000 UTC m=+1.564325364" watchObservedRunningTime="2025-01-16 08:59:47.184936999 +0000 UTC m=+9.026635898" Jan 16 08:59:47.636287 kubelet[2523]: E0116 08:59:47.636237 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:49.432745 kubelet[2523]: E0116 08:59:49.432697 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:49.694688 kubelet[2523]: E0116 08:59:49.693824 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:51.300757 kubelet[2523]: I0116 08:59:51.299711 2523 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 16 08:59:51.303005 containerd[1466]: time="2025-01-16T08:59:51.301684509Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 16 08:59:51.303548 kubelet[2523]: I0116 08:59:51.302033 2523 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 16 08:59:51.535489 kubelet[2523]: I0116 08:59:51.534866 2523 topology_manager.go:215] "Topology Admit Handler" podUID="2144d981-6b7b-423b-9c6c-319e05ff398d" podNamespace="kube-system" podName="kube-proxy-xcqtq" Jan 16 08:59:51.546145 kubelet[2523]: I0116 08:59:51.546033 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2144d981-6b7b-423b-9c6c-319e05ff398d-lib-modules\") pod \"kube-proxy-xcqtq\" (UID: \"2144d981-6b7b-423b-9c6c-319e05ff398d\") " pod="kube-system/kube-proxy-xcqtq" Jan 16 08:59:51.546145 kubelet[2523]: I0116 08:59:51.546087 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmj9v\" (UniqueName: \"kubernetes.io/projected/2144d981-6b7b-423b-9c6c-319e05ff398d-kube-api-access-rmj9v\") pod \"kube-proxy-xcqtq\" (UID: \"2144d981-6b7b-423b-9c6c-319e05ff398d\") " pod="kube-system/kube-proxy-xcqtq" Jan 16 08:59:51.546145 kubelet[2523]: I0116 08:59:51.546114 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/2144d981-6b7b-423b-9c6c-319e05ff398d-kube-proxy\") pod \"kube-proxy-xcqtq\" (UID: \"2144d981-6b7b-423b-9c6c-319e05ff398d\") " pod="kube-system/kube-proxy-xcqtq" Jan 16 08:59:51.546769 kubelet[2523]: I0116 08:59:51.546591 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2144d981-6b7b-423b-9c6c-319e05ff398d-xtables-lock\") pod \"kube-proxy-xcqtq\" (UID: \"2144d981-6b7b-423b-9c6c-319e05ff398d\") " pod="kube-system/kube-proxy-xcqtq" Jan 16 08:59:51.554205 systemd[1]: Created slice kubepods-besteffort-pod2144d981_6b7b_423b_9c6c_319e05ff398d.slice - libcontainer container kubepods-besteffort-pod2144d981_6b7b_423b_9c6c_319e05ff398d.slice. Jan 16 08:59:51.614121 kubelet[2523]: I0116 08:59:51.613221 2523 topology_manager.go:215] "Topology Admit Handler" podUID="3c63d751-d26c-4f74-8f5e-e23b51b03a5b" podNamespace="kube-flannel" podName="kube-flannel-ds-2pr6v" Jan 16 08:59:51.634356 systemd[1]: Created slice kubepods-burstable-pod3c63d751_d26c_4f74_8f5e_e23b51b03a5b.slice - libcontainer container kubepods-burstable-pod3c63d751_d26c_4f74_8f5e_e23b51b03a5b.slice. Jan 16 08:59:51.647428 kubelet[2523]: I0116 08:59:51.646986 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-run\") pod \"kube-flannel-ds-2pr6v\" (UID: \"3c63d751-d26c-4f74-8f5e-e23b51b03a5b\") " pod="kube-flannel/kube-flannel-ds-2pr6v" Jan 16 08:59:51.647428 kubelet[2523]: I0116 08:59:51.647055 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-xtables-lock\") pod \"kube-flannel-ds-2pr6v\" (UID: \"3c63d751-d26c-4f74-8f5e-e23b51b03a5b\") " pod="kube-flannel/kube-flannel-ds-2pr6v" Jan 16 08:59:51.647428 kubelet[2523]: I0116 08:59:51.647082 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-cni-plugin\") pod \"kube-flannel-ds-2pr6v\" (UID: \"3c63d751-d26c-4f74-8f5e-e23b51b03a5b\") " pod="kube-flannel/kube-flannel-ds-2pr6v" Jan 16 08:59:51.647428 kubelet[2523]: I0116 08:59:51.647113 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-flannel-cfg\") pod \"kube-flannel-ds-2pr6v\" (UID: \"3c63d751-d26c-4f74-8f5e-e23b51b03a5b\") " pod="kube-flannel/kube-flannel-ds-2pr6v" Jan 16 08:59:51.647428 kubelet[2523]: I0116 08:59:51.647146 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6d8f\" (UniqueName: \"kubernetes.io/projected/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-kube-api-access-b6d8f\") pod \"kube-flannel-ds-2pr6v\" (UID: \"3c63d751-d26c-4f74-8f5e-e23b51b03a5b\") " pod="kube-flannel/kube-flannel-ds-2pr6v" Jan 16 08:59:51.648262 kubelet[2523]: I0116 08:59:51.647223 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-cni\") pod \"kube-flannel-ds-2pr6v\" (UID: \"3c63d751-d26c-4f74-8f5e-e23b51b03a5b\") " pod="kube-flannel/kube-flannel-ds-2pr6v" Jan 16 08:59:51.659179 kubelet[2523]: W0116 08:59:51.659027 2523 reflector.go:547] object-"kube-flannel"/"kube-flannel-cfg": failed to list *v1.ConfigMap: configmaps "kube-flannel-cfg" is forbidden: User "system:node:ci-4081.3.0-f-8a831eefc3" cannot list resource "configmaps" in API group "" in the namespace "kube-flannel": no relationship found between node 'ci-4081.3.0-f-8a831eefc3' and this object Jan 16 08:59:51.659179 kubelet[2523]: E0116 08:59:51.659081 2523 reflector.go:150] object-"kube-flannel"/"kube-flannel-cfg": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-flannel-cfg" is forbidden: User "system:node:ci-4081.3.0-f-8a831eefc3" cannot list resource "configmaps" in API group "" in the namespace "kube-flannel": no relationship found between node 'ci-4081.3.0-f-8a831eefc3' and this object Jan 16 08:59:51.659179 kubelet[2523]: W0116 08:59:51.659122 2523 reflector.go:547] object-"kube-flannel"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4081.3.0-f-8a831eefc3" cannot list resource "configmaps" in API group "" in the namespace "kube-flannel": no relationship found between node 'ci-4081.3.0-f-8a831eefc3' and this object Jan 16 08:59:51.659179 kubelet[2523]: E0116 08:59:51.659136 2523 reflector.go:150] object-"kube-flannel"/"kube-root-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4081.3.0-f-8a831eefc3" cannot list resource "configmaps" in API group "" in the namespace "kube-flannel": no relationship found between node 'ci-4081.3.0-f-8a831eefc3' and this object Jan 16 08:59:51.868686 kubelet[2523]: E0116 08:59:51.868185 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:51.870404 containerd[1466]: time="2025-01-16T08:59:51.869745986Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xcqtq,Uid:2144d981-6b7b-423b-9c6c-319e05ff398d,Namespace:kube-system,Attempt:0,}" Jan 16 08:59:51.918130 containerd[1466]: time="2025-01-16T08:59:51.917126998Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 16 08:59:51.918840 containerd[1466]: time="2025-01-16T08:59:51.918760017Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 16 08:59:51.919023 containerd[1466]: time="2025-01-16T08:59:51.918991549Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:51.919327 containerd[1466]: time="2025-01-16T08:59:51.919277051Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:51.956012 systemd[1]: Started cri-containerd-4f73401aca48fb6d1ab1361cc4daad830ceebafc1a56871d005f133c0c3ba2d1.scope - libcontainer container 4f73401aca48fb6d1ab1361cc4daad830ceebafc1a56871d005f133c0c3ba2d1. Jan 16 08:59:51.993385 containerd[1466]: time="2025-01-16T08:59:51.993125539Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xcqtq,Uid:2144d981-6b7b-423b-9c6c-319e05ff398d,Namespace:kube-system,Attempt:0,} returns sandbox id \"4f73401aca48fb6d1ab1361cc4daad830ceebafc1a56871d005f133c0c3ba2d1\"" Jan 16 08:59:51.995706 kubelet[2523]: E0116 08:59:51.994377 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:51.998607 containerd[1466]: time="2025-01-16T08:59:51.998441746Z" level=info msg="CreateContainer within sandbox \"4f73401aca48fb6d1ab1361cc4daad830ceebafc1a56871d005f133c0c3ba2d1\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 16 08:59:52.019643 containerd[1466]: time="2025-01-16T08:59:52.019587606Z" level=info msg="CreateContainer within sandbox \"4f73401aca48fb6d1ab1361cc4daad830ceebafc1a56871d005f133c0c3ba2d1\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"5abd0df50d650f26a64477b9e272c5a4c2ac9bf47aeecceaf548e3cec4637125\"" Jan 16 08:59:52.022699 containerd[1466]: time="2025-01-16T08:59:52.020733467Z" level=info msg="StartContainer for \"5abd0df50d650f26a64477b9e272c5a4c2ac9bf47aeecceaf548e3cec4637125\"" Jan 16 08:59:52.074970 systemd[1]: Started cri-containerd-5abd0df50d650f26a64477b9e272c5a4c2ac9bf47aeecceaf548e3cec4637125.scope - libcontainer container 5abd0df50d650f26a64477b9e272c5a4c2ac9bf47aeecceaf548e3cec4637125. Jan 16 08:59:52.154395 containerd[1466]: time="2025-01-16T08:59:52.154261559Z" level=info msg="StartContainer for \"5abd0df50d650f26a64477b9e272c5a4c2ac9bf47aeecceaf548e3cec4637125\" returns successfully" Jan 16 08:59:52.665339 kubelet[2523]: E0116 08:59:52.664072 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:52.750273 kubelet[2523]: E0116 08:59:52.749888 2523 configmap.go:199] Couldn't get configMap kube-flannel/kube-flannel-cfg: failed to sync configmap cache: timed out waiting for the condition Jan 16 08:59:52.750273 kubelet[2523]: E0116 08:59:52.750038 2523 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-flannel-cfg podName:3c63d751-d26c-4f74-8f5e-e23b51b03a5b nodeName:}" failed. No retries permitted until 2025-01-16 08:59:53.250007822 +0000 UTC m=+15.091706728 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "flannel-cfg" (UniqueName: "kubernetes.io/configmap/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-flannel-cfg") pod "kube-flannel-ds-2pr6v" (UID: "3c63d751-d26c-4f74-8f5e-e23b51b03a5b") : failed to sync configmap cache: timed out waiting for the condition Jan 16 08:59:52.769718 kubelet[2523]: E0116 08:59:52.768465 2523 projected.go:294] Couldn't get configMap kube-flannel/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 16 08:59:52.769718 kubelet[2523]: E0116 08:59:52.768554 2523 projected.go:200] Error preparing data for projected volume kube-api-access-b6d8f for pod kube-flannel/kube-flannel-ds-2pr6v: failed to sync configmap cache: timed out waiting for the condition Jan 16 08:59:52.770078 kubelet[2523]: E0116 08:59:52.770035 2523 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-kube-api-access-b6d8f podName:3c63d751-d26c-4f74-8f5e-e23b51b03a5b nodeName:}" failed. No retries permitted until 2025-01-16 08:59:53.268718558 +0000 UTC m=+15.110417467 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-b6d8f" (UniqueName: "kubernetes.io/projected/3c63d751-d26c-4f74-8f5e-e23b51b03a5b-kube-api-access-b6d8f") pod "kube-flannel-ds-2pr6v" (UID: "3c63d751-d26c-4f74-8f5e-e23b51b03a5b") : failed to sync configmap cache: timed out waiting for the condition Jan 16 08:59:53.441438 kubelet[2523]: E0116 08:59:53.441376 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:53.444223 containerd[1466]: time="2025-01-16T08:59:53.444149809Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-2pr6v,Uid:3c63d751-d26c-4f74-8f5e-e23b51b03a5b,Namespace:kube-flannel,Attempt:0,}" Jan 16 08:59:53.488634 containerd[1466]: time="2025-01-16T08:59:53.487744392Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 16 08:59:53.488634 containerd[1466]: time="2025-01-16T08:59:53.488366241Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 16 08:59:53.488634 containerd[1466]: time="2025-01-16T08:59:53.488454365Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:53.490508 containerd[1466]: time="2025-01-16T08:59:53.490179107Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 08:59:53.530082 systemd[1]: Started cri-containerd-cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b.scope - libcontainer container cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b. Jan 16 08:59:53.604894 containerd[1466]: time="2025-01-16T08:59:53.604809159Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-2pr6v,Uid:3c63d751-d26c-4f74-8f5e-e23b51b03a5b,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b\"" Jan 16 08:59:53.607471 kubelet[2523]: E0116 08:59:53.607167 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:53.611918 containerd[1466]: time="2025-01-16T08:59:53.611851749Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\"" Jan 16 08:59:55.687510 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3738107954.mount: Deactivated successfully. Jan 16 08:59:55.757709 containerd[1466]: time="2025-01-16T08:59:55.756386144Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:55.757709 containerd[1466]: time="2025-01-16T08:59:55.757342200Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3852936" Jan 16 08:59:55.757709 containerd[1466]: time="2025-01-16T08:59:55.757575348Z" level=info msg="ImageCreate event name:\"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:55.761542 containerd[1466]: time="2025-01-16T08:59:55.761471325Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 08:59:55.762802 containerd[1466]: time="2025-01-16T08:59:55.762759445Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3842055\" in 2.150846088s" Jan 16 08:59:55.762972 containerd[1466]: time="2025-01-16T08:59:55.762955415Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\"" Jan 16 08:59:55.766815 containerd[1466]: time="2025-01-16T08:59:55.766756789Z" level=info msg="CreateContainer within sandbox \"cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Jan 16 08:59:55.783516 containerd[1466]: time="2025-01-16T08:59:55.781603716Z" level=info msg="CreateContainer within sandbox \"cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d\"" Jan 16 08:59:55.787199 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2797512896.mount: Deactivated successfully. Jan 16 08:59:55.788886 containerd[1466]: time="2025-01-16T08:59:55.788277248Z" level=info msg="StartContainer for \"c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d\"" Jan 16 08:59:55.835190 systemd[1]: Started cri-containerd-c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d.scope - libcontainer container c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d. Jan 16 08:59:55.882769 containerd[1466]: time="2025-01-16T08:59:55.882072045Z" level=info msg="StartContainer for \"c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d\" returns successfully" Jan 16 08:59:55.884351 systemd[1]: cri-containerd-c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d.scope: Deactivated successfully. Jan 16 08:59:55.935224 containerd[1466]: time="2025-01-16T08:59:55.934849191Z" level=info msg="shim disconnected" id=c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d namespace=k8s.io Jan 16 08:59:55.935224 containerd[1466]: time="2025-01-16T08:59:55.934947092Z" level=warning msg="cleaning up after shim disconnected" id=c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d namespace=k8s.io Jan 16 08:59:55.935224 containerd[1466]: time="2025-01-16T08:59:55.934963559Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 16 08:59:56.551309 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c6314e7945be5da7128aef6ab484d6c80dfb332d9498b68f94eefe79095c688d-rootfs.mount: Deactivated successfully. Jan 16 08:59:56.680108 kubelet[2523]: E0116 08:59:56.680019 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 08:59:56.685640 containerd[1466]: time="2025-01-16T08:59:56.685280670Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\"" Jan 16 08:59:56.701691 kubelet[2523]: I0116 08:59:56.700459 2523 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-xcqtq" podStartSLOduration=5.700429646 podStartE2EDuration="5.700429646s" podCreationTimestamp="2025-01-16 08:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-16 08:59:52.683639001 +0000 UTC m=+14.525337909" watchObservedRunningTime="2025-01-16 08:59:56.700429646 +0000 UTC m=+18.542128557" Jan 16 08:59:58.752328 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2276085057.mount: Deactivated successfully. Jan 16 09:00:02.045952 containerd[1466]: time="2025-01-16T09:00:02.043549239Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 09:00:02.049550 containerd[1466]: time="2025-01-16T09:00:02.049363557Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26866358" Jan 16 09:00:02.052679 containerd[1466]: time="2025-01-16T09:00:02.052041215Z" level=info msg="ImageCreate event name:\"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 09:00:02.070790 containerd[1466]: time="2025-01-16T09:00:02.063091656Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26855532\" in 5.377734236s" Jan 16 09:00:02.070790 containerd[1466]: time="2025-01-16T09:00:02.063179715Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\"" Jan 16 09:00:02.070790 containerd[1466]: time="2025-01-16T09:00:02.068054853Z" level=info msg="CreateContainer within sandbox \"cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 16 09:00:02.073143 containerd[1466]: time="2025-01-16T09:00:02.071581665Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 16 09:00:02.112333 containerd[1466]: time="2025-01-16T09:00:02.112213527Z" level=info msg="CreateContainer within sandbox \"cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125\"" Jan 16 09:00:02.129128 containerd[1466]: time="2025-01-16T09:00:02.129016224Z" level=info msg="StartContainer for \"345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125\"" Jan 16 09:00:02.277482 systemd[1]: Started cri-containerd-345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125.scope - libcontainer container 345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125. Jan 16 09:00:02.370983 systemd[1]: cri-containerd-345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125.scope: Deactivated successfully. Jan 16 09:00:02.376916 containerd[1466]: time="2025-01-16T09:00:02.376412503Z" level=info msg="StartContainer for \"345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125\" returns successfully" Jan 16 09:00:02.441724 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125-rootfs.mount: Deactivated successfully. Jan 16 09:00:02.453886 kubelet[2523]: I0116 09:00:02.453208 2523 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Jan 16 09:00:02.560070 containerd[1466]: time="2025-01-16T09:00:02.557748771Z" level=info msg="shim disconnected" id=345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125 namespace=k8s.io Jan 16 09:00:02.560070 containerd[1466]: time="2025-01-16T09:00:02.557870484Z" level=warning msg="cleaning up after shim disconnected" id=345766518667abe426ffb69e24c4a41591150d87ab5d3a0ea7e0217034a05125 namespace=k8s.io Jan 16 09:00:02.560070 containerd[1466]: time="2025-01-16T09:00:02.557885257Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 16 09:00:02.665779 kubelet[2523]: I0116 09:00:02.654640 2523 topology_manager.go:215] "Topology Admit Handler" podUID="461479dd-2bdf-49c0-8725-ff3c929deabe" podNamespace="kube-system" podName="coredns-7db6d8ff4d-78csp" Jan 16 09:00:02.665779 kubelet[2523]: I0116 09:00:02.654967 2523 topology_manager.go:215] "Topology Admit Handler" podUID="7a88a9e6-6e07-4b82-aaa3-21dd3ca30602" podNamespace="kube-system" podName="coredns-7db6d8ff4d-24pmf" Jan 16 09:00:02.687988 systemd[1]: Created slice kubepods-burstable-pod7a88a9e6_6e07_4b82_aaa3_21dd3ca30602.slice - libcontainer container kubepods-burstable-pod7a88a9e6_6e07_4b82_aaa3_21dd3ca30602.slice. Jan 16 09:00:02.702592 systemd[1]: Created slice kubepods-burstable-pod461479dd_2bdf_49c0_8725_ff3c929deabe.slice - libcontainer container kubepods-burstable-pod461479dd_2bdf_49c0_8725_ff3c929deabe.slice. Jan 16 09:00:02.738276 kubelet[2523]: E0116 09:00:02.730071 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:02.738464 containerd[1466]: time="2025-01-16T09:00:02.735683454Z" level=info msg="CreateContainer within sandbox \"cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Jan 16 09:00:02.799686 kubelet[2523]: I0116 09:00:02.797588 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmgn8\" (UniqueName: \"kubernetes.io/projected/461479dd-2bdf-49c0-8725-ff3c929deabe-kube-api-access-mmgn8\") pod \"coredns-7db6d8ff4d-78csp\" (UID: \"461479dd-2bdf-49c0-8725-ff3c929deabe\") " pod="kube-system/coredns-7db6d8ff4d-78csp" Jan 16 09:00:02.806140 kubelet[2523]: I0116 09:00:02.800895 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/461479dd-2bdf-49c0-8725-ff3c929deabe-config-volume\") pod \"coredns-7db6d8ff4d-78csp\" (UID: \"461479dd-2bdf-49c0-8725-ff3c929deabe\") " pod="kube-system/coredns-7db6d8ff4d-78csp" Jan 16 09:00:02.806140 kubelet[2523]: I0116 09:00:02.800955 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a88a9e6-6e07-4b82-aaa3-21dd3ca30602-config-volume\") pod \"coredns-7db6d8ff4d-24pmf\" (UID: \"7a88a9e6-6e07-4b82-aaa3-21dd3ca30602\") " pod="kube-system/coredns-7db6d8ff4d-24pmf" Jan 16 09:00:02.806140 kubelet[2523]: I0116 09:00:02.800991 2523 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8bpv\" (UniqueName: \"kubernetes.io/projected/7a88a9e6-6e07-4b82-aaa3-21dd3ca30602-kube-api-access-m8bpv\") pod \"coredns-7db6d8ff4d-24pmf\" (UID: \"7a88a9e6-6e07-4b82-aaa3-21dd3ca30602\") " pod="kube-system/coredns-7db6d8ff4d-24pmf" Jan 16 09:00:02.829322 containerd[1466]: time="2025-01-16T09:00:02.813986668Z" level=info msg="CreateContainer within sandbox \"cc741ae34e7535dbeb3867fa43648e1b825091fa554a3e02c84f553c2594339b\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"e0ba52b7c96096d5935d86d679d2e7f2e2ac72b40598672fe4090bfdcbe337bc\"" Jan 16 09:00:02.831390 containerd[1466]: time="2025-01-16T09:00:02.830528079Z" level=info msg="StartContainer for \"e0ba52b7c96096d5935d86d679d2e7f2e2ac72b40598672fe4090bfdcbe337bc\"" Jan 16 09:00:03.005178 kubelet[2523]: E0116 09:00:03.003154 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:03.013512 containerd[1466]: time="2025-01-16T09:00:03.008274114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-24pmf,Uid:7a88a9e6-6e07-4b82-aaa3-21dd3ca30602,Namespace:kube-system,Attempt:0,}" Jan 16 09:00:03.032991 kubelet[2523]: E0116 09:00:03.028043 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:03.033231 containerd[1466]: time="2025-01-16T09:00:03.031489168Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-78csp,Uid:461479dd-2bdf-49c0-8725-ff3c929deabe,Namespace:kube-system,Attempt:0,}" Jan 16 09:00:03.042104 systemd[1]: Started cri-containerd-e0ba52b7c96096d5935d86d679d2e7f2e2ac72b40598672fe4090bfdcbe337bc.scope - libcontainer container e0ba52b7c96096d5935d86d679d2e7f2e2ac72b40598672fe4090bfdcbe337bc. Jan 16 09:00:03.138212 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3913597524.mount: Deactivated successfully. Jan 16 09:00:03.233141 containerd[1466]: time="2025-01-16T09:00:03.229989795Z" level=info msg="StartContainer for \"e0ba52b7c96096d5935d86d679d2e7f2e2ac72b40598672fe4090bfdcbe337bc\" returns successfully" Jan 16 09:00:03.378162 systemd[1]: run-netns-cni\x2d87ca41d1\x2db2de\x2d4ad8\x2d98f0\x2d19dc0cd375a4.mount: Deactivated successfully. Jan 16 09:00:03.378422 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4035b8492cc19619e6a3b1c8bf32ab00c0ac97c1a8a1fe453c7087cb623927e3-shm.mount: Deactivated successfully. Jan 16 09:00:03.405970 systemd[1]: run-netns-cni\x2dafe79343\x2d595c\x2d16bb\x2d1c88\x2d1c652ed5aa1c.mount: Deactivated successfully. Jan 16 09:00:03.414734 containerd[1466]: time="2025-01-16T09:00:03.413332680Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-24pmf,Uid:7a88a9e6-6e07-4b82-aaa3-21dd3ca30602,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"4035b8492cc19619e6a3b1c8bf32ab00c0ac97c1a8a1fe453c7087cb623927e3\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 16 09:00:03.415942 kubelet[2523]: E0116 09:00:03.415192 2523 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4035b8492cc19619e6a3b1c8bf32ab00c0ac97c1a8a1fe453c7087cb623927e3\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 16 09:00:03.415942 kubelet[2523]: E0116 09:00:03.415303 2523 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4035b8492cc19619e6a3b1c8bf32ab00c0ac97c1a8a1fe453c7087cb623927e3\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-24pmf" Jan 16 09:00:03.415942 kubelet[2523]: E0116 09:00:03.415341 2523 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4035b8492cc19619e6a3b1c8bf32ab00c0ac97c1a8a1fe453c7087cb623927e3\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-24pmf" Jan 16 09:00:03.415942 kubelet[2523]: E0116 09:00:03.415400 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-24pmf_kube-system(7a88a9e6-6e07-4b82-aaa3-21dd3ca30602)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-24pmf_kube-system(7a88a9e6-6e07-4b82-aaa3-21dd3ca30602)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4035b8492cc19619e6a3b1c8bf32ab00c0ac97c1a8a1fe453c7087cb623927e3\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-7db6d8ff4d-24pmf" podUID="7a88a9e6-6e07-4b82-aaa3-21dd3ca30602" Jan 16 09:00:03.424697 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-886f28c717652914ba4f9b1399ad1c7125c9657a701f79c0235dfc7d91ed1f1b-shm.mount: Deactivated successfully. Jan 16 09:00:03.439695 containerd[1466]: time="2025-01-16T09:00:03.437327003Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-78csp,Uid:461479dd-2bdf-49c0-8725-ff3c929deabe,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"886f28c717652914ba4f9b1399ad1c7125c9657a701f79c0235dfc7d91ed1f1b\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 16 09:00:03.439923 kubelet[2523]: E0116 09:00:03.439315 2523 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"886f28c717652914ba4f9b1399ad1c7125c9657a701f79c0235dfc7d91ed1f1b\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 16 09:00:03.439923 kubelet[2523]: E0116 09:00:03.439408 2523 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"886f28c717652914ba4f9b1399ad1c7125c9657a701f79c0235dfc7d91ed1f1b\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-78csp" Jan 16 09:00:03.439923 kubelet[2523]: E0116 09:00:03.439439 2523 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"886f28c717652914ba4f9b1399ad1c7125c9657a701f79c0235dfc7d91ed1f1b\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-7db6d8ff4d-78csp" Jan 16 09:00:03.439923 kubelet[2523]: E0116 09:00:03.439507 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-78csp_kube-system(461479dd-2bdf-49c0-8725-ff3c929deabe)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-78csp_kube-system(461479dd-2bdf-49c0-8725-ff3c929deabe)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"886f28c717652914ba4f9b1399ad1c7125c9657a701f79c0235dfc7d91ed1f1b\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-7db6d8ff4d-78csp" podUID="461479dd-2bdf-49c0-8725-ff3c929deabe" Jan 16 09:00:03.746786 kubelet[2523]: E0116 09:00:03.745277 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:04.523193 systemd-networkd[1372]: flannel.1: Link UP Jan 16 09:00:04.523208 systemd-networkd[1372]: flannel.1: Gained carrier Jan 16 09:00:04.760578 kubelet[2523]: E0116 09:00:04.758020 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:05.917631 systemd-networkd[1372]: flannel.1: Gained IPv6LL Jan 16 09:00:16.557547 kubelet[2523]: E0116 09:00:16.555554 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:16.559983 containerd[1466]: time="2025-01-16T09:00:16.558762562Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-78csp,Uid:461479dd-2bdf-49c0-8725-ff3c929deabe,Namespace:kube-system,Attempt:0,}" Jan 16 09:00:16.676633 systemd-networkd[1372]: cni0: Link UP Jan 16 09:00:16.676663 systemd-networkd[1372]: cni0: Gained carrier Jan 16 09:00:16.702028 systemd-networkd[1372]: cni0: Lost carrier Jan 16 09:00:16.797363 systemd-networkd[1372]: veth4a5f8cda: Link UP Jan 16 09:00:16.802837 kernel: cni0: port 1(veth4a5f8cda) entered blocking state Jan 16 09:00:16.803005 kernel: cni0: port 1(veth4a5f8cda) entered disabled state Jan 16 09:00:16.803038 kernel: veth4a5f8cda: entered allmulticast mode Jan 16 09:00:16.804801 kernel: veth4a5f8cda: entered promiscuous mode Jan 16 09:00:16.820243 kernel: cni0: port 1(veth4a5f8cda) entered blocking state Jan 16 09:00:16.820468 kernel: cni0: port 1(veth4a5f8cda) entered forwarding state Jan 16 09:00:16.820514 kernel: cni0: port 1(veth4a5f8cda) entered disabled state Jan 16 09:00:16.830960 kernel: cni0: port 1(veth4a5f8cda) entered blocking state Jan 16 09:00:16.831186 kernel: cni0: port 1(veth4a5f8cda) entered forwarding state Jan 16 09:00:16.835437 systemd-networkd[1372]: veth4a5f8cda: Gained carrier Jan 16 09:00:16.838985 systemd-networkd[1372]: cni0: Gained carrier Jan 16 09:00:16.849718 containerd[1466]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc00001a938), "name":"cbr0", "type":"bridge"} Jan 16 09:00:16.849718 containerd[1466]: delegateAdd: netconf sent to delegate plugin: Jan 16 09:00:16.939825 containerd[1466]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2025-01-16T09:00:16.938774654Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 16 09:00:16.939825 containerd[1466]: time="2025-01-16T09:00:16.938889253Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 16 09:00:16.939825 containerd[1466]: time="2025-01-16T09:00:16.938909371Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 09:00:16.939825 containerd[1466]: time="2025-01-16T09:00:16.939117648Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 09:00:16.996457 systemd[1]: Started cri-containerd-647fa20b811f16f20f7618860894df65e33dadd90e98bb03db4bd05ea9d96cb4.scope - libcontainer container 647fa20b811f16f20f7618860894df65e33dadd90e98bb03db4bd05ea9d96cb4. Jan 16 09:00:17.144404 containerd[1466]: time="2025-01-16T09:00:17.144342658Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-78csp,Uid:461479dd-2bdf-49c0-8725-ff3c929deabe,Namespace:kube-system,Attempt:0,} returns sandbox id \"647fa20b811f16f20f7618860894df65e33dadd90e98bb03db4bd05ea9d96cb4\"" Jan 16 09:00:17.147948 kubelet[2523]: E0116 09:00:17.147899 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:17.160522 containerd[1466]: time="2025-01-16T09:00:17.160455574Z" level=info msg="CreateContainer within sandbox \"647fa20b811f16f20f7618860894df65e33dadd90e98bb03db4bd05ea9d96cb4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 16 09:00:17.203611 containerd[1466]: time="2025-01-16T09:00:17.203526416Z" level=info msg="CreateContainer within sandbox \"647fa20b811f16f20f7618860894df65e33dadd90e98bb03db4bd05ea9d96cb4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d9681c7bd9a29e9ac836da1f555ac1f3d999314e8f72d5aa744fd2bc56c973e1\"" Jan 16 09:00:17.206803 containerd[1466]: time="2025-01-16T09:00:17.204814462Z" level=info msg="StartContainer for \"d9681c7bd9a29e9ac836da1f555ac1f3d999314e8f72d5aa744fd2bc56c973e1\"" Jan 16 09:00:17.317023 systemd[1]: Started cri-containerd-d9681c7bd9a29e9ac836da1f555ac1f3d999314e8f72d5aa744fd2bc56c973e1.scope - libcontainer container d9681c7bd9a29e9ac836da1f555ac1f3d999314e8f72d5aa744fd2bc56c973e1. Jan 16 09:00:17.386408 containerd[1466]: time="2025-01-16T09:00:17.383499746Z" level=info msg="StartContainer for \"d9681c7bd9a29e9ac836da1f555ac1f3d999314e8f72d5aa744fd2bc56c973e1\" returns successfully" Jan 16 09:00:17.557786 kubelet[2523]: E0116 09:00:17.556420 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:17.560916 containerd[1466]: time="2025-01-16T09:00:17.560172975Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-24pmf,Uid:7a88a9e6-6e07-4b82-aaa3-21dd3ca30602,Namespace:kube-system,Attempt:0,}" Jan 16 09:00:17.622007 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1991242501.mount: Deactivated successfully. Jan 16 09:00:17.674316 systemd-networkd[1372]: veth7f21be31: Link UP Jan 16 09:00:17.679294 kernel: cni0: port 2(veth7f21be31) entered blocking state Jan 16 09:00:17.679436 kernel: cni0: port 2(veth7f21be31) entered disabled state Jan 16 09:00:17.689697 kernel: veth7f21be31: entered allmulticast mode Jan 16 09:00:17.703118 kernel: veth7f21be31: entered promiscuous mode Jan 16 09:00:17.706293 kernel: cni0: port 2(veth7f21be31) entered blocking state Jan 16 09:00:17.706473 kernel: cni0: port 2(veth7f21be31) entered forwarding state Jan 16 09:00:17.725157 kernel: cni0: port 2(veth7f21be31) entered disabled state Jan 16 09:00:17.744104 kernel: cni0: port 2(veth7f21be31) entered blocking state Jan 16 09:00:17.744291 kernel: cni0: port 2(veth7f21be31) entered forwarding state Jan 16 09:00:17.758702 systemd-networkd[1372]: veth7f21be31: Gained carrier Jan 16 09:00:17.760737 containerd[1466]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc0000a48e8), "name":"cbr0", "type":"bridge"} Jan 16 09:00:17.760737 containerd[1466]: delegateAdd: netconf sent to delegate plugin: Jan 16 09:00:17.839299 containerd[1466]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2025-01-16T09:00:17.838261270Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 16 09:00:17.839299 containerd[1466]: time="2025-01-16T09:00:17.838367129Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 16 09:00:17.839299 containerd[1466]: time="2025-01-16T09:00:17.838409496Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 09:00:17.839299 containerd[1466]: time="2025-01-16T09:00:17.838575202Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 16 09:00:17.914381 systemd[1]: Started cri-containerd-6a5b03f2098f8a47a38653b30c2bf3c1d22e3d803e05e4ea7e4357f3b28762a5.scope - libcontainer container 6a5b03f2098f8a47a38653b30c2bf3c1d22e3d803e05e4ea7e4357f3b28762a5. Jan 16 09:00:17.942813 kubelet[2523]: E0116 09:00:17.942747 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:18.048623 containerd[1466]: time="2025-01-16T09:00:18.048011077Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-24pmf,Uid:7a88a9e6-6e07-4b82-aaa3-21dd3ca30602,Namespace:kube-system,Attempt:0,} returns sandbox id \"6a5b03f2098f8a47a38653b30c2bf3c1d22e3d803e05e4ea7e4357f3b28762a5\"" Jan 16 09:00:18.053963 kubelet[2523]: E0116 09:00:18.053906 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:18.088029 containerd[1466]: time="2025-01-16T09:00:18.080757687Z" level=info msg="CreateContainer within sandbox \"6a5b03f2098f8a47a38653b30c2bf3c1d22e3d803e05e4ea7e4357f3b28762a5\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 16 09:00:18.134079 kubelet[2523]: I0116 09:00:18.133255 2523 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-2pr6v" podStartSLOduration=18.679175651 podStartE2EDuration="27.133226844s" podCreationTimestamp="2025-01-16 08:59:51 +0000 UTC" firstStartedPulling="2025-01-16 08:59:53.610396744 +0000 UTC m=+15.452095648" lastFinishedPulling="2025-01-16 09:00:02.064447937 +0000 UTC m=+23.906146841" observedRunningTime="2025-01-16 09:00:03.789888007 +0000 UTC m=+25.631586919" watchObservedRunningTime="2025-01-16 09:00:18.133226844 +0000 UTC m=+39.974925757" Jan 16 09:00:18.134079 kubelet[2523]: I0116 09:00:18.133688 2523 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-78csp" podStartSLOduration=27.133516041 podStartE2EDuration="27.133516041s" podCreationTimestamp="2025-01-16 08:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-16 09:00:18.133204627 +0000 UTC m=+39.974903537" watchObservedRunningTime="2025-01-16 09:00:18.133516041 +0000 UTC m=+39.975214952" Jan 16 09:00:18.162492 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount280818197.mount: Deactivated successfully. Jan 16 09:00:18.225116 containerd[1466]: time="2025-01-16T09:00:18.225032575Z" level=info msg="CreateContainer within sandbox \"6a5b03f2098f8a47a38653b30c2bf3c1d22e3d803e05e4ea7e4357f3b28762a5\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ee9537832a857ea91ee6af19f5f2e373ea7e197f3e8a05bfd22d16dc02ac4431\"" Jan 16 09:00:18.227135 containerd[1466]: time="2025-01-16T09:00:18.227063940Z" level=info msg="StartContainer for \"ee9537832a857ea91ee6af19f5f2e373ea7e197f3e8a05bfd22d16dc02ac4431\"" Jan 16 09:00:18.307782 systemd[1]: Started cri-containerd-ee9537832a857ea91ee6af19f5f2e373ea7e197f3e8a05bfd22d16dc02ac4431.scope - libcontainer container ee9537832a857ea91ee6af19f5f2e373ea7e197f3e8a05bfd22d16dc02ac4431. Jan 16 09:00:18.456842 containerd[1466]: time="2025-01-16T09:00:18.454746455Z" level=info msg="StartContainer for \"ee9537832a857ea91ee6af19f5f2e373ea7e197f3e8a05bfd22d16dc02ac4431\" returns successfully" Jan 16 09:00:18.521373 systemd-networkd[1372]: cni0: Gained IPv6LL Jan 16 09:00:18.587864 systemd-networkd[1372]: veth4a5f8cda: Gained IPv6LL Jan 16 09:00:18.900962 kubelet[2523]: E0116 09:00:18.900831 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:19.028867 kubelet[2523]: I0116 09:00:19.028764 2523 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-24pmf" podStartSLOduration=28.02872737 podStartE2EDuration="28.02872737s" podCreationTimestamp="2025-01-16 08:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-16 09:00:18.971176529 +0000 UTC m=+40.812875438" watchObservedRunningTime="2025-01-16 09:00:19.02872737 +0000 UTC m=+40.870426302" Jan 16 09:00:19.545168 systemd-networkd[1372]: veth7f21be31: Gained IPv6LL Jan 16 09:00:19.904363 kubelet[2523]: E0116 09:00:19.904291 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:20.907814 kubelet[2523]: E0116 09:00:20.906917 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:23.033950 kubelet[2523]: E0116 09:00:23.033493 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:23.925691 kubelet[2523]: E0116 09:00:23.925431 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:00:30.018798 systemd[1]: Started sshd@5-147.182.199.56:22-139.178.68.195:60468.service - OpenSSH per-connection server daemon (139.178.68.195:60468). Jan 16 09:00:30.147999 sshd[3471]: Accepted publickey for core from 139.178.68.195 port 60468 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:30.146335 sshd[3471]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:30.166507 systemd-logind[1451]: New session 6 of user core. Jan 16 09:00:30.182152 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 16 09:00:30.447967 sshd[3471]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:30.454802 systemd-logind[1451]: Session 6 logged out. Waiting for processes to exit. Jan 16 09:00:30.456564 systemd[1]: sshd@5-147.182.199.56:22-139.178.68.195:60468.service: Deactivated successfully. Jan 16 09:00:30.461745 systemd[1]: session-6.scope: Deactivated successfully. Jan 16 09:00:30.467448 systemd-logind[1451]: Removed session 6. Jan 16 09:00:35.483243 systemd[1]: Started sshd@6-147.182.199.56:22-139.178.68.195:50024.service - OpenSSH per-connection server daemon (139.178.68.195:50024). Jan 16 09:00:35.549907 sshd[3521]: Accepted publickey for core from 139.178.68.195 port 50024 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:35.552860 sshd[3521]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:35.584227 systemd-logind[1451]: New session 7 of user core. Jan 16 09:00:35.598531 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 16 09:00:35.849120 sshd[3521]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:35.858091 systemd[1]: sshd@6-147.182.199.56:22-139.178.68.195:50024.service: Deactivated successfully. Jan 16 09:00:35.867122 systemd[1]: session-7.scope: Deactivated successfully. Jan 16 09:00:35.871590 systemd-logind[1451]: Session 7 logged out. Waiting for processes to exit. Jan 16 09:00:35.874084 systemd-logind[1451]: Removed session 7. Jan 16 09:00:40.871331 systemd[1]: Started sshd@7-147.182.199.56:22-139.178.68.195:50038.service - OpenSSH per-connection server daemon (139.178.68.195:50038). Jan 16 09:00:40.946856 sshd[3558]: Accepted publickey for core from 139.178.68.195 port 50038 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:40.953480 sshd[3558]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:40.968869 systemd-logind[1451]: New session 8 of user core. Jan 16 09:00:40.973007 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 16 09:00:41.207095 sshd[3558]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:41.218902 systemd[1]: sshd@7-147.182.199.56:22-139.178.68.195:50038.service: Deactivated successfully. Jan 16 09:00:41.223918 systemd[1]: session-8.scope: Deactivated successfully. Jan 16 09:00:41.229145 systemd-logind[1451]: Session 8 logged out. Waiting for processes to exit. Jan 16 09:00:41.231269 systemd-logind[1451]: Removed session 8. Jan 16 09:00:46.238171 systemd[1]: Started sshd@8-147.182.199.56:22-139.178.68.195:57450.service - OpenSSH per-connection server daemon (139.178.68.195:57450). Jan 16 09:00:46.300705 sshd[3593]: Accepted publickey for core from 139.178.68.195 port 57450 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:46.303127 sshd[3593]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:46.334403 systemd-logind[1451]: New session 9 of user core. Jan 16 09:00:46.344529 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 16 09:00:46.542791 sshd[3593]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:46.561925 systemd[1]: sshd@8-147.182.199.56:22-139.178.68.195:57450.service: Deactivated successfully. Jan 16 09:00:46.569424 systemd[1]: session-9.scope: Deactivated successfully. Jan 16 09:00:46.571031 systemd-logind[1451]: Session 9 logged out. Waiting for processes to exit. Jan 16 09:00:46.590541 systemd[1]: Started sshd@9-147.182.199.56:22-139.178.68.195:57460.service - OpenSSH per-connection server daemon (139.178.68.195:57460). Jan 16 09:00:46.595239 systemd-logind[1451]: Removed session 9. Jan 16 09:00:46.651706 sshd[3607]: Accepted publickey for core from 139.178.68.195 port 57460 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:46.655187 sshd[3607]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:46.667194 systemd-logind[1451]: New session 10 of user core. Jan 16 09:00:46.675979 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 16 09:00:46.965850 sshd[3607]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:46.978503 systemd[1]: sshd@9-147.182.199.56:22-139.178.68.195:57460.service: Deactivated successfully. Jan 16 09:00:46.983420 systemd[1]: session-10.scope: Deactivated successfully. Jan 16 09:00:46.986063 systemd-logind[1451]: Session 10 logged out. Waiting for processes to exit. Jan 16 09:00:46.997854 systemd[1]: Started sshd@10-147.182.199.56:22-139.178.68.195:57468.service - OpenSSH per-connection server daemon (139.178.68.195:57468). Jan 16 09:00:47.003433 systemd-logind[1451]: Removed session 10. Jan 16 09:00:47.110962 sshd[3618]: Accepted publickey for core from 139.178.68.195 port 57468 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:47.114983 sshd[3618]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:47.126170 systemd-logind[1451]: New session 11 of user core. Jan 16 09:00:47.135859 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 16 09:00:47.361892 sshd[3618]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:47.372208 systemd[1]: sshd@10-147.182.199.56:22-139.178.68.195:57468.service: Deactivated successfully. Jan 16 09:00:47.375932 systemd[1]: session-11.scope: Deactivated successfully. Jan 16 09:00:47.378320 systemd-logind[1451]: Session 11 logged out. Waiting for processes to exit. Jan 16 09:00:47.381380 systemd-logind[1451]: Removed session 11. Jan 16 09:00:52.390031 systemd[1]: Started sshd@11-147.182.199.56:22-139.178.68.195:57472.service - OpenSSH per-connection server daemon (139.178.68.195:57472). Jan 16 09:00:52.474179 sshd[3655]: Accepted publickey for core from 139.178.68.195 port 57472 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:52.477954 sshd[3655]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:52.492232 systemd-logind[1451]: New session 12 of user core. Jan 16 09:00:52.495986 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 16 09:00:52.721730 sshd[3655]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:52.728011 systemd-logind[1451]: Session 12 logged out. Waiting for processes to exit. Jan 16 09:00:52.729420 systemd[1]: sshd@11-147.182.199.56:22-139.178.68.195:57472.service: Deactivated successfully. Jan 16 09:00:52.735750 systemd[1]: session-12.scope: Deactivated successfully. Jan 16 09:00:52.739734 systemd-logind[1451]: Removed session 12. Jan 16 09:00:54.507161 systemd[1]: Started sshd@12-147.182.199.56:22-198.235.24.208:52230.service - OpenSSH per-connection server daemon (198.235.24.208:52230). Jan 16 09:00:54.799723 sshd[3668]: Connection closed by 198.235.24.208 port 52230 Jan 16 09:00:54.801269 systemd[1]: sshd@12-147.182.199.56:22-198.235.24.208:52230.service: Deactivated successfully. Jan 16 09:00:57.771393 systemd[1]: Started sshd@13-147.182.199.56:22-139.178.68.195:42632.service - OpenSSH per-connection server daemon (139.178.68.195:42632). Jan 16 09:00:57.843750 sshd[3693]: Accepted publickey for core from 139.178.68.195 port 42632 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:57.846619 sshd[3693]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:57.854500 systemd-logind[1451]: New session 13 of user core. Jan 16 09:00:57.860019 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 16 09:00:58.097950 sshd[3693]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:58.117149 systemd[1]: sshd@13-147.182.199.56:22-139.178.68.195:42632.service: Deactivated successfully. Jan 16 09:00:58.123435 systemd[1]: session-13.scope: Deactivated successfully. Jan 16 09:00:58.127261 systemd-logind[1451]: Session 13 logged out. Waiting for processes to exit. Jan 16 09:00:58.141301 systemd[1]: Started sshd@14-147.182.199.56:22-139.178.68.195:42644.service - OpenSSH per-connection server daemon (139.178.68.195:42644). Jan 16 09:00:58.144446 systemd-logind[1451]: Removed session 13. Jan 16 09:00:58.207945 sshd[3705]: Accepted publickey for core from 139.178.68.195 port 42644 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:58.209257 sshd[3705]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:58.224447 systemd-logind[1451]: New session 14 of user core. Jan 16 09:00:58.231075 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 16 09:00:58.729302 sshd[3705]: pam_unix(sshd:session): session closed for user core Jan 16 09:00:58.742783 systemd[1]: sshd@14-147.182.199.56:22-139.178.68.195:42644.service: Deactivated successfully. Jan 16 09:00:58.746820 systemd[1]: session-14.scope: Deactivated successfully. Jan 16 09:00:58.757093 systemd-logind[1451]: Session 14 logged out. Waiting for processes to exit. Jan 16 09:00:58.766332 systemd[1]: Started sshd@15-147.182.199.56:22-139.178.68.195:42656.service - OpenSSH per-connection server daemon (139.178.68.195:42656). Jan 16 09:00:58.769130 systemd-logind[1451]: Removed session 14. Jan 16 09:00:58.856339 sshd[3716]: Accepted publickey for core from 139.178.68.195 port 42656 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:00:58.859331 sshd[3716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:00:58.867757 systemd-logind[1451]: New session 15 of user core. Jan 16 09:00:58.875019 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 16 09:01:01.971711 sshd[3716]: pam_unix(sshd:session): session closed for user core Jan 16 09:01:02.002321 systemd[1]: sshd@15-147.182.199.56:22-139.178.68.195:42656.service: Deactivated successfully. Jan 16 09:01:02.009914 systemd[1]: session-15.scope: Deactivated successfully. Jan 16 09:01:02.020408 systemd-logind[1451]: Session 15 logged out. Waiting for processes to exit. Jan 16 09:01:02.056905 systemd[1]: Started sshd@16-147.182.199.56:22-139.178.68.195:42660.service - OpenSSH per-connection server daemon (139.178.68.195:42660). Jan 16 09:01:02.063508 systemd-logind[1451]: Removed session 15. Jan 16 09:01:02.154832 sshd[3755]: Accepted publickey for core from 139.178.68.195 port 42660 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:01:02.160915 sshd[3755]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:01:02.179909 systemd-logind[1451]: New session 16 of user core. Jan 16 09:01:02.203108 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 16 09:01:02.574074 kubelet[2523]: E0116 09:01:02.573393 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:01:02.957753 sshd[3755]: pam_unix(sshd:session): session closed for user core Jan 16 09:01:02.986397 systemd[1]: sshd@16-147.182.199.56:22-139.178.68.195:42660.service: Deactivated successfully. Jan 16 09:01:02.992243 systemd[1]: session-16.scope: Deactivated successfully. Jan 16 09:01:02.997451 systemd-logind[1451]: Session 16 logged out. Waiting for processes to exit. Jan 16 09:01:03.003608 systemd-logind[1451]: Removed session 16. Jan 16 09:01:03.017635 systemd[1]: Started sshd@17-147.182.199.56:22-139.178.68.195:42676.service - OpenSSH per-connection server daemon (139.178.68.195:42676). Jan 16 09:01:03.115759 sshd[3767]: Accepted publickey for core from 139.178.68.195 port 42676 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:01:03.119422 sshd[3767]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:01:03.137119 systemd-logind[1451]: New session 17 of user core. Jan 16 09:01:03.144280 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 16 09:01:03.393310 sshd[3767]: pam_unix(sshd:session): session closed for user core Jan 16 09:01:03.399642 systemd[1]: sshd@17-147.182.199.56:22-139.178.68.195:42676.service: Deactivated successfully. Jan 16 09:01:03.403290 systemd[1]: session-17.scope: Deactivated successfully. Jan 16 09:01:03.408141 systemd-logind[1451]: Session 17 logged out. Waiting for processes to exit. Jan 16 09:01:03.411604 systemd-logind[1451]: Removed session 17. Jan 16 09:01:03.556157 kubelet[2523]: E0116 09:01:03.555246 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:01:08.418967 systemd[1]: Started sshd@18-147.182.199.56:22-139.178.68.195:48910.service - OpenSSH per-connection server daemon (139.178.68.195:48910). Jan 16 09:01:08.508302 sshd[3801]: Accepted publickey for core from 139.178.68.195 port 48910 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:01:08.513198 sshd[3801]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:01:08.532093 systemd-logind[1451]: New session 18 of user core. Jan 16 09:01:08.543968 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 16 09:01:08.558689 kubelet[2523]: E0116 09:01:08.558212 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:01:08.561757 kubelet[2523]: E0116 09:01:08.561401 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:01:08.809158 sshd[3801]: pam_unix(sshd:session): session closed for user core Jan 16 09:01:08.824021 systemd[1]: sshd@18-147.182.199.56:22-139.178.68.195:48910.service: Deactivated successfully. Jan 16 09:01:08.829623 systemd[1]: session-18.scope: Deactivated successfully. Jan 16 09:01:08.832237 systemd-logind[1451]: Session 18 logged out. Waiting for processes to exit. Jan 16 09:01:08.837987 systemd-logind[1451]: Removed session 18. Jan 16 09:01:09.559690 kubelet[2523]: E0116 09:01:09.557353 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" Jan 16 09:01:13.830357 systemd[1]: Started sshd@19-147.182.199.56:22-139.178.68.195:48916.service - OpenSSH per-connection server daemon (139.178.68.195:48916). Jan 16 09:01:13.889457 sshd[3837]: Accepted publickey for core from 139.178.68.195 port 48916 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:01:13.892412 sshd[3837]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:01:13.901878 systemd-logind[1451]: New session 19 of user core. Jan 16 09:01:13.909005 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 16 09:01:14.088596 sshd[3837]: pam_unix(sshd:session): session closed for user core Jan 16 09:01:14.095572 systemd[1]: sshd@19-147.182.199.56:22-139.178.68.195:48916.service: Deactivated successfully. Jan 16 09:01:14.099211 systemd[1]: session-19.scope: Deactivated successfully. Jan 16 09:01:14.101251 systemd-logind[1451]: Session 19 logged out. Waiting for processes to exit. Jan 16 09:01:14.104297 systemd-logind[1451]: Removed session 19. Jan 16 09:01:15.536273 systemd[1]: Started sshd@20-147.182.199.56:22-92.118.39.87:50758.service - OpenSSH per-connection server daemon (92.118.39.87:50758). Jan 16 09:01:16.254157 sshd[3855]: Invalid user ubuntu from 92.118.39.87 port 50758 Jan 16 09:01:16.421898 sshd[3855]: Connection closed by invalid user ubuntu 92.118.39.87 port 50758 [preauth] Jan 16 09:01:16.424420 systemd[1]: sshd@20-147.182.199.56:22-92.118.39.87:50758.service: Deactivated successfully. Jan 16 09:01:19.114722 systemd[1]: Started sshd@21-147.182.199.56:22-139.178.68.195:46508.service - OpenSSH per-connection server daemon (139.178.68.195:46508). Jan 16 09:01:19.169525 sshd[3875]: Accepted publickey for core from 139.178.68.195 port 46508 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:01:19.172497 sshd[3875]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:01:19.182590 systemd-logind[1451]: New session 20 of user core. Jan 16 09:01:19.189347 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 16 09:01:19.381405 sshd[3875]: pam_unix(sshd:session): session closed for user core Jan 16 09:01:19.391344 systemd-logind[1451]: Session 20 logged out. Waiting for processes to exit. Jan 16 09:01:19.391649 systemd[1]: sshd@21-147.182.199.56:22-139.178.68.195:46508.service: Deactivated successfully. Jan 16 09:01:19.397195 systemd[1]: session-20.scope: Deactivated successfully. Jan 16 09:01:19.399512 systemd-logind[1451]: Removed session 20. Jan 16 09:01:24.409975 systemd[1]: Started sshd@22-147.182.199.56:22-139.178.68.195:46516.service - OpenSSH per-connection server daemon (139.178.68.195:46516). Jan 16 09:01:24.479727 sshd[3910]: Accepted publickey for core from 139.178.68.195 port 46516 ssh2: RSA SHA256:fWXAJ6WCtHVKvQlmcI2C6JuFf3oBdxh55gZP5IlKwm0 Jan 16 09:01:24.482329 sshd[3910]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 16 09:01:24.494009 systemd-logind[1451]: New session 21 of user core. Jan 16 09:01:24.501079 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 16 09:01:24.759159 sshd[3910]: pam_unix(sshd:session): session closed for user core Jan 16 09:01:24.767856 systemd[1]: sshd@22-147.182.199.56:22-139.178.68.195:46516.service: Deactivated successfully. Jan 16 09:01:24.773384 systemd[1]: session-21.scope: Deactivated successfully. Jan 16 09:01:24.775385 systemd-logind[1451]: Session 21 logged out. Waiting for processes to exit. Jan 16 09:01:24.778180 systemd-logind[1451]: Removed session 21.