Jan 17 00:15:46.998082 kernel: Linux version 6.6.119-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Jan 16 22:25:55 -00 2026 Jan 17 00:15:46.998116 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=5950c0a3c50d11b7bc07a3e3bf06049ed0b5a605b5e0b52a981b78f1c63eeedd Jan 17 00:15:46.998978 kernel: BIOS-provided physical RAM map: Jan 17 00:15:46.998991 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Jan 17 00:15:46.999002 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Jan 17 00:15:46.999012 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Jan 17 00:15:46.999025 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Jan 17 00:15:46.999052 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Jan 17 00:15:46.999064 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Jan 17 00:15:46.999084 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Jan 17 00:15:46.999095 kernel: NX (Execute Disable) protection: active Jan 17 00:15:46.999106 kernel: APIC: Static calls initialized Jan 17 00:15:46.999136 kernel: SMBIOS 2.8 present. Jan 17 00:15:46.999151 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Jan 17 00:15:46.999165 kernel: Hypervisor detected: KVM Jan 17 00:15:46.999183 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jan 17 00:15:46.999195 kernel: kvm-clock: using sched offset of 3424572062 cycles Jan 17 00:15:46.999204 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jan 17 00:15:46.999213 kernel: tsc: Detected 2494.140 MHz processor Jan 17 00:15:46.999231 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 17 00:15:46.999240 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 17 00:15:46.999248 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Jan 17 00:15:46.999256 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Jan 17 00:15:46.999265 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 17 00:15:46.999277 kernel: ACPI: Early table checksum verification disabled Jan 17 00:15:46.999285 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Jan 17 00:15:46.999293 kernel: ACPI: RSDT 0x000000007FFE19FD 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 17 00:15:46.999301 kernel: ACPI: FACP 0x000000007FFE17E1 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 17 00:15:46.999310 kernel: ACPI: DSDT 0x000000007FFE0040 0017A1 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 17 00:15:46.999318 kernel: ACPI: FACS 0x000000007FFE0000 000040 Jan 17 00:15:46.999333 kernel: ACPI: APIC 0x000000007FFE1855 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 17 00:15:46.999345 kernel: ACPI: HPET 0x000000007FFE18D5 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 17 00:15:46.999366 kernel: ACPI: SRAT 0x000000007FFE190D 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 17 00:15:46.999383 kernel: ACPI: WAET 0x000000007FFE19D5 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 17 00:15:46.999391 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe17e1-0x7ffe1854] Jan 17 00:15:46.999399 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe17e0] Jan 17 00:15:46.999407 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Jan 17 00:15:46.999415 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe1855-0x7ffe18d4] Jan 17 00:15:46.999423 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe18d5-0x7ffe190c] Jan 17 00:15:46.999431 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe190d-0x7ffe19d4] Jan 17 00:15:46.999448 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe19d5-0x7ffe19fc] Jan 17 00:15:46.999456 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Jan 17 00:15:46.999465 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Jan 17 00:15:46.999474 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Jan 17 00:15:46.999482 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Jan 17 00:15:46.999494 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Jan 17 00:15:46.999502 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Jan 17 00:15:46.999515 kernel: Zone ranges: Jan 17 00:15:46.999528 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 17 00:15:46.999541 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Jan 17 00:15:46.999554 kernel: Normal empty Jan 17 00:15:46.999566 kernel: Movable zone start for each node Jan 17 00:15:46.999574 kernel: Early memory node ranges Jan 17 00:15:46.999582 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Jan 17 00:15:46.999591 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Jan 17 00:15:46.999599 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Jan 17 00:15:46.999612 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 17 00:15:46.999626 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Jan 17 00:15:46.999644 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Jan 17 00:15:46.999657 kernel: ACPI: PM-Timer IO Port: 0x608 Jan 17 00:15:46.999665 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jan 17 00:15:46.999674 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jan 17 00:15:46.999683 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jan 17 00:15:46.999691 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jan 17 00:15:46.999701 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jan 17 00:15:46.999721 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jan 17 00:15:46.999737 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jan 17 00:15:46.999751 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 17 00:15:46.999765 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jan 17 00:15:46.999777 kernel: TSC deadline timer available Jan 17 00:15:46.999788 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Jan 17 00:15:46.999803 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jan 17 00:15:46.999818 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Jan 17 00:15:46.999837 kernel: Booting paravirtualized kernel on KVM Jan 17 00:15:46.999852 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 17 00:15:46.999870 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Jan 17 00:15:46.999884 kernel: percpu: Embedded 57 pages/cpu s196328 r8192 d28952 u1048576 Jan 17 00:15:46.999896 kernel: pcpu-alloc: s196328 r8192 d28952 u1048576 alloc=1*2097152 Jan 17 00:15:46.999907 kernel: pcpu-alloc: [0] 0 1 Jan 17 00:15:46.999920 kernel: kvm-guest: PV spinlocks disabled, no host support Jan 17 00:15:46.999934 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=5950c0a3c50d11b7bc07a3e3bf06049ed0b5a605b5e0b52a981b78f1c63eeedd Jan 17 00:15:46.999945 kernel: random: crng init done Jan 17 00:15:46.999957 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 17 00:15:46.999974 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jan 17 00:15:46.999986 kernel: Fallback order for Node 0: 0 Jan 17 00:15:46.999998 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Jan 17 00:15:47.000011 kernel: Policy zone: DMA32 Jan 17 00:15:47.000023 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 17 00:15:47.000039 kernel: Memory: 1971212K/2096612K available (12288K kernel code, 2288K rwdata, 22752K rodata, 42884K init, 2312K bss, 125140K reserved, 0K cma-reserved) Jan 17 00:15:47.000052 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jan 17 00:15:47.000067 kernel: Kernel/User page tables isolation: enabled Jan 17 00:15:47.000086 kernel: ftrace: allocating 37989 entries in 149 pages Jan 17 00:15:47.000099 kernel: ftrace: allocated 149 pages with 4 groups Jan 17 00:15:47.000112 kernel: Dynamic Preempt: voluntary Jan 17 00:15:47.000148 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 17 00:15:47.000164 kernel: rcu: RCU event tracing is enabled. Jan 17 00:15:47.000176 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jan 17 00:15:47.000190 kernel: Trampoline variant of Tasks RCU enabled. Jan 17 00:15:47.000201 kernel: Rude variant of Tasks RCU enabled. Jan 17 00:15:47.000213 kernel: Tracing variant of Tasks RCU enabled. Jan 17 00:15:47.000224 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 17 00:15:47.000243 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jan 17 00:15:47.000254 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Jan 17 00:15:47.000266 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 17 00:15:47.000278 kernel: Console: colour VGA+ 80x25 Jan 17 00:15:47.000295 kernel: printk: console [tty0] enabled Jan 17 00:15:47.000307 kernel: printk: console [ttyS0] enabled Jan 17 00:15:47.000319 kernel: ACPI: Core revision 20230628 Jan 17 00:15:47.000331 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jan 17 00:15:47.000342 kernel: APIC: Switch to symmetric I/O mode setup Jan 17 00:15:47.000359 kernel: x2apic enabled Jan 17 00:15:47.000371 kernel: APIC: Switched APIC routing to: physical x2apic Jan 17 00:15:47.000382 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 17 00:15:47.000394 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Jan 17 00:15:47.000406 kernel: Calibrating delay loop (skipped) preset value.. 4988.28 BogoMIPS (lpj=2494140) Jan 17 00:15:47.000418 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Jan 17 00:15:47.000431 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Jan 17 00:15:47.000443 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 17 00:15:47.000473 kernel: Spectre V2 : Mitigation: Retpolines Jan 17 00:15:47.000485 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jan 17 00:15:47.000497 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Jan 17 00:15:47.000514 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jan 17 00:15:47.000526 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jan 17 00:15:47.000539 kernel: MDS: Mitigation: Clear CPU buffers Jan 17 00:15:47.000551 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jan 17 00:15:47.000564 kernel: active return thunk: its_return_thunk Jan 17 00:15:47.000581 kernel: ITS: Mitigation: Aligned branch/return thunks Jan 17 00:15:47.000600 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 17 00:15:47.000613 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 17 00:15:47.000625 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 17 00:15:47.000637 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 17 00:15:47.000650 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Jan 17 00:15:47.000662 kernel: Freeing SMP alternatives memory: 32K Jan 17 00:15:47.000674 kernel: pid_max: default: 32768 minimum: 301 Jan 17 00:15:47.000687 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 17 00:15:47.000704 kernel: landlock: Up and running. Jan 17 00:15:47.000717 kernel: SELinux: Initializing. Jan 17 00:15:47.000730 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 17 00:15:47.000743 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 17 00:15:47.000756 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Jan 17 00:15:47.000768 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 17 00:15:47.000781 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 17 00:15:47.000794 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 17 00:15:47.000807 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Jan 17 00:15:47.000825 kernel: signal: max sigframe size: 1776 Jan 17 00:15:47.000838 kernel: rcu: Hierarchical SRCU implementation. Jan 17 00:15:47.000851 kernel: rcu: Max phase no-delay instances is 400. Jan 17 00:15:47.000863 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 17 00:15:47.000875 kernel: smp: Bringing up secondary CPUs ... Jan 17 00:15:47.000889 kernel: smpboot: x86: Booting SMP configuration: Jan 17 00:15:47.000901 kernel: .... node #0, CPUs: #1 Jan 17 00:15:47.000914 kernel: smp: Brought up 1 node, 2 CPUs Jan 17 00:15:47.000932 kernel: smpboot: Max logical packages: 1 Jan 17 00:15:47.000950 kernel: smpboot: Total of 2 processors activated (9976.56 BogoMIPS) Jan 17 00:15:47.000965 kernel: devtmpfs: initialized Jan 17 00:15:47.000980 kernel: x86/mm: Memory block size: 128MB Jan 17 00:15:47.000993 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 17 00:15:47.001006 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jan 17 00:15:47.001020 kernel: pinctrl core: initialized pinctrl subsystem Jan 17 00:15:47.001034 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 17 00:15:47.001048 kernel: audit: initializing netlink subsys (disabled) Jan 17 00:15:47.001060 kernel: audit: type=2000 audit(1768608945.413:1): state=initialized audit_enabled=0 res=1 Jan 17 00:15:47.001080 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 17 00:15:47.001094 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 17 00:15:47.001108 kernel: cpuidle: using governor menu Jan 17 00:15:47.001117 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 17 00:15:47.001139 kernel: dca service started, version 1.12.1 Jan 17 00:15:47.001148 kernel: PCI: Using configuration type 1 for base access Jan 17 00:15:47.001158 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 17 00:15:47.001167 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 17 00:15:47.001176 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 17 00:15:47.001190 kernel: ACPI: Added _OSI(Module Device) Jan 17 00:15:47.001199 kernel: ACPI: Added _OSI(Processor Device) Jan 17 00:15:47.001209 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 17 00:15:47.001218 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 17 00:15:47.001227 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jan 17 00:15:47.001236 kernel: ACPI: Interpreter enabled Jan 17 00:15:47.001245 kernel: ACPI: PM: (supports S0 S5) Jan 17 00:15:47.001254 kernel: ACPI: Using IOAPIC for interrupt routing Jan 17 00:15:47.001263 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 17 00:15:47.001276 kernel: PCI: Using E820 reservations for host bridge windows Jan 17 00:15:47.001285 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Jan 17 00:15:47.001294 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 17 00:15:47.001546 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Jan 17 00:15:47.001665 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Jan 17 00:15:47.001762 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Jan 17 00:15:47.001774 kernel: acpiphp: Slot [3] registered Jan 17 00:15:47.001788 kernel: acpiphp: Slot [4] registered Jan 17 00:15:47.001797 kernel: acpiphp: Slot [5] registered Jan 17 00:15:47.001806 kernel: acpiphp: Slot [6] registered Jan 17 00:15:47.001815 kernel: acpiphp: Slot [7] registered Jan 17 00:15:47.001824 kernel: acpiphp: Slot [8] registered Jan 17 00:15:47.001833 kernel: acpiphp: Slot [9] registered Jan 17 00:15:47.001842 kernel: acpiphp: Slot [10] registered Jan 17 00:15:47.001851 kernel: acpiphp: Slot [11] registered Jan 17 00:15:47.001860 kernel: acpiphp: Slot [12] registered Jan 17 00:15:47.001873 kernel: acpiphp: Slot [13] registered Jan 17 00:15:47.001882 kernel: acpiphp: Slot [14] registered Jan 17 00:15:47.001891 kernel: acpiphp: Slot [15] registered Jan 17 00:15:47.001899 kernel: acpiphp: Slot [16] registered Jan 17 00:15:47.001908 kernel: acpiphp: Slot [17] registered Jan 17 00:15:47.001917 kernel: acpiphp: Slot [18] registered Jan 17 00:15:47.001926 kernel: acpiphp: Slot [19] registered Jan 17 00:15:47.001935 kernel: acpiphp: Slot [20] registered Jan 17 00:15:47.001944 kernel: acpiphp: Slot [21] registered Jan 17 00:15:47.001953 kernel: acpiphp: Slot [22] registered Jan 17 00:15:47.001965 kernel: acpiphp: Slot [23] registered Jan 17 00:15:47.001974 kernel: acpiphp: Slot [24] registered Jan 17 00:15:47.001983 kernel: acpiphp: Slot [25] registered Jan 17 00:15:47.001991 kernel: acpiphp: Slot [26] registered Jan 17 00:15:47.002004 kernel: acpiphp: Slot [27] registered Jan 17 00:15:47.002019 kernel: acpiphp: Slot [28] registered Jan 17 00:15:47.002032 kernel: acpiphp: Slot [29] registered Jan 17 00:15:47.002045 kernel: acpiphp: Slot [30] registered Jan 17 00:15:47.002060 kernel: acpiphp: Slot [31] registered Jan 17 00:15:47.002074 kernel: PCI host bridge to bus 0000:00 Jan 17 00:15:47.002235 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 17 00:15:47.002330 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jan 17 00:15:47.002416 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 17 00:15:47.002502 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Jan 17 00:15:47.002659 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Jan 17 00:15:47.002773 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 17 00:15:47.002978 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Jan 17 00:15:47.003211 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Jan 17 00:15:47.003367 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Jan 17 00:15:47.003485 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Jan 17 00:15:47.003651 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Jan 17 00:15:47.003759 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Jan 17 00:15:47.003882 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Jan 17 00:15:47.004006 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Jan 17 00:15:47.004142 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Jan 17 00:15:47.004245 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Jan 17 00:15:47.004401 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Jan 17 00:15:47.004518 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Jan 17 00:15:47.004621 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Jan 17 00:15:47.004753 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Jan 17 00:15:47.004860 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Jan 17 00:15:47.004989 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Jan 17 00:15:47.005152 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Jan 17 00:15:47.005273 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Jan 17 00:15:47.005371 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 17 00:15:47.005493 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Jan 17 00:15:47.005602 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Jan 17 00:15:47.005714 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Jan 17 00:15:47.005812 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Jan 17 00:15:47.005933 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Jan 17 00:15:47.006030 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Jan 17 00:15:47.006150 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Jan 17 00:15:47.006258 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Jan 17 00:15:47.006396 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Jan 17 00:15:47.006505 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Jan 17 00:15:47.006604 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Jan 17 00:15:47.006703 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Jan 17 00:15:47.006814 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Jan 17 00:15:47.006915 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Jan 17 00:15:47.007024 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Jan 17 00:15:47.007330 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Jan 17 00:15:47.007454 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Jan 17 00:15:47.007557 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Jan 17 00:15:47.007659 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Jan 17 00:15:47.007806 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Jan 17 00:15:47.007954 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Jan 17 00:15:47.008064 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Jan 17 00:15:47.008171 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Jan 17 00:15:47.008184 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jan 17 00:15:47.008194 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jan 17 00:15:47.008203 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 17 00:15:47.008213 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jan 17 00:15:47.008222 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Jan 17 00:15:47.008231 kernel: iommu: Default domain type: Translated Jan 17 00:15:47.008245 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 17 00:15:47.008254 kernel: PCI: Using ACPI for IRQ routing Jan 17 00:15:47.008264 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 17 00:15:47.008273 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Jan 17 00:15:47.008282 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Jan 17 00:15:47.008380 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Jan 17 00:15:47.008519 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Jan 17 00:15:47.008655 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 17 00:15:47.008677 kernel: vgaarb: loaded Jan 17 00:15:47.008687 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jan 17 00:15:47.008697 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jan 17 00:15:47.008706 kernel: clocksource: Switched to clocksource kvm-clock Jan 17 00:15:47.008715 kernel: VFS: Disk quotas dquot_6.6.0 Jan 17 00:15:47.008725 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 17 00:15:47.008734 kernel: pnp: PnP ACPI init Jan 17 00:15:47.008744 kernel: pnp: PnP ACPI: found 4 devices Jan 17 00:15:47.008753 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 17 00:15:47.008766 kernel: NET: Registered PF_INET protocol family Jan 17 00:15:47.008776 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 17 00:15:47.008785 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Jan 17 00:15:47.008795 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 17 00:15:47.008804 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Jan 17 00:15:47.008813 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Jan 17 00:15:47.008822 kernel: TCP: Hash tables configured (established 16384 bind 16384) Jan 17 00:15:47.008831 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 17 00:15:47.008840 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 17 00:15:47.008853 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 17 00:15:47.008862 kernel: NET: Registered PF_XDP protocol family Jan 17 00:15:47.008990 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jan 17 00:15:47.009083 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jan 17 00:15:47.009233 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jan 17 00:15:47.009322 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Jan 17 00:15:47.009434 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Jan 17 00:15:47.009571 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Jan 17 00:15:47.009705 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jan 17 00:15:47.009721 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Jan 17 00:15:47.009821 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x7b0 took 36572 usecs Jan 17 00:15:47.009834 kernel: PCI: CLS 0 bytes, default 64 Jan 17 00:15:47.009843 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jan 17 00:15:47.009853 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Jan 17 00:15:47.009862 kernel: Initialise system trusted keyrings Jan 17 00:15:47.009872 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Jan 17 00:15:47.009886 kernel: Key type asymmetric registered Jan 17 00:15:47.009895 kernel: Asymmetric key parser 'x509' registered Jan 17 00:15:47.009904 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jan 17 00:15:47.009913 kernel: io scheduler mq-deadline registered Jan 17 00:15:47.009923 kernel: io scheduler kyber registered Jan 17 00:15:47.009932 kernel: io scheduler bfq registered Jan 17 00:15:47.009942 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 17 00:15:47.009951 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Jan 17 00:15:47.009960 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Jan 17 00:15:47.009969 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Jan 17 00:15:47.009983 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 17 00:15:47.009992 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 17 00:15:47.010001 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jan 17 00:15:47.010011 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 17 00:15:47.010020 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 17 00:15:47.010226 kernel: rtc_cmos 00:03: RTC can wake from S4 Jan 17 00:15:47.010249 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jan 17 00:15:47.010366 kernel: rtc_cmos 00:03: registered as rtc0 Jan 17 00:15:47.010467 kernel: rtc_cmos 00:03: setting system clock to 2026-01-17T00:15:46 UTC (1768608946) Jan 17 00:15:47.010588 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Jan 17 00:15:47.010606 kernel: intel_pstate: CPU model not supported Jan 17 00:15:47.010619 kernel: NET: Registered PF_INET6 protocol family Jan 17 00:15:47.010635 kernel: Segment Routing with IPv6 Jan 17 00:15:47.010649 kernel: In-situ OAM (IOAM) with IPv6 Jan 17 00:15:47.010663 kernel: NET: Registered PF_PACKET protocol family Jan 17 00:15:47.010673 kernel: Key type dns_resolver registered Jan 17 00:15:47.010688 kernel: IPI shorthand broadcast: enabled Jan 17 00:15:47.010698 kernel: sched_clock: Marking stable (1174003918, 171311572)->(1379444805, -34129315) Jan 17 00:15:47.010707 kernel: registered taskstats version 1 Jan 17 00:15:47.010717 kernel: Loading compiled-in X.509 certificates Jan 17 00:15:47.010726 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.119-flatcar: b6a847a3a522371f15b0d5425f12279a240740e4' Jan 17 00:15:47.010735 kernel: Key type .fscrypt registered Jan 17 00:15:47.010744 kernel: Key type fscrypt-provisioning registered Jan 17 00:15:47.010753 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 17 00:15:47.010762 kernel: ima: Allocated hash algorithm: sha1 Jan 17 00:15:47.010775 kernel: ima: No architecture policies found Jan 17 00:15:47.010784 kernel: clk: Disabling unused clocks Jan 17 00:15:47.010793 kernel: Freeing unused kernel image (initmem) memory: 42884K Jan 17 00:15:47.010803 kernel: Write protecting the kernel read-only data: 36864k Jan 17 00:15:47.010812 kernel: Freeing unused kernel image (rodata/data gap) memory: 1824K Jan 17 00:15:47.010845 kernel: Run /init as init process Jan 17 00:15:47.010858 kernel: with arguments: Jan 17 00:15:47.010868 kernel: /init Jan 17 00:15:47.010878 kernel: with environment: Jan 17 00:15:47.010887 kernel: HOME=/ Jan 17 00:15:47.010900 kernel: TERM=linux Jan 17 00:15:47.010914 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 17 00:15:47.010927 systemd[1]: Detected virtualization kvm. Jan 17 00:15:47.010937 systemd[1]: Detected architecture x86-64. Jan 17 00:15:47.010947 systemd[1]: Running in initrd. Jan 17 00:15:47.010957 systemd[1]: No hostname configured, using default hostname. Jan 17 00:15:47.010966 systemd[1]: Hostname set to . Jan 17 00:15:47.010980 systemd[1]: Initializing machine ID from VM UUID. Jan 17 00:15:47.010990 systemd[1]: Queued start job for default target initrd.target. Jan 17 00:15:47.011001 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 00:15:47.011015 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 00:15:47.011030 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 17 00:15:47.011059 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 17 00:15:47.011075 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 17 00:15:47.011090 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 17 00:15:47.011112 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 17 00:15:47.011142 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 17 00:15:47.011159 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 00:15:47.011172 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 17 00:15:47.011182 systemd[1]: Reached target paths.target - Path Units. Jan 17 00:15:47.011192 systemd[1]: Reached target slices.target - Slice Units. Jan 17 00:15:47.011203 systemd[1]: Reached target swap.target - Swaps. Jan 17 00:15:47.011218 systemd[1]: Reached target timers.target - Timer Units. Jan 17 00:15:47.011228 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 17 00:15:47.011238 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 17 00:15:47.011248 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 17 00:15:47.011258 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 17 00:15:47.011272 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 17 00:15:47.011283 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 17 00:15:47.011297 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 00:15:47.011307 systemd[1]: Reached target sockets.target - Socket Units. Jan 17 00:15:47.011318 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 17 00:15:47.011328 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 17 00:15:47.011339 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 17 00:15:47.011349 systemd[1]: Starting systemd-fsck-usr.service... Jan 17 00:15:47.011360 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 17 00:15:47.011375 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 17 00:15:47.011385 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 00:15:47.011395 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 17 00:15:47.011409 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 00:15:47.011420 systemd[1]: Finished systemd-fsck-usr.service. Jan 17 00:15:47.011473 systemd-journald[186]: Collecting audit messages is disabled. Jan 17 00:15:47.011502 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 17 00:15:47.011515 systemd-journald[186]: Journal started Jan 17 00:15:47.011542 systemd-journald[186]: Runtime Journal (/run/log/journal/243af4b5561a4f9b93af0aee0d527157) is 4.9M, max 39.3M, 34.4M free. Jan 17 00:15:47.004062 systemd-modules-load[187]: Inserted module 'overlay' Jan 17 00:15:47.067996 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 17 00:15:47.068036 kernel: Bridge firewalling registered Jan 17 00:15:47.068056 systemd[1]: Started systemd-journald.service - Journal Service. Jan 17 00:15:47.039233 systemd-modules-load[187]: Inserted module 'br_netfilter' Jan 17 00:15:47.067696 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 17 00:15:47.075543 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 00:15:47.080477 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 00:15:47.097741 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 17 00:15:47.103443 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 17 00:15:47.105287 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 00:15:47.119520 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 17 00:15:47.125190 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 17 00:15:47.137680 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 00:15:47.138774 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 00:15:47.140670 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 00:15:47.147544 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 17 00:15:47.152374 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 17 00:15:47.174177 dracut-cmdline[219]: dracut-dracut-053 Jan 17 00:15:47.180100 dracut-cmdline[219]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=5950c0a3c50d11b7bc07a3e3bf06049ed0b5a605b5e0b52a981b78f1c63eeedd Jan 17 00:15:47.198936 systemd-resolved[221]: Positive Trust Anchors: Jan 17 00:15:47.198958 systemd-resolved[221]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 17 00:15:47.198995 systemd-resolved[221]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 17 00:15:47.202781 systemd-resolved[221]: Defaulting to hostname 'linux'. Jan 17 00:15:47.205364 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 17 00:15:47.206065 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 17 00:15:47.311192 kernel: SCSI subsystem initialized Jan 17 00:15:47.322171 kernel: Loading iSCSI transport class v2.0-870. Jan 17 00:15:47.336225 kernel: iscsi: registered transport (tcp) Jan 17 00:15:47.361282 kernel: iscsi: registered transport (qla4xxx) Jan 17 00:15:47.361418 kernel: QLogic iSCSI HBA Driver Jan 17 00:15:47.431871 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 17 00:15:47.439436 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 17 00:15:47.472689 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 17 00:15:47.472791 kernel: device-mapper: uevent: version 1.0.3 Jan 17 00:15:47.472816 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 17 00:15:47.524183 kernel: raid6: avx2x4 gen() 20412 MB/s Jan 17 00:15:47.541189 kernel: raid6: avx2x2 gen() 20742 MB/s Jan 17 00:15:47.558429 kernel: raid6: avx2x1 gen() 17187 MB/s Jan 17 00:15:47.558562 kernel: raid6: using algorithm avx2x2 gen() 20742 MB/s Jan 17 00:15:47.578370 kernel: raid6: .... xor() 17674 MB/s, rmw enabled Jan 17 00:15:47.578515 kernel: raid6: using avx2x2 recovery algorithm Jan 17 00:15:47.605193 kernel: xor: automatically using best checksumming function avx Jan 17 00:15:47.784179 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 17 00:15:47.801314 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 17 00:15:47.808474 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 00:15:47.839568 systemd-udevd[404]: Using default interface naming scheme 'v255'. Jan 17 00:15:47.845426 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 00:15:47.852551 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 17 00:15:47.881072 dracut-pre-trigger[407]: rd.md=0: removing MD RAID activation Jan 17 00:15:47.930766 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 17 00:15:47.937429 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 17 00:15:48.034055 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 00:15:48.042459 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 17 00:15:48.066891 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 17 00:15:48.071535 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 17 00:15:48.072578 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 00:15:48.074780 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 17 00:15:48.082679 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 17 00:15:48.120534 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 17 00:15:48.136228 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Jan 17 00:15:48.146167 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Jan 17 00:15:48.157222 kernel: scsi host0: Virtio SCSI HBA Jan 17 00:15:48.165272 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 17 00:15:48.165355 kernel: GPT:9289727 != 125829119 Jan 17 00:15:48.165368 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 17 00:15:48.165381 kernel: GPT:9289727 != 125829119 Jan 17 00:15:48.166261 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 17 00:15:48.171680 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 17 00:15:48.201167 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Jan 17 00:15:48.207012 kernel: virtio_blk virtio5: [vdb] 964 512-byte logical blocks (494 kB/482 KiB) Jan 17 00:15:48.214160 kernel: libata version 3.00 loaded. Jan 17 00:15:48.220058 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 17 00:15:48.222249 kernel: cryptd: max_cpu_qlen set to 1000 Jan 17 00:15:48.220235 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 00:15:48.233695 kernel: ACPI: bus type USB registered Jan 17 00:15:48.233734 kernel: usbcore: registered new interface driver usbfs Jan 17 00:15:48.233748 kernel: usbcore: registered new interface driver hub Jan 17 00:15:48.233760 kernel: usbcore: registered new device driver usb Jan 17 00:15:48.233498 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 00:15:48.234092 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 00:15:48.234303 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 00:15:48.235578 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 00:15:48.244180 kernel: ata_piix 0000:00:01.1: version 2.13 Jan 17 00:15:48.243491 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 00:15:48.266223 kernel: AVX2 version of gcm_enc/dec engaged. Jan 17 00:15:48.266352 kernel: AES CTR mode by8 optimization enabled Jan 17 00:15:48.269187 kernel: scsi host1: ata_piix Jan 17 00:15:48.278674 kernel: scsi host2: ata_piix Jan 17 00:15:48.278956 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Jan 17 00:15:48.278972 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Jan 17 00:15:48.312161 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/vda6 scanned by (udev-worker) (458) Jan 17 00:15:48.316190 kernel: BTRFS: device fsid a67b5ac0-cdfd-426d-9386-e029282f433a devid 1 transid 33 /dev/vda3 scanned by (udev-worker) (465) Jan 17 00:15:48.320849 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jan 17 00:15:48.387287 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Jan 17 00:15:48.387555 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Jan 17 00:15:48.387681 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Jan 17 00:15:48.387798 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Jan 17 00:15:48.387923 kernel: hub 1-0:1.0: USB hub found Jan 17 00:15:48.388081 kernel: hub 1-0:1.0: 2 ports detected Jan 17 00:15:48.392232 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 00:15:48.398799 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jan 17 00:15:48.403356 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Jan 17 00:15:48.403992 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jan 17 00:15:48.413469 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 17 00:15:48.422402 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 17 00:15:48.425498 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 00:15:48.429854 disk-uuid[530]: Primary Header is updated. Jan 17 00:15:48.429854 disk-uuid[530]: Secondary Entries is updated. Jan 17 00:15:48.429854 disk-uuid[530]: Secondary Header is updated. Jan 17 00:15:48.447741 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 17 00:15:48.456071 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 00:15:48.459178 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 17 00:15:48.476172 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 17 00:15:49.467164 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jan 17 00:15:49.467249 disk-uuid[531]: The operation has completed successfully. Jan 17 00:15:49.518282 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 17 00:15:49.518405 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 17 00:15:49.523383 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 17 00:15:49.536317 sh[564]: Success Jan 17 00:15:49.552171 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Jan 17 00:15:49.635318 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 17 00:15:49.648325 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 17 00:15:49.653782 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 17 00:15:49.675508 kernel: BTRFS info (device dm-0): first mount of filesystem a67b5ac0-cdfd-426d-9386-e029282f433a Jan 17 00:15:49.675596 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 17 00:15:49.678450 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 17 00:15:49.678548 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 17 00:15:49.680059 kernel: BTRFS info (device dm-0): using free space tree Jan 17 00:15:49.691077 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 17 00:15:49.692759 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 17 00:15:49.706456 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 17 00:15:49.709786 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 17 00:15:49.724943 kernel: BTRFS info (device vda6): first mount of filesystem 0f2efc88-79cd-4337-a46a-d3848e5a06b0 Jan 17 00:15:49.725013 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 00:15:49.725027 kernel: BTRFS info (device vda6): using free space tree Jan 17 00:15:49.732170 kernel: BTRFS info (device vda6): auto enabling async discard Jan 17 00:15:49.747526 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 17 00:15:49.749729 kernel: BTRFS info (device vda6): last unmount of filesystem 0f2efc88-79cd-4337-a46a-d3848e5a06b0 Jan 17 00:15:49.758947 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 17 00:15:49.766438 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 17 00:15:49.844733 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 17 00:15:49.861537 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 17 00:15:49.917823 systemd-networkd[746]: lo: Link UP Jan 17 00:15:49.917841 systemd-networkd[746]: lo: Gained carrier Jan 17 00:15:49.925501 systemd-networkd[746]: Enumeration completed Jan 17 00:15:49.925906 ignition[657]: Ignition 2.19.0 Jan 17 00:15:49.926598 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 17 00:15:49.925918 ignition[657]: Stage: fetch-offline Jan 17 00:15:49.927699 systemd[1]: Reached target network.target - Network. Jan 17 00:15:49.925987 ignition[657]: no configs at "/usr/lib/ignition/base.d" Jan 17 00:15:49.928591 systemd-networkd[746]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Jan 17 00:15:49.926006 ignition[657]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 17 00:15:49.928597 systemd-networkd[746]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Jan 17 00:15:49.926211 ignition[657]: parsed url from cmdline: "" Jan 17 00:15:49.930117 systemd-networkd[746]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 17 00:15:49.926218 ignition[657]: no config URL provided Jan 17 00:15:49.930224 systemd-networkd[746]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 17 00:15:49.926227 ignition[657]: reading system config file "/usr/lib/ignition/user.ign" Jan 17 00:15:49.932279 systemd-networkd[746]: eth0: Link UP Jan 17 00:15:49.926242 ignition[657]: no config at "/usr/lib/ignition/user.ign" Jan 17 00:15:49.932290 systemd-networkd[746]: eth0: Gained carrier Jan 17 00:15:49.926254 ignition[657]: failed to fetch config: resource requires networking Jan 17 00:15:49.932308 systemd-networkd[746]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Jan 17 00:15:49.927150 ignition[657]: Ignition finished successfully Jan 17 00:15:49.933970 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 17 00:15:49.936670 systemd-networkd[746]: eth1: Link UP Jan 17 00:15:49.936677 systemd-networkd[746]: eth1: Gained carrier Jan 17 00:15:49.936697 systemd-networkd[746]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 17 00:15:49.945507 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jan 17 00:15:49.952260 systemd-networkd[746]: eth0: DHCPv4 address 143.198.57.57/20, gateway 143.198.48.1 acquired from 169.254.169.253 Jan 17 00:15:49.960364 systemd-networkd[746]: eth1: DHCPv4 address 10.124.0.24/20 acquired from 169.254.169.253 Jan 17 00:15:49.980607 ignition[754]: Ignition 2.19.0 Jan 17 00:15:49.980624 ignition[754]: Stage: fetch Jan 17 00:15:49.980976 ignition[754]: no configs at "/usr/lib/ignition/base.d" Jan 17 00:15:49.980999 ignition[754]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 17 00:15:49.981184 ignition[754]: parsed url from cmdline: "" Jan 17 00:15:49.981190 ignition[754]: no config URL provided Jan 17 00:15:49.981199 ignition[754]: reading system config file "/usr/lib/ignition/user.ign" Jan 17 00:15:49.981215 ignition[754]: no config at "/usr/lib/ignition/user.ign" Jan 17 00:15:49.981244 ignition[754]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Jan 17 00:15:49.998835 ignition[754]: GET result: OK Jan 17 00:15:49.999092 ignition[754]: parsing config with SHA512: 60dcf3356fbf753c2e3406a5388d9b3e8c58b5a50e219a1404b5ad348fa7a4d5f95b1b730fcadc04591656e11c289ecf4262adea646426bdc5c383168e1e6630 Jan 17 00:15:50.006947 unknown[754]: fetched base config from "system" Jan 17 00:15:50.006970 unknown[754]: fetched base config from "system" Jan 17 00:15:50.008094 ignition[754]: fetch: fetch complete Jan 17 00:15:50.006982 unknown[754]: fetched user config from "digitalocean" Jan 17 00:15:50.008105 ignition[754]: fetch: fetch passed Jan 17 00:15:50.010924 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jan 17 00:15:50.008219 ignition[754]: Ignition finished successfully Jan 17 00:15:50.017505 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 17 00:15:50.055406 ignition[761]: Ignition 2.19.0 Jan 17 00:15:50.055420 ignition[761]: Stage: kargs Jan 17 00:15:50.055787 ignition[761]: no configs at "/usr/lib/ignition/base.d" Jan 17 00:15:50.055808 ignition[761]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 17 00:15:50.058794 ignition[761]: kargs: kargs passed Jan 17 00:15:50.058910 ignition[761]: Ignition finished successfully Jan 17 00:15:50.061374 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 17 00:15:50.071520 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 17 00:15:50.092467 ignition[769]: Ignition 2.19.0 Jan 17 00:15:50.092479 ignition[769]: Stage: disks Jan 17 00:15:50.092766 ignition[769]: no configs at "/usr/lib/ignition/base.d" Jan 17 00:15:50.092780 ignition[769]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 17 00:15:50.093805 ignition[769]: disks: disks passed Jan 17 00:15:50.096237 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 17 00:15:50.093874 ignition[769]: Ignition finished successfully Jan 17 00:15:50.102731 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 17 00:15:50.103844 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 17 00:15:50.104834 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 17 00:15:50.105997 systemd[1]: Reached target sysinit.target - System Initialization. Jan 17 00:15:50.107189 systemd[1]: Reached target basic.target - Basic System. Jan 17 00:15:50.113479 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 17 00:15:50.136573 systemd-fsck[778]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jan 17 00:15:50.140876 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 17 00:15:50.150357 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 17 00:15:50.272176 kernel: EXT4-fs (vda9): mounted filesystem ab055cfb-d92d-4784-aa05-26ea844796bc r/w with ordered data mode. Quota mode: none. Jan 17 00:15:50.272428 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 17 00:15:50.273926 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 17 00:15:50.283416 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 17 00:15:50.291440 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 17 00:15:50.294798 systemd[1]: Starting flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent... Jan 17 00:15:50.299557 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Jan 17 00:15:50.300327 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 17 00:15:50.300374 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 17 00:15:50.307679 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 17 00:15:50.315936 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/vda6 scanned by mount (786) Jan 17 00:15:50.316040 kernel: BTRFS info (device vda6): first mount of filesystem 0f2efc88-79cd-4337-a46a-d3848e5a06b0 Jan 17 00:15:50.320157 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 00:15:50.320237 kernel: BTRFS info (device vda6): using free space tree Jan 17 00:15:50.319367 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 17 00:15:50.328238 kernel: BTRFS info (device vda6): auto enabling async discard Jan 17 00:15:50.330529 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 17 00:15:50.398837 coreos-metadata[789]: Jan 17 00:15:50.398 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 17 00:15:50.413173 coreos-metadata[789]: Jan 17 00:15:50.410 INFO Fetch successful Jan 17 00:15:50.416038 coreos-metadata[789]: Jan 17 00:15:50.415 INFO wrote hostname ci-4081.3.6-n-09ad808b32 to /sysroot/etc/hostname Jan 17 00:15:50.417745 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 17 00:15:50.421389 coreos-metadata[788]: Jan 17 00:15:50.421 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 17 00:15:50.422768 initrd-setup-root[817]: cut: /sysroot/etc/passwd: No such file or directory Jan 17 00:15:50.428668 initrd-setup-root[825]: cut: /sysroot/etc/group: No such file or directory Jan 17 00:15:50.434672 initrd-setup-root[832]: cut: /sysroot/etc/shadow: No such file or directory Jan 17 00:15:50.436112 coreos-metadata[788]: Jan 17 00:15:50.435 INFO Fetch successful Jan 17 00:15:50.442820 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. Jan 17 00:15:50.442977 systemd[1]: Finished flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent. Jan 17 00:15:50.445450 initrd-setup-root[840]: cut: /sysroot/etc/gshadow: No such file or directory Jan 17 00:15:50.574881 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 17 00:15:50.583382 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 17 00:15:50.585788 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 17 00:15:50.600210 kernel: BTRFS info (device vda6): last unmount of filesystem 0f2efc88-79cd-4337-a46a-d3848e5a06b0 Jan 17 00:15:50.630388 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 17 00:15:50.642219 ignition[907]: INFO : Ignition 2.19.0 Jan 17 00:15:50.642219 ignition[907]: INFO : Stage: mount Jan 17 00:15:50.643697 ignition[907]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 00:15:50.643697 ignition[907]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 17 00:15:50.643697 ignition[907]: INFO : mount: mount passed Jan 17 00:15:50.643697 ignition[907]: INFO : Ignition finished successfully Jan 17 00:15:50.645687 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 17 00:15:50.653323 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 17 00:15:50.674395 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 17 00:15:50.678496 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 17 00:15:50.697169 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/vda6 scanned by mount (920) Jan 17 00:15:50.700368 kernel: BTRFS info (device vda6): first mount of filesystem 0f2efc88-79cd-4337-a46a-d3848e5a06b0 Jan 17 00:15:50.700478 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 00:15:50.702852 kernel: BTRFS info (device vda6): using free space tree Jan 17 00:15:50.709187 kernel: BTRFS info (device vda6): auto enabling async discard Jan 17 00:15:50.710103 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 17 00:15:50.753872 ignition[937]: INFO : Ignition 2.19.0 Jan 17 00:15:50.753872 ignition[937]: INFO : Stage: files Jan 17 00:15:50.755913 ignition[937]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 00:15:50.755913 ignition[937]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 17 00:15:50.755913 ignition[937]: DEBUG : files: compiled without relabeling support, skipping Jan 17 00:15:50.758513 ignition[937]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 17 00:15:50.758513 ignition[937]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 17 00:15:50.760752 ignition[937]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 17 00:15:50.761857 ignition[937]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 17 00:15:50.763358 unknown[937]: wrote ssh authorized keys file for user: core Jan 17 00:15:50.764581 ignition[937]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 17 00:15:50.767265 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Jan 17 00:15:50.768898 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Jan 17 00:15:50.834243 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 17 00:15:50.942164 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 17 00:15:50.951075 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 17 00:15:50.951075 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 17 00:15:50.951075 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Jan 17 00:15:50.951075 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Jan 17 00:15:50.951075 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Jan 17 00:15:50.951075 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Jan 17 00:15:51.127478 systemd-networkd[746]: eth0: Gained IPv6LL Jan 17 00:15:51.306683 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 17 00:15:51.383805 systemd-networkd[746]: eth1: Gained IPv6LL Jan 17 00:15:52.708857 ignition[937]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Jan 17 00:15:52.708857 ignition[937]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 17 00:15:52.711349 ignition[937]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 17 00:15:52.712195 ignition[937]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 17 00:15:52.712195 ignition[937]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 17 00:15:52.712195 ignition[937]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Jan 17 00:15:52.712195 ignition[937]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Jan 17 00:15:52.712195 ignition[937]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 17 00:15:52.716259 ignition[937]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 17 00:15:52.716259 ignition[937]: INFO : files: files passed Jan 17 00:15:52.716259 ignition[937]: INFO : Ignition finished successfully Jan 17 00:15:52.713921 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 17 00:15:52.734353 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 17 00:15:52.737386 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 17 00:15:52.741276 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 17 00:15:52.741421 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 17 00:15:52.766896 initrd-setup-root-after-ignition[967]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 17 00:15:52.766896 initrd-setup-root-after-ignition[967]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 17 00:15:52.770492 initrd-setup-root-after-ignition[971]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 17 00:15:52.773121 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 17 00:15:52.774444 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 17 00:15:52.780441 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 17 00:15:52.823465 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 17 00:15:52.823648 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 17 00:15:52.825817 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 17 00:15:52.827220 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 17 00:15:52.828010 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 17 00:15:52.834548 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 17 00:15:52.864467 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 17 00:15:52.873429 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 17 00:15:52.890733 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 17 00:15:52.891729 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 00:15:52.893149 systemd[1]: Stopped target timers.target - Timer Units. Jan 17 00:15:52.894584 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 17 00:15:52.894817 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 17 00:15:52.896454 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 17 00:15:52.898181 systemd[1]: Stopped target basic.target - Basic System. Jan 17 00:15:52.899419 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 17 00:15:52.900532 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 17 00:15:52.901853 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 17 00:15:52.903152 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 17 00:15:52.904390 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 17 00:15:52.905662 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 17 00:15:52.907060 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 17 00:15:52.908098 systemd[1]: Stopped target swap.target - Swaps. Jan 17 00:15:52.909177 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 17 00:15:52.909413 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 17 00:15:52.910857 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 17 00:15:52.912361 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 00:15:52.913430 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 17 00:15:52.914449 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 00:15:52.916071 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 17 00:15:52.916343 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 17 00:15:52.918017 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 17 00:15:52.918311 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 17 00:15:52.919446 systemd[1]: ignition-files.service: Deactivated successfully. Jan 17 00:15:52.919569 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 17 00:15:52.920596 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Jan 17 00:15:52.920761 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 17 00:15:52.935342 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 17 00:15:52.937411 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 17 00:15:52.938264 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 17 00:15:52.939289 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 00:15:52.941551 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 17 00:15:52.941921 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 17 00:15:52.949506 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 17 00:15:52.949628 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 17 00:15:52.968539 ignition[991]: INFO : Ignition 2.19.0 Jan 17 00:15:52.971274 ignition[991]: INFO : Stage: umount Jan 17 00:15:52.971274 ignition[991]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 00:15:52.971274 ignition[991]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Jan 17 00:15:52.975884 ignition[991]: INFO : umount: umount passed Jan 17 00:15:52.975884 ignition[991]: INFO : Ignition finished successfully Jan 17 00:15:52.976969 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 17 00:15:52.977118 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 17 00:15:52.981594 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 17 00:15:52.982192 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 17 00:15:52.982244 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 17 00:15:52.983858 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 17 00:15:52.983918 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 17 00:15:52.989837 systemd[1]: ignition-fetch.service: Deactivated successfully. Jan 17 00:15:52.989907 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jan 17 00:15:52.990875 systemd[1]: Stopped target network.target - Network. Jan 17 00:15:52.992026 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 17 00:15:52.992105 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 17 00:15:52.993120 systemd[1]: Stopped target paths.target - Path Units. Jan 17 00:15:52.994014 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 17 00:15:52.996212 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 00:15:52.997482 systemd[1]: Stopped target slices.target - Slice Units. Jan 17 00:15:52.998476 systemd[1]: Stopped target sockets.target - Socket Units. Jan 17 00:15:52.999664 systemd[1]: iscsid.socket: Deactivated successfully. Jan 17 00:15:52.999722 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 17 00:15:53.000950 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 17 00:15:53.001004 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 17 00:15:53.001966 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 17 00:15:53.002053 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 17 00:15:53.003245 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 17 00:15:53.003325 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 17 00:15:53.004449 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 17 00:15:53.005951 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 17 00:15:53.007534 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 17 00:15:53.007684 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 17 00:15:53.008214 systemd-networkd[746]: eth0: DHCPv6 lease lost Jan 17 00:15:53.009018 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 17 00:15:53.009160 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 17 00:15:53.013262 systemd-networkd[746]: eth1: DHCPv6 lease lost Jan 17 00:15:53.015854 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 17 00:15:53.016020 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 17 00:15:53.019375 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 17 00:15:53.019437 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 17 00:15:53.025426 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 17 00:15:53.026248 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 17 00:15:53.026328 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 17 00:15:53.027580 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 00:15:53.034484 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 17 00:15:53.034603 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 17 00:15:53.045600 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 17 00:15:53.045725 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 17 00:15:53.046856 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 17 00:15:53.046907 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 17 00:15:53.047839 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 17 00:15:53.047888 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 00:15:53.049437 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 17 00:15:53.049603 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 00:15:53.050605 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 17 00:15:53.050714 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 17 00:15:53.052399 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 17 00:15:53.052495 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 17 00:15:53.053754 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 17 00:15:53.053797 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 00:15:53.055569 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 17 00:15:53.055629 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 17 00:15:53.056949 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 17 00:15:53.057003 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 17 00:15:53.057983 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 17 00:15:53.058032 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 00:15:53.066506 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 17 00:15:53.067504 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 17 00:15:53.067578 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 00:15:53.068072 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 17 00:15:53.068115 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 00:15:53.068658 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 17 00:15:53.068707 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 00:15:53.069188 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 00:15:53.069227 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 00:15:53.078682 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 17 00:15:53.078841 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 17 00:15:53.080653 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 17 00:15:53.091573 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 17 00:15:53.103637 systemd[1]: Switching root. Jan 17 00:15:53.137801 systemd-journald[186]: Journal stopped Jan 17 00:15:54.431444 systemd-journald[186]: Received SIGTERM from PID 1 (systemd). Jan 17 00:15:54.431564 kernel: SELinux: policy capability network_peer_controls=1 Jan 17 00:15:54.431586 kernel: SELinux: policy capability open_perms=1 Jan 17 00:15:54.431608 kernel: SELinux: policy capability extended_socket_class=1 Jan 17 00:15:54.431629 kernel: SELinux: policy capability always_check_network=0 Jan 17 00:15:54.431645 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 17 00:15:54.431672 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 17 00:15:54.431701 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 17 00:15:54.431723 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 17 00:15:54.431743 kernel: audit: type=1403 audit(1768608953.306:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 17 00:15:54.431773 systemd[1]: Successfully loaded SELinux policy in 43.632ms. Jan 17 00:15:54.431814 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.768ms. Jan 17 00:15:54.431840 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 17 00:15:54.431855 systemd[1]: Detected virtualization kvm. Jan 17 00:15:54.431868 systemd[1]: Detected architecture x86-64. Jan 17 00:15:54.431884 systemd[1]: Detected first boot. Jan 17 00:15:54.431898 systemd[1]: Hostname set to . Jan 17 00:15:54.431917 systemd[1]: Initializing machine ID from VM UUID. Jan 17 00:15:54.431931 zram_generator::config[1035]: No configuration found. Jan 17 00:15:54.431946 systemd[1]: Populated /etc with preset unit settings. Jan 17 00:15:54.431959 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 17 00:15:54.431973 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 17 00:15:54.431985 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 17 00:15:54.432003 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 17 00:15:54.432018 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 17 00:15:54.432031 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 17 00:15:54.432045 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 17 00:15:54.432058 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 17 00:15:54.432071 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 17 00:15:54.432085 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 17 00:15:54.432098 systemd[1]: Created slice user.slice - User and Session Slice. Jan 17 00:15:54.432117 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 00:15:54.432242 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 00:15:54.432257 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 17 00:15:54.432272 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 17 00:15:54.432287 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 17 00:15:54.432301 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 17 00:15:54.432314 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 17 00:15:54.432328 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 00:15:54.432341 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 17 00:15:54.432359 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 17 00:15:54.432372 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 17 00:15:54.432385 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 17 00:15:54.432398 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 00:15:54.432412 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 17 00:15:54.432424 systemd[1]: Reached target slices.target - Slice Units. Jan 17 00:15:54.432437 systemd[1]: Reached target swap.target - Swaps. Jan 17 00:15:54.432453 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 17 00:15:54.432468 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 17 00:15:54.432481 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 17 00:15:54.432494 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 17 00:15:54.432509 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 00:15:54.432522 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 17 00:15:54.432541 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 17 00:15:54.432554 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 17 00:15:54.432569 systemd[1]: Mounting media.mount - External Media Directory... Jan 17 00:15:54.432585 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:54.432605 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 17 00:15:54.432619 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 17 00:15:54.432631 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 17 00:15:54.432645 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 17 00:15:54.432658 systemd[1]: Reached target machines.target - Containers. Jan 17 00:15:54.432672 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 17 00:15:54.432693 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 17 00:15:54.432717 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 17 00:15:54.432734 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 17 00:15:54.432753 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 17 00:15:54.432773 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 17 00:15:54.432793 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 17 00:15:54.432811 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 17 00:15:54.432830 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 17 00:15:54.432850 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 17 00:15:54.432876 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 17 00:15:54.432897 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 17 00:15:54.432918 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 17 00:15:54.432938 systemd[1]: Stopped systemd-fsck-usr.service. Jan 17 00:15:54.432960 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 17 00:15:54.432980 kernel: loop: module loaded Jan 17 00:15:54.433002 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 17 00:15:54.433022 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 17 00:15:54.433041 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 17 00:15:54.433064 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 17 00:15:54.433079 systemd[1]: verity-setup.service: Deactivated successfully. Jan 17 00:15:54.433092 systemd[1]: Stopped verity-setup.service. Jan 17 00:15:54.433107 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:54.433119 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 17 00:15:54.434185 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 17 00:15:54.434205 systemd[1]: Mounted media.mount - External Media Directory. Jan 17 00:15:54.434219 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 17 00:15:54.434240 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 17 00:15:54.434253 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 17 00:15:54.434267 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 00:15:54.434283 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 17 00:15:54.434297 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 17 00:15:54.434310 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 17 00:15:54.434323 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 17 00:15:54.434337 kernel: ACPI: bus type drm_connector registered Jan 17 00:15:54.434351 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 17 00:15:54.434365 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 17 00:15:54.434377 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 17 00:15:54.434394 kernel: fuse: init (API version 7.39) Jan 17 00:15:54.434406 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 17 00:15:54.434420 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 17 00:15:54.434433 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 17 00:15:54.434446 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 17 00:15:54.434459 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 17 00:15:54.434473 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 17 00:15:54.434486 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 17 00:15:54.434499 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 17 00:15:54.434515 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 17 00:15:54.434572 systemd-journald[1111]: Collecting audit messages is disabled. Jan 17 00:15:54.434599 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 17 00:15:54.434613 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 17 00:15:54.434629 systemd-journald[1111]: Journal started Jan 17 00:15:54.434663 systemd-journald[1111]: Runtime Journal (/run/log/journal/243af4b5561a4f9b93af0aee0d527157) is 4.9M, max 39.3M, 34.4M free. Jan 17 00:15:54.007282 systemd[1]: Queued start job for default target multi-user.target. Jan 17 00:15:54.025228 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jan 17 00:15:54.025774 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 17 00:15:54.440248 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 17 00:15:54.443354 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 17 00:15:54.446154 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 17 00:15:54.453821 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 17 00:15:54.459309 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 17 00:15:54.462227 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 00:15:54.467459 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 17 00:15:54.476866 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 17 00:15:54.486270 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 17 00:15:54.495464 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 17 00:15:54.504182 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 17 00:15:54.516346 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 17 00:15:54.530938 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 17 00:15:54.536529 systemd[1]: Started systemd-journald.service - Journal Service. Jan 17 00:15:54.539033 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 17 00:15:54.539866 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 17 00:15:54.541016 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 17 00:15:54.542491 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 17 00:15:54.544540 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 17 00:15:54.554233 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 00:15:54.591370 kernel: loop0: detected capacity change from 0 to 142488 Jan 17 00:15:54.594568 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 17 00:15:54.605467 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 17 00:15:54.617389 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 17 00:15:54.626417 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 17 00:15:54.642673 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 17 00:15:54.641712 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 17 00:15:54.665873 systemd-tmpfiles[1138]: ACLs are not supported, ignoring. Jan 17 00:15:54.665893 systemd-tmpfiles[1138]: ACLs are not supported, ignoring. Jan 17 00:15:54.675305 systemd-journald[1111]: Time spent on flushing to /var/log/journal/243af4b5561a4f9b93af0aee0d527157 is 92.855ms for 997 entries. Jan 17 00:15:54.675305 systemd-journald[1111]: System Journal (/var/log/journal/243af4b5561a4f9b93af0aee0d527157) is 8.0M, max 195.6M, 187.6M free. Jan 17 00:15:54.818476 systemd-journald[1111]: Received client request to flush runtime journal. Jan 17 00:15:54.818549 kernel: loop1: detected capacity change from 0 to 224512 Jan 17 00:15:54.818575 kernel: loop2: detected capacity change from 0 to 140768 Jan 17 00:15:54.818596 kernel: loop3: detected capacity change from 0 to 8 Jan 17 00:15:54.675710 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 17 00:15:54.679070 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 17 00:15:54.712742 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 00:15:54.725575 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 17 00:15:54.738356 udevadm[1166]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jan 17 00:15:54.799238 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 17 00:15:54.808553 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 17 00:15:54.827868 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 17 00:15:54.859172 kernel: loop4: detected capacity change from 0 to 142488 Jan 17 00:15:54.859890 systemd-tmpfiles[1178]: ACLs are not supported, ignoring. Jan 17 00:15:54.859910 systemd-tmpfiles[1178]: ACLs are not supported, ignoring. Jan 17 00:15:54.872185 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 00:15:54.881190 kernel: loop5: detected capacity change from 0 to 224512 Jan 17 00:15:54.901301 kernel: loop6: detected capacity change from 0 to 140768 Jan 17 00:15:54.928165 kernel: loop7: detected capacity change from 0 to 8 Jan 17 00:15:54.930636 (sd-merge)[1184]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. Jan 17 00:15:54.932377 (sd-merge)[1184]: Merged extensions into '/usr'. Jan 17 00:15:54.955522 systemd[1]: Reloading requested from client PID 1137 ('systemd-sysext') (unit systemd-sysext.service)... Jan 17 00:15:54.955576 systemd[1]: Reloading... Jan 17 00:15:55.186170 zram_generator::config[1211]: No configuration found. Jan 17 00:15:55.213846 ldconfig[1133]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 17 00:15:55.321652 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 00:15:55.378546 systemd[1]: Reloading finished in 422 ms. Jan 17 00:15:55.406365 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 17 00:15:55.408755 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 17 00:15:55.422507 systemd[1]: Starting ensure-sysext.service... Jan 17 00:15:55.434271 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 17 00:15:55.459955 systemd[1]: Reloading requested from client PID 1254 ('systemctl') (unit ensure-sysext.service)... Jan 17 00:15:55.460175 systemd[1]: Reloading... Jan 17 00:15:55.492117 systemd-tmpfiles[1255]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 17 00:15:55.493094 systemd-tmpfiles[1255]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 17 00:15:55.496734 systemd-tmpfiles[1255]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 17 00:15:55.499669 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. Jan 17 00:15:55.499756 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. Jan 17 00:15:55.506516 systemd-tmpfiles[1255]: Detected autofs mount point /boot during canonicalization of boot. Jan 17 00:15:55.506705 systemd-tmpfiles[1255]: Skipping /boot Jan 17 00:15:55.525431 systemd-tmpfiles[1255]: Detected autofs mount point /boot during canonicalization of boot. Jan 17 00:15:55.526340 systemd-tmpfiles[1255]: Skipping /boot Jan 17 00:15:55.631182 zram_generator::config[1291]: No configuration found. Jan 17 00:15:55.762106 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 00:15:55.818924 systemd[1]: Reloading finished in 358 ms. Jan 17 00:15:55.839189 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 17 00:15:55.844845 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 00:15:55.858616 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 17 00:15:55.863608 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 17 00:15:55.873379 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 17 00:15:55.879422 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 17 00:15:55.882666 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 00:15:55.888219 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 17 00:15:55.898061 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:55.898640 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 17 00:15:55.902575 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 17 00:15:55.906496 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 17 00:15:55.917274 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 17 00:15:55.919381 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 00:15:55.919552 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:55.923766 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:55.923963 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 17 00:15:55.924773 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 00:15:55.924909 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:55.929303 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:55.929584 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 17 00:15:55.941538 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 17 00:15:55.942323 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 00:15:55.942492 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:55.946703 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 17 00:15:55.950592 systemd[1]: Finished ensure-sysext.service. Jan 17 00:15:55.959526 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 17 00:15:55.991947 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 17 00:15:56.014082 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 17 00:15:56.014368 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 17 00:15:56.015711 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 17 00:15:56.016227 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 17 00:15:56.028830 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 17 00:15:56.029110 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 17 00:15:56.031740 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 17 00:15:56.033030 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 17 00:15:56.033308 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 17 00:15:56.035511 augenrules[1358]: No rules Jan 17 00:15:56.035465 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 17 00:15:56.038693 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 17 00:15:56.044844 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 17 00:15:56.050700 systemd-udevd[1338]: Using default interface naming scheme 'v255'. Jan 17 00:15:56.056466 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 17 00:15:56.066016 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 17 00:15:56.072750 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 17 00:15:56.096575 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 17 00:15:56.102305 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 17 00:15:56.110585 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 00:15:56.120468 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 17 00:15:56.276061 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 17 00:15:56.324228 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 17 00:15:56.325511 systemd[1]: Reached target time-set.target - System Time Set. Jan 17 00:15:56.338523 systemd-networkd[1378]: lo: Link UP Jan 17 00:15:56.339387 systemd-networkd[1378]: lo: Gained carrier Jan 17 00:15:56.340584 systemd-networkd[1378]: Enumeration completed Jan 17 00:15:56.343959 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 17 00:15:56.344180 systemd-timesyncd[1351]: No network connectivity, watching for changes. Jan 17 00:15:56.344453 systemd-resolved[1334]: Positive Trust Anchors: Jan 17 00:15:56.345144 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (1384) Jan 17 00:15:56.344463 systemd-resolved[1334]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 17 00:15:56.344501 systemd-resolved[1334]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 17 00:15:56.350276 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Jan 17 00:15:56.350795 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:56.350964 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 17 00:15:56.352238 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 17 00:15:56.358396 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 17 00:15:56.359264 systemd-resolved[1334]: Using system hostname 'ci-4081.3.6-n-09ad808b32'. Jan 17 00:15:56.360605 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 17 00:15:56.361212 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 00:15:56.363923 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 17 00:15:56.364492 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 17 00:15:56.364524 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 00:15:56.368407 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 17 00:15:56.369021 systemd[1]: Reached target network.target - Network. Jan 17 00:15:56.370489 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 17 00:15:56.396147 kernel: ISO 9660 Extensions: RRIP_1991A Jan 17 00:15:56.401734 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Jan 17 00:15:56.403658 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 17 00:15:56.403853 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 17 00:15:56.409379 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 17 00:15:56.409610 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 17 00:15:56.413994 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 17 00:15:56.417195 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 17 00:15:56.417707 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 17 00:15:56.419156 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 17 00:15:56.425155 kernel: ACPI: button: Power Button [PWRF] Jan 17 00:15:56.428132 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 17 00:15:56.449154 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Jan 17 00:15:56.511239 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Jan 17 00:15:56.520512 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Jan 17 00:15:56.536027 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jan 17 00:15:56.539165 kernel: Console: switching to colour dummy device 80x25 Jan 17 00:15:56.541099 systemd-networkd[1378]: eth1: Configuring with /run/systemd/network/10-b2:c1:1b:5b:12:11.network. Jan 17 00:15:56.544406 systemd-networkd[1378]: eth1: Link UP Jan 17 00:15:56.544666 systemd-networkd[1378]: eth1: Gained carrier Jan 17 00:15:56.551168 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Jan 17 00:15:56.554769 systemd-timesyncd[1351]: Network configuration changed, trying to establish connection. Jan 17 00:15:56.565483 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Jan 17 00:15:56.565599 kernel: [drm] features: -context_init Jan 17 00:15:56.569611 systemd-networkd[1378]: eth0: Configuring with /run/systemd/network/10-9e:f2:89:ad:90:4d.network. Jan 17 00:15:56.571516 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 17 00:15:56.574454 systemd-networkd[1378]: eth0: Link UP Jan 17 00:15:56.574471 systemd-networkd[1378]: eth0: Gained carrier Jan 17 00:15:56.613635 kernel: [drm] number of scanouts: 1 Jan 17 00:15:56.613746 kernel: [drm] number of cap sets: 0 Jan 17 00:15:56.616236 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 Jan 17 00:15:56.624820 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 17 00:15:56.636044 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Jan 17 00:15:56.636170 kernel: Console: switching to colour frame buffer device 128x48 Jan 17 00:15:56.639741 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 00:15:56.642178 kernel: mousedev: PS/2 mouse device common for all mice Jan 17 00:15:56.657167 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Jan 17 00:15:56.705689 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 00:15:56.706187 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 00:15:56.727202 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 00:15:56.733653 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 00:15:56.733877 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 00:15:56.741520 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 00:15:56.815903 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 00:15:56.849204 kernel: EDAC MC: Ver: 3.0.0 Jan 17 00:15:56.874511 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 17 00:15:56.881492 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 17 00:15:56.905711 lvm[1435]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 17 00:15:56.936837 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 17 00:15:56.939258 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 17 00:15:56.939391 systemd[1]: Reached target sysinit.target - System Initialization. Jan 17 00:15:56.940580 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 17 00:15:56.940765 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 17 00:15:56.941101 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 17 00:15:56.941294 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 17 00:15:56.941365 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 17 00:15:56.941422 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 17 00:15:56.941460 systemd[1]: Reached target paths.target - Path Units. Jan 17 00:15:56.941534 systemd[1]: Reached target timers.target - Timer Units. Jan 17 00:15:56.942858 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 17 00:15:56.947447 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 17 00:15:56.955142 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 17 00:15:56.958071 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 17 00:15:56.962255 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 17 00:15:56.963377 systemd[1]: Reached target sockets.target - Socket Units. Jan 17 00:15:56.966543 systemd[1]: Reached target basic.target - Basic System. Jan 17 00:15:56.967301 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 17 00:15:56.967340 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 17 00:15:56.970224 systemd[1]: Starting containerd.service - containerd container runtime... Jan 17 00:15:56.971551 lvm[1439]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 17 00:15:56.986662 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jan 17 00:15:56.993387 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 17 00:15:56.997991 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 17 00:15:57.010618 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 17 00:15:57.013772 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 17 00:15:57.017349 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 17 00:15:57.018323 jq[1443]: false Jan 17 00:15:57.024341 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 17 00:15:57.034393 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 17 00:15:57.039486 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 17 00:15:57.055411 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 17 00:15:57.057818 extend-filesystems[1444]: Found loop4 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found loop5 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found loop6 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found loop7 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda1 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda2 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda3 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found usr Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda4 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda6 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda7 Jan 17 00:15:57.064218 extend-filesystems[1444]: Found vda9 Jan 17 00:15:57.064218 extend-filesystems[1444]: Checking size of /dev/vda9 Jan 17 00:15:57.129792 coreos-metadata[1441]: Jan 17 00:15:57.120 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 17 00:15:57.156237 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Jan 17 00:15:57.156299 extend-filesystems[1444]: Resized partition /dev/vda9 Jan 17 00:15:57.104057 dbus-daemon[1442]: [system] SELinux support is enabled Jan 17 00:15:57.066036 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 17 00:15:57.160778 coreos-metadata[1441]: Jan 17 00:15:57.148 INFO Fetch successful Jan 17 00:15:57.160818 extend-filesystems[1462]: resize2fs 1.47.1 (20-May-2024) Jan 17 00:15:57.067547 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 17 00:15:57.070414 systemd[1]: Starting update-engine.service - Update Engine... Jan 17 00:15:57.092327 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 17 00:15:57.106497 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 17 00:15:57.204601 jq[1456]: true Jan 17 00:15:57.115726 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 17 00:15:57.133764 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 17 00:15:57.134028 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 17 00:15:57.143493 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 17 00:15:57.144436 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 17 00:15:57.191759 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 17 00:15:57.191803 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 17 00:15:57.200814 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 17 00:15:57.200992 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Jan 17 00:15:57.201029 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 17 00:15:57.210372 systemd-logind[1452]: New seat seat0. Jan 17 00:15:57.218181 systemd[1]: motdgen.service: Deactivated successfully. Jan 17 00:15:57.218820 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 17 00:15:57.221605 systemd-logind[1452]: Watching system buttons on /dev/input/event1 (Power Button) Jan 17 00:15:57.221641 systemd-logind[1452]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 17 00:15:57.222049 systemd[1]: Started systemd-logind.service - User Login Management. Jan 17 00:15:57.253753 tar[1466]: linux-amd64/LICENSE Jan 17 00:15:57.261151 tar[1466]: linux-amd64/helm Jan 17 00:15:57.280542 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Jan 17 00:15:57.282832 jq[1477]: true Jan 17 00:15:57.294292 extend-filesystems[1462]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jan 17 00:15:57.294292 extend-filesystems[1462]: old_desc_blocks = 1, new_desc_blocks = 8 Jan 17 00:15:57.294292 extend-filesystems[1462]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Jan 17 00:15:57.322744 extend-filesystems[1444]: Resized filesystem in /dev/vda9 Jan 17 00:15:57.322744 extend-filesystems[1444]: Found vdb Jan 17 00:15:57.337757 update_engine[1455]: I20260117 00:15:57.316476 1455 main.cc:92] Flatcar Update Engine starting Jan 17 00:15:57.304719 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 17 00:15:57.304969 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 17 00:15:57.344270 update_engine[1455]: I20260117 00:15:57.338832 1455 update_check_scheduler.cc:74] Next update check in 4m45s Jan 17 00:15:57.339197 systemd[1]: Started update-engine.service - Update Engine. Jan 17 00:15:57.349474 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 17 00:15:57.353311 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jan 17 00:15:57.358408 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 17 00:15:57.370773 (ntainerd)[1483]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 17 00:15:57.490102 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (1389) Jan 17 00:15:57.528158 bash[1506]: Updated "/home/core/.ssh/authorized_keys" Jan 17 00:15:57.534317 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 17 00:15:57.587776 systemd[1]: Starting sshkeys.service... Jan 17 00:15:57.678962 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jan 17 00:15:57.694494 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jan 17 00:15:57.731416 systemd-networkd[1378]: eth0: Gained IPv6LL Jan 17 00:15:57.739232 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 17 00:15:57.743686 systemd[1]: Reached target network-online.target - Network is Online. Jan 17 00:15:57.757704 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 00:15:57.773253 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 17 00:15:57.849069 locksmithd[1490]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 17 00:15:57.853202 coreos-metadata[1516]: Jan 17 00:15:57.851 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Jan 17 00:15:57.859173 coreos-metadata[1516]: Jan 17 00:15:57.858 INFO Fetch successful Jan 17 00:15:57.885344 unknown[1516]: wrote ssh authorized keys file for user: core Jan 17 00:15:57.935336 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 17 00:15:57.945223 containerd[1483]: time="2026-01-17T00:15:57.944603669Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Jan 17 00:15:57.972860 update-ssh-keys[1529]: Updated "/home/core/.ssh/authorized_keys" Jan 17 00:15:57.974720 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jan 17 00:15:57.977023 systemd-networkd[1378]: eth1: Gained IPv6LL Jan 17 00:15:57.982733 systemd[1]: Finished sshkeys.service. Jan 17 00:15:58.046182 containerd[1483]: time="2026-01-17T00:15:58.044804437Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 17 00:15:58.049503 containerd[1483]: time="2026-01-17T00:15:58.047749668Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.119-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 17 00:15:58.049503 containerd[1483]: time="2026-01-17T00:15:58.047838414Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 17 00:15:58.049503 containerd[1483]: time="2026-01-17T00:15:58.047874467Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 17 00:15:58.049503 containerd[1483]: time="2026-01-17T00:15:58.048104650Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 17 00:15:58.049503 containerd[1483]: time="2026-01-17T00:15:58.048140312Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 17 00:15:58.050813 containerd[1483]: time="2026-01-17T00:15:58.050276823Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 00:15:58.050813 containerd[1483]: time="2026-01-17T00:15:58.050325714Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 17 00:15:58.050813 containerd[1483]: time="2026-01-17T00:15:58.050628891Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 00:15:58.050813 containerd[1483]: time="2026-01-17T00:15:58.050646349Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 17 00:15:58.050813 containerd[1483]: time="2026-01-17T00:15:58.050663425Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 00:15:58.050813 containerd[1483]: time="2026-01-17T00:15:58.050676441Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 17 00:15:58.050813 containerd[1483]: time="2026-01-17T00:15:58.050786437Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 17 00:15:58.054691 containerd[1483]: time="2026-01-17T00:15:58.054230769Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 17 00:15:58.054691 containerd[1483]: time="2026-01-17T00:15:58.054565378Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 00:15:58.054691 containerd[1483]: time="2026-01-17T00:15:58.054592494Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 17 00:15:58.054962 containerd[1483]: time="2026-01-17T00:15:58.054797189Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 17 00:15:58.054962 containerd[1483]: time="2026-01-17T00:15:58.054896948Z" level=info msg="metadata content store policy set" policy=shared Jan 17 00:15:58.067275 containerd[1483]: time="2026-01-17T00:15:58.066514256Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 17 00:15:58.067275 containerd[1483]: time="2026-01-17T00:15:58.066626200Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 17 00:15:58.067275 containerd[1483]: time="2026-01-17T00:15:58.066657377Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 17 00:15:58.067275 containerd[1483]: time="2026-01-17T00:15:58.066679772Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 17 00:15:58.067275 containerd[1483]: time="2026-01-17T00:15:58.066720445Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 17 00:15:58.067275 containerd[1483]: time="2026-01-17T00:15:58.066994302Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 17 00:15:58.069830 containerd[1483]: time="2026-01-17T00:15:58.069395782Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 17 00:15:58.070761 containerd[1483]: time="2026-01-17T00:15:58.070694993Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 17 00:15:58.070761 containerd[1483]: time="2026-01-17T00:15:58.070763990Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 17 00:15:58.070892 containerd[1483]: time="2026-01-17T00:15:58.070804922Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 17 00:15:58.070892 containerd[1483]: time="2026-01-17T00:15:58.070833614Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.070892 containerd[1483]: time="2026-01-17T00:15:58.070856063Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.070892 containerd[1483]: time="2026-01-17T00:15:58.070885978Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.070905815Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072269487Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072305541Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072346566Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072367596Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072418825Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072446932Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072472096Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072518448Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072540567Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072578901Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072601312Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072625354Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.072668 containerd[1483]: time="2026-01-17T00:15:58.072660126Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074169656Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074224058Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074260910Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074278684Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074300202Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074346408Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074364876Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.074447 containerd[1483]: time="2026-01-17T00:15:58.074377970Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 17 00:15:58.074750 containerd[1483]: time="2026-01-17T00:15:58.074512661Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 17 00:15:58.074750 containerd[1483]: time="2026-01-17T00:15:58.074655397Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 17 00:15:58.074750 containerd[1483]: time="2026-01-17T00:15:58.074671599Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 17 00:15:58.074750 containerd[1483]: time="2026-01-17T00:15:58.074687587Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 17 00:15:58.076969 containerd[1483]: time="2026-01-17T00:15:58.076394500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.082601 containerd[1483]: time="2026-01-17T00:15:58.080085728Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 17 00:15:58.082601 containerd[1483]: time="2026-01-17T00:15:58.080170683Z" level=info msg="NRI interface is disabled by configuration." Jan 17 00:15:58.082601 containerd[1483]: time="2026-01-17T00:15:58.080189537Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 17 00:15:58.082861 containerd[1483]: time="2026-01-17T00:15:58.080573384Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 17 00:15:58.082861 containerd[1483]: time="2026-01-17T00:15:58.080702432Z" level=info msg="Connect containerd service" Jan 17 00:15:58.082861 containerd[1483]: time="2026-01-17T00:15:58.080775963Z" level=info msg="using legacy CRI server" Jan 17 00:15:58.082861 containerd[1483]: time="2026-01-17T00:15:58.080785353Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 17 00:15:58.082861 containerd[1483]: time="2026-01-17T00:15:58.080947527Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 17 00:15:58.082861 containerd[1483]: time="2026-01-17T00:15:58.082779871Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.084552887Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.084669514Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.084795467Z" level=info msg="Start subscribing containerd event" Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.084868525Z" level=info msg="Start recovering state" Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.084999500Z" level=info msg="Start event monitor" Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.085020296Z" level=info msg="Start snapshots syncer" Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.085037394Z" level=info msg="Start cni network conf syncer for default" Jan 17 00:15:58.086218 containerd[1483]: time="2026-01-17T00:15:58.085060448Z" level=info msg="Start streaming server" Jan 17 00:15:58.087440 systemd[1]: Started containerd.service - containerd container runtime. Jan 17 00:15:58.088667 containerd[1483]: time="2026-01-17T00:15:58.087708288Z" level=info msg="containerd successfully booted in 0.144797s" Jan 17 00:15:58.131419 sshd_keygen[1480]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 17 00:15:58.218491 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 17 00:15:58.236590 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 17 00:15:58.255857 systemd[1]: issuegen.service: Deactivated successfully. Jan 17 00:15:58.256192 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 17 00:15:58.270697 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 17 00:15:58.328146 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 17 00:15:58.339321 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 17 00:15:58.346653 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 17 00:15:58.349213 systemd[1]: Reached target getty.target - Login Prompts. Jan 17 00:15:58.623547 tar[1466]: linux-amd64/README.md Jan 17 00:15:58.652367 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 17 00:15:58.703760 kernel: hrtimer: interrupt took 5993780 ns Jan 17 00:15:59.443851 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:15:59.446635 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 17 00:15:59.449958 systemd[1]: Startup finished in 1.323s (kernel) + 6.582s (initrd) + 6.185s (userspace) = 14.091s. Jan 17 00:15:59.453147 (kubelet)[1563]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 00:16:00.239809 kubelet[1563]: E0117 00:16:00.239709 1563 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 00:16:00.243373 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 00:16:00.243575 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 00:16:00.243940 systemd[1]: kubelet.service: Consumed 1.501s CPU time. Jan 17 00:16:01.002565 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 17 00:16:01.007594 systemd[1]: Started sshd@0-143.198.57.57:22-4.153.228.146:48906.service - OpenSSH per-connection server daemon (4.153.228.146:48906). Jan 17 00:16:01.482326 sshd[1574]: Accepted publickey for core from 4.153.228.146 port 48906 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:16:01.485668 sshd[1574]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:16:01.501214 systemd-logind[1452]: New session 1 of user core. Jan 17 00:16:01.502217 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 17 00:16:01.515600 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 17 00:16:01.537773 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 17 00:16:01.544733 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 17 00:16:01.558953 (systemd)[1578]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 17 00:16:01.720978 systemd[1578]: Queued start job for default target default.target. Jan 17 00:16:01.732391 systemd[1578]: Created slice app.slice - User Application Slice. Jan 17 00:16:01.732449 systemd[1578]: Reached target paths.target - Paths. Jan 17 00:16:01.732474 systemd[1578]: Reached target timers.target - Timers. Jan 17 00:16:01.734813 systemd[1578]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 17 00:16:01.754062 systemd[1578]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 17 00:16:01.754336 systemd[1578]: Reached target sockets.target - Sockets. Jan 17 00:16:01.754377 systemd[1578]: Reached target basic.target - Basic System. Jan 17 00:16:01.754582 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 17 00:16:01.755598 systemd[1578]: Reached target default.target - Main User Target. Jan 17 00:16:01.755686 systemd[1578]: Startup finished in 187ms. Jan 17 00:16:01.759456 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 17 00:16:02.089742 systemd[1]: Started sshd@1-143.198.57.57:22-4.153.228.146:48914.service - OpenSSH per-connection server daemon (4.153.228.146:48914). Jan 17 00:16:02.498393 sshd[1589]: Accepted publickey for core from 4.153.228.146 port 48914 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:16:02.500605 sshd[1589]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:16:02.508477 systemd-logind[1452]: New session 2 of user core. Jan 17 00:16:02.518822 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 17 00:16:02.789483 sshd[1589]: pam_unix(sshd:session): session closed for user core Jan 17 00:16:02.795120 systemd[1]: sshd@1-143.198.57.57:22-4.153.228.146:48914.service: Deactivated successfully. Jan 17 00:16:02.797459 systemd[1]: session-2.scope: Deactivated successfully. Jan 17 00:16:02.798395 systemd-logind[1452]: Session 2 logged out. Waiting for processes to exit. Jan 17 00:16:02.799956 systemd-logind[1452]: Removed session 2. Jan 17 00:16:02.882634 systemd[1]: Started sshd@2-143.198.57.57:22-4.153.228.146:48926.service - OpenSSH per-connection server daemon (4.153.228.146:48926). Jan 17 00:16:02.935108 systemd-timesyncd[1351]: Contacted time server 204.2.134.173:123 (1.flatcar.pool.ntp.org). Jan 17 00:16:02.935243 systemd-timesyncd[1351]: Initial clock synchronization to Sat 2026-01-17 00:16:03.069252 UTC. Jan 17 00:16:03.329956 sshd[1596]: Accepted publickey for core from 4.153.228.146 port 48926 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:16:03.332008 sshd[1596]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:16:03.338276 systemd-logind[1452]: New session 3 of user core. Jan 17 00:16:03.345452 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 17 00:16:03.644460 sshd[1596]: pam_unix(sshd:session): session closed for user core Jan 17 00:16:03.649066 systemd[1]: sshd@2-143.198.57.57:22-4.153.228.146:48926.service: Deactivated successfully. Jan 17 00:16:03.651767 systemd[1]: session-3.scope: Deactivated successfully. Jan 17 00:16:03.652452 systemd-logind[1452]: Session 3 logged out. Waiting for processes to exit. Jan 17 00:16:03.654268 systemd-logind[1452]: Removed session 3. Jan 17 00:16:03.731393 systemd[1]: Started sshd@3-143.198.57.57:22-4.153.228.146:48928.service - OpenSSH per-connection server daemon (4.153.228.146:48928). Jan 17 00:16:04.192548 sshd[1603]: Accepted publickey for core from 4.153.228.146 port 48928 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:16:04.195306 sshd[1603]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:16:04.203954 systemd-logind[1452]: New session 4 of user core. Jan 17 00:16:04.210549 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 17 00:16:04.525068 sshd[1603]: pam_unix(sshd:session): session closed for user core Jan 17 00:16:04.532348 systemd[1]: sshd@3-143.198.57.57:22-4.153.228.146:48928.service: Deactivated successfully. Jan 17 00:16:04.535396 systemd[1]: session-4.scope: Deactivated successfully. Jan 17 00:16:04.537245 systemd-logind[1452]: Session 4 logged out. Waiting for processes to exit. Jan 17 00:16:04.539430 systemd-logind[1452]: Removed session 4. Jan 17 00:16:04.622654 systemd[1]: Started sshd@4-143.198.57.57:22-4.153.228.146:51854.service - OpenSSH per-connection server daemon (4.153.228.146:51854). Jan 17 00:16:05.094833 sshd[1610]: Accepted publickey for core from 4.153.228.146 port 51854 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:16:05.097481 sshd[1610]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:16:05.105095 systemd-logind[1452]: New session 5 of user core. Jan 17 00:16:05.111947 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 17 00:16:05.374934 sudo[1613]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 17 00:16:05.376188 sudo[1613]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 17 00:16:05.861608 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 17 00:16:05.861735 (dockerd)[1629]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 17 00:16:06.392556 dockerd[1629]: time="2026-01-17T00:16:06.392477858Z" level=info msg="Starting up" Jan 17 00:16:06.555349 dockerd[1629]: time="2026-01-17T00:16:06.555265087Z" level=info msg="Loading containers: start." Jan 17 00:16:06.715177 kernel: Initializing XFRM netlink socket Jan 17 00:16:06.829055 systemd-networkd[1378]: docker0: Link UP Jan 17 00:16:06.855026 dockerd[1629]: time="2026-01-17T00:16:06.854968773Z" level=info msg="Loading containers: done." Jan 17 00:16:06.875554 dockerd[1629]: time="2026-01-17T00:16:06.875475674Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 17 00:16:06.875826 dockerd[1629]: time="2026-01-17T00:16:06.875641934Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Jan 17 00:16:06.875826 dockerd[1629]: time="2026-01-17T00:16:06.875805473Z" level=info msg="Daemon has completed initialization" Jan 17 00:16:06.879460 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1500703020-merged.mount: Deactivated successfully. Jan 17 00:16:06.922667 dockerd[1629]: time="2026-01-17T00:16:06.921905540Z" level=info msg="API listen on /run/docker.sock" Jan 17 00:16:06.922257 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 17 00:16:08.054168 containerd[1483]: time="2026-01-17T00:16:08.053177382Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.11\"" Jan 17 00:16:08.767021 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount739216270.mount: Deactivated successfully. Jan 17 00:16:08.855517 systemd-resolved[1334]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Jan 17 00:16:10.236179 containerd[1483]: time="2026-01-17T00:16:10.235804240Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:10.238613 containerd[1483]: time="2026-01-17T00:16:10.238532464Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.11: active requests=0, bytes read=29070647" Jan 17 00:16:10.239362 containerd[1483]: time="2026-01-17T00:16:10.239311829Z" level=info msg="ImageCreate event name:\"sha256:7757c58248a29fc7474a8072796848689852b0477adf16765f38b3d1a9bacadf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:10.243605 containerd[1483]: time="2026-01-17T00:16:10.243419327Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:41eaecaed9af0ca8ab36d7794819c7df199e68c6c6ee0649114d713c495f8bd5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:10.244832 containerd[1483]: time="2026-01-17T00:16:10.244203065Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.11\" with image id \"sha256:7757c58248a29fc7474a8072796848689852b0477adf16765f38b3d1a9bacadf\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:41eaecaed9af0ca8ab36d7794819c7df199e68c6c6ee0649114d713c495f8bd5\", size \"29067246\" in 2.190932701s" Jan 17 00:16:10.244832 containerd[1483]: time="2026-01-17T00:16:10.244253798Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.11\" returns image reference \"sha256:7757c58248a29fc7474a8072796848689852b0477adf16765f38b3d1a9bacadf\"" Jan 17 00:16:10.245357 containerd[1483]: time="2026-01-17T00:16:10.245322904Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.11\"" Jan 17 00:16:10.493876 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 17 00:16:10.501514 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 00:16:10.694507 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:16:10.697724 (kubelet)[1842]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 00:16:10.779583 kubelet[1842]: E0117 00:16:10.779400 1842 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 00:16:10.785126 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 00:16:10.785383 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 00:16:11.898435 containerd[1483]: time="2026-01-17T00:16:11.897021385Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:11.898435 containerd[1483]: time="2026-01-17T00:16:11.898092148Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.11: active requests=0, bytes read=24993354" Jan 17 00:16:11.898435 containerd[1483]: time="2026-01-17T00:16:11.898367260Z" level=info msg="ImageCreate event name:\"sha256:0175d0a8243db520e3caa6d5c1e4248fddbc32447a9e8b5f4630831bc1e2489e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:11.903016 containerd[1483]: time="2026-01-17T00:16:11.902943533Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:ce7b2ead5eef1a1554ef28b2b79596c6a8c6d506a87a7ab1381e77fe3d72f55f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:11.905016 containerd[1483]: time="2026-01-17T00:16:11.904937496Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.11\" with image id \"sha256:0175d0a8243db520e3caa6d5c1e4248fddbc32447a9e8b5f4630831bc1e2489e\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:ce7b2ead5eef1a1554ef28b2b79596c6a8c6d506a87a7ab1381e77fe3d72f55f\", size \"26650388\" in 1.65957088s" Jan 17 00:16:11.905016 containerd[1483]: time="2026-01-17T00:16:11.905000638Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.11\" returns image reference \"sha256:0175d0a8243db520e3caa6d5c1e4248fddbc32447a9e8b5f4630831bc1e2489e\"" Jan 17 00:16:11.905804 containerd[1483]: time="2026-01-17T00:16:11.905730923Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.11\"" Jan 17 00:16:11.949920 systemd-resolved[1334]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. Jan 17 00:16:13.604578 containerd[1483]: time="2026-01-17T00:16:13.604517263Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:13.606583 containerd[1483]: time="2026-01-17T00:16:13.606504699Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.11: active requests=0, bytes read=19405076" Jan 17 00:16:13.606942 containerd[1483]: time="2026-01-17T00:16:13.606893569Z" level=info msg="ImageCreate event name:\"sha256:23d6a1fb92fda53b787f364351c610e55f073e8bdf0de5831974df7875b13f21\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:13.609930 containerd[1483]: time="2026-01-17T00:16:13.609866415Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:b3039587bbe70e61a6aeaff56c21fdeeef104524a31f835bcc80887d40b8e6b2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:13.612574 containerd[1483]: time="2026-01-17T00:16:13.611866203Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.11\" with image id \"sha256:23d6a1fb92fda53b787f364351c610e55f073e8bdf0de5831974df7875b13f21\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:b3039587bbe70e61a6aeaff56c21fdeeef104524a31f835bcc80887d40b8e6b2\", size \"21062128\" in 1.706080874s" Jan 17 00:16:13.612574 containerd[1483]: time="2026-01-17T00:16:13.611930463Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.11\" returns image reference \"sha256:23d6a1fb92fda53b787f364351c610e55f073e8bdf0de5831974df7875b13f21\"" Jan 17 00:16:13.612749 containerd[1483]: time="2026-01-17T00:16:13.612575770Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.11\"" Jan 17 00:16:14.888046 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount879674336.mount: Deactivated successfully. Jan 17 00:16:15.529108 containerd[1483]: time="2026-01-17T00:16:15.529016983Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:15.531043 containerd[1483]: time="2026-01-17T00:16:15.530968337Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.11: active requests=0, bytes read=31161899" Jan 17 00:16:15.531797 containerd[1483]: time="2026-01-17T00:16:15.531745154Z" level=info msg="ImageCreate event name:\"sha256:4d8fb2dc5751966f058943ff7c5f10551e603d726ab8648c7c7b7f95a2663e3d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:15.533432 containerd[1483]: time="2026-01-17T00:16:15.533377669Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:4204f9136c23a867929d32046032fe069b49ad94cf168042405e7d0ec88bdba9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:15.534333 containerd[1483]: time="2026-01-17T00:16:15.534282745Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.11\" with image id \"sha256:4d8fb2dc5751966f058943ff7c5f10551e603d726ab8648c7c7b7f95a2663e3d\", repo tag \"registry.k8s.io/kube-proxy:v1.32.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:4204f9136c23a867929d32046032fe069b49ad94cf168042405e7d0ec88bdba9\", size \"31160918\" in 1.921670554s" Jan 17 00:16:15.534591 containerd[1483]: time="2026-01-17T00:16:15.534474116Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.11\" returns image reference \"sha256:4d8fb2dc5751966f058943ff7c5f10551e603d726ab8648c7c7b7f95a2663e3d\"" Jan 17 00:16:15.535304 containerd[1483]: time="2026-01-17T00:16:15.535269233Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Jan 17 00:16:16.212478 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2947113038.mount: Deactivated successfully. Jan 17 00:16:17.190475 containerd[1483]: time="2026-01-17T00:16:17.190408278Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:17.192102 containerd[1483]: time="2026-01-17T00:16:17.191746881Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Jan 17 00:16:17.194215 containerd[1483]: time="2026-01-17T00:16:17.193091637Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:17.196199 containerd[1483]: time="2026-01-17T00:16:17.196112714Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:17.198016 containerd[1483]: time="2026-01-17T00:16:17.197516663Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.662206236s" Jan 17 00:16:17.198016 containerd[1483]: time="2026-01-17T00:16:17.197568240Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Jan 17 00:16:17.198410 containerd[1483]: time="2026-01-17T00:16:17.198386216Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jan 17 00:16:17.734308 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2114630089.mount: Deactivated successfully. Jan 17 00:16:17.740160 containerd[1483]: time="2026-01-17T00:16:17.739277754Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:17.741000 containerd[1483]: time="2026-01-17T00:16:17.740954063Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Jan 17 00:16:17.742017 containerd[1483]: time="2026-01-17T00:16:17.741984475Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:17.744274 containerd[1483]: time="2026-01-17T00:16:17.744223092Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:17.745592 containerd[1483]: time="2026-01-17T00:16:17.745552150Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 547.066173ms" Jan 17 00:16:17.745876 containerd[1483]: time="2026-01-17T00:16:17.745765577Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jan 17 00:16:17.746741 containerd[1483]: time="2026-01-17T00:16:17.746706210Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Jan 17 00:16:18.333249 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2750741941.mount: Deactivated successfully. Jan 17 00:16:20.375826 containerd[1483]: time="2026-01-17T00:16:20.374259799Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:20.375826 containerd[1483]: time="2026-01-17T00:16:20.375348652Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Jan 17 00:16:20.375826 containerd[1483]: time="2026-01-17T00:16:20.375766318Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:20.379607 containerd[1483]: time="2026-01-17T00:16:20.379563180Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:20.381517 containerd[1483]: time="2026-01-17T00:16:20.381465487Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.634718642s" Jan 17 00:16:20.381624 containerd[1483]: time="2026-01-17T00:16:20.381521011Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Jan 17 00:16:20.798551 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 17 00:16:20.806620 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 00:16:21.042430 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:16:21.052923 (kubelet)[2003]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 00:16:21.128936 kubelet[2003]: E0117 00:16:21.128840 2003 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 00:16:21.132811 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 00:16:21.133006 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 00:16:23.059811 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:16:23.067547 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 00:16:23.126996 systemd[1]: Reloading requested from client PID 2018 ('systemctl') (unit session-5.scope)... Jan 17 00:16:23.127022 systemd[1]: Reloading... Jan 17 00:16:23.278715 zram_generator::config[2057]: No configuration found. Jan 17 00:16:23.465864 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 00:16:23.587686 systemd[1]: Reloading finished in 459 ms. Jan 17 00:16:23.643483 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 17 00:16:23.643606 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 17 00:16:23.643952 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:16:23.659718 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 00:16:23.844437 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:16:23.844803 (kubelet)[2110]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 17 00:16:23.905403 kubelet[2110]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 00:16:23.905896 kubelet[2110]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 17 00:16:23.905959 kubelet[2110]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 00:16:23.906171 kubelet[2110]: I0117 00:16:23.906101 2110 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 17 00:16:24.592198 kubelet[2110]: I0117 00:16:24.591377 2110 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Jan 17 00:16:24.592198 kubelet[2110]: I0117 00:16:24.591421 2110 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 17 00:16:24.592198 kubelet[2110]: I0117 00:16:24.591709 2110 server.go:954] "Client rotation is on, will bootstrap in background" Jan 17 00:16:24.626792 kubelet[2110]: I0117 00:16:24.626264 2110 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 17 00:16:24.627847 kubelet[2110]: E0117 00:16:24.627099 2110 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://143.198.57.57:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:24.640174 kubelet[2110]: E0117 00:16:24.640011 2110 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 17 00:16:24.640174 kubelet[2110]: I0117 00:16:24.640079 2110 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 17 00:16:24.644985 kubelet[2110]: I0117 00:16:24.644940 2110 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 17 00:16:24.647865 kubelet[2110]: I0117 00:16:24.647759 2110 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 17 00:16:24.648113 kubelet[2110]: I0117 00:16:24.647863 2110 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.6-n-09ad808b32","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 17 00:16:24.648286 kubelet[2110]: I0117 00:16:24.648124 2110 topology_manager.go:138] "Creating topology manager with none policy" Jan 17 00:16:24.648286 kubelet[2110]: I0117 00:16:24.648160 2110 container_manager_linux.go:304] "Creating device plugin manager" Jan 17 00:16:24.649746 kubelet[2110]: I0117 00:16:24.649668 2110 state_mem.go:36] "Initialized new in-memory state store" Jan 17 00:16:24.654529 kubelet[2110]: I0117 00:16:24.654468 2110 kubelet.go:446] "Attempting to sync node with API server" Jan 17 00:16:24.654529 kubelet[2110]: I0117 00:16:24.654532 2110 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 17 00:16:24.655986 kubelet[2110]: I0117 00:16:24.654566 2110 kubelet.go:352] "Adding apiserver pod source" Jan 17 00:16:24.655986 kubelet[2110]: I0117 00:16:24.654582 2110 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 17 00:16:24.666263 kubelet[2110]: I0117 00:16:24.666219 2110 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 17 00:16:24.672895 kubelet[2110]: W0117 00:16:24.669462 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://143.198.57.57:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-09ad808b32&limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:24.673349 kubelet[2110]: E0117 00:16:24.673308 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://143.198.57.57:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-09ad808b32&limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:24.673711 kubelet[2110]: W0117 00:16:24.673659 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://143.198.57.57:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:24.673838 kubelet[2110]: E0117 00:16:24.673818 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://143.198.57.57:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:24.674681 kubelet[2110]: I0117 00:16:24.674620 2110 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 17 00:16:24.674775 kubelet[2110]: W0117 00:16:24.674733 2110 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 17 00:16:24.676176 kubelet[2110]: I0117 00:16:24.675630 2110 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 17 00:16:24.676176 kubelet[2110]: I0117 00:16:24.675679 2110 server.go:1287] "Started kubelet" Jan 17 00:16:24.680008 kubelet[2110]: I0117 00:16:24.679735 2110 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Jan 17 00:16:24.681419 kubelet[2110]: I0117 00:16:24.681380 2110 server.go:479] "Adding debug handlers to kubelet server" Jan 17 00:16:24.683157 kubelet[2110]: I0117 00:16:24.682893 2110 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 17 00:16:24.683538 kubelet[2110]: I0117 00:16:24.683516 2110 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 17 00:16:24.687261 kubelet[2110]: I0117 00:16:24.685642 2110 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 17 00:16:24.695232 kubelet[2110]: E0117 00:16:24.693478 2110 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://143.198.57.57:6443/api/v1/namespaces/default/events\": dial tcp 143.198.57.57:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.6-n-09ad808b32.188b5c8046280772 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.6-n-09ad808b32,UID:ci-4081.3.6-n-09ad808b32,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.6-n-09ad808b32,},FirstTimestamp:2026-01-17 00:16:24.675649394 +0000 UTC m=+0.823202519,LastTimestamp:2026-01-17 00:16:24.675649394 +0000 UTC m=+0.823202519,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.6-n-09ad808b32,}" Jan 17 00:16:24.695794 kubelet[2110]: I0117 00:16:24.695758 2110 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 17 00:16:24.702613 kubelet[2110]: I0117 00:16:24.702578 2110 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 17 00:16:24.702997 kubelet[2110]: E0117 00:16:24.702974 2110 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.6-n-09ad808b32\" not found" Jan 17 00:16:24.703201 kubelet[2110]: I0117 00:16:24.703185 2110 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 17 00:16:24.703351 kubelet[2110]: I0117 00:16:24.703339 2110 reconciler.go:26] "Reconciler: start to sync state" Jan 17 00:16:24.706987 kubelet[2110]: I0117 00:16:24.706941 2110 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 17 00:16:24.707641 kubelet[2110]: E0117 00:16:24.707596 2110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.57.57:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-09ad808b32?timeout=10s\": dial tcp 143.198.57.57:6443: connect: connection refused" interval="200ms" Jan 17 00:16:24.708337 kubelet[2110]: W0117 00:16:24.708083 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://143.198.57.57:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:24.708337 kubelet[2110]: E0117 00:16:24.708177 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://143.198.57.57:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:24.711060 kubelet[2110]: I0117 00:16:24.711026 2110 factory.go:221] Registration of the containerd container factory successfully Jan 17 00:16:24.711847 kubelet[2110]: I0117 00:16:24.711291 2110 factory.go:221] Registration of the systemd container factory successfully Jan 17 00:16:24.719320 kubelet[2110]: I0117 00:16:24.719251 2110 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 17 00:16:24.721960 kubelet[2110]: I0117 00:16:24.720928 2110 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 17 00:16:24.723801 kubelet[2110]: I0117 00:16:24.723354 2110 status_manager.go:227] "Starting to sync pod status with apiserver" Jan 17 00:16:24.723801 kubelet[2110]: I0117 00:16:24.723429 2110 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 17 00:16:24.723801 kubelet[2110]: I0117 00:16:24.723441 2110 kubelet.go:2382] "Starting kubelet main sync loop" Jan 17 00:16:24.723801 kubelet[2110]: E0117 00:16:24.723513 2110 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 17 00:16:24.726405 kubelet[2110]: E0117 00:16:24.726363 2110 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 17 00:16:24.742229 kubelet[2110]: W0117 00:16:24.742107 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://143.198.57.57:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:24.743091 kubelet[2110]: E0117 00:16:24.742480 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://143.198.57.57:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:24.753655 kubelet[2110]: I0117 00:16:24.753622 2110 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 17 00:16:24.754227 kubelet[2110]: I0117 00:16:24.753899 2110 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 17 00:16:24.754227 kubelet[2110]: I0117 00:16:24.753935 2110 state_mem.go:36] "Initialized new in-memory state store" Jan 17 00:16:24.756474 kubelet[2110]: I0117 00:16:24.756013 2110 policy_none.go:49] "None policy: Start" Jan 17 00:16:24.756474 kubelet[2110]: I0117 00:16:24.756059 2110 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 17 00:16:24.756474 kubelet[2110]: I0117 00:16:24.756078 2110 state_mem.go:35] "Initializing new in-memory state store" Jan 17 00:16:24.766098 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 17 00:16:24.777940 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 17 00:16:24.782493 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 17 00:16:24.794463 kubelet[2110]: I0117 00:16:24.793650 2110 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 17 00:16:24.794463 kubelet[2110]: I0117 00:16:24.793887 2110 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 17 00:16:24.794463 kubelet[2110]: I0117 00:16:24.794431 2110 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 17 00:16:24.794872 kubelet[2110]: I0117 00:16:24.794738 2110 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 17 00:16:24.797965 kubelet[2110]: E0117 00:16:24.797877 2110 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 17 00:16:24.797965 kubelet[2110]: E0117 00:16:24.797939 2110 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.6-n-09ad808b32\" not found" Jan 17 00:16:24.836684 systemd[1]: Created slice kubepods-burstable-pod05899a951d27afed2430cb3966b87e48.slice - libcontainer container kubepods-burstable-pod05899a951d27afed2430cb3966b87e48.slice. Jan 17 00:16:24.861172 kubelet[2110]: E0117 00:16:24.861014 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.866593 systemd[1]: Created slice kubepods-burstable-pode2620ab1e85978569b630bfbb9aa0ed5.slice - libcontainer container kubepods-burstable-pode2620ab1e85978569b630bfbb9aa0ed5.slice. Jan 17 00:16:24.869687 kubelet[2110]: E0117 00:16:24.869647 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.872414 systemd[1]: Created slice kubepods-burstable-pod953358aaeeadfb7bfa465aefc2a50423.slice - libcontainer container kubepods-burstable-pod953358aaeeadfb7bfa465aefc2a50423.slice. Jan 17 00:16:24.874941 kubelet[2110]: E0117 00:16:24.874369 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.898090 kubelet[2110]: I0117 00:16:24.897534 2110 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.898090 kubelet[2110]: E0117 00:16:24.898037 2110 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://143.198.57.57:6443/api/v1/nodes\": dial tcp 143.198.57.57:6443: connect: connection refused" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.904558 kubelet[2110]: I0117 00:16:24.904490 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-ca-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.904558 kubelet[2110]: I0117 00:16:24.904554 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.904922 kubelet[2110]: I0117 00:16:24.904590 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.904922 kubelet[2110]: I0117 00:16:24.904623 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/05899a951d27afed2430cb3966b87e48-ca-certs\") pod \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" (UID: \"05899a951d27afed2430cb3966b87e48\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.904922 kubelet[2110]: I0117 00:16:24.904649 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/05899a951d27afed2430cb3966b87e48-k8s-certs\") pod \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" (UID: \"05899a951d27afed2430cb3966b87e48\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.904922 kubelet[2110]: I0117 00:16:24.904718 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.904922 kubelet[2110]: I0117 00:16:24.904753 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.905212 kubelet[2110]: I0117 00:16:24.904778 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e2620ab1e85978569b630bfbb9aa0ed5-kubeconfig\") pod \"kube-scheduler-ci-4081.3.6-n-09ad808b32\" (UID: \"e2620ab1e85978569b630bfbb9aa0ed5\") " pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.905212 kubelet[2110]: I0117 00:16:24.904798 2110 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/05899a951d27afed2430cb3966b87e48-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" (UID: \"05899a951d27afed2430cb3966b87e48\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:24.909524 kubelet[2110]: E0117 00:16:24.909449 2110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.57.57:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-09ad808b32?timeout=10s\": dial tcp 143.198.57.57:6443: connect: connection refused" interval="400ms" Jan 17 00:16:25.099650 kubelet[2110]: I0117 00:16:25.099207 2110 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:25.099650 kubelet[2110]: E0117 00:16:25.099601 2110 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://143.198.57.57:6443/api/v1/nodes\": dial tcp 143.198.57.57:6443: connect: connection refused" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:25.162428 kubelet[2110]: E0117 00:16:25.162282 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:25.166332 containerd[1483]: time="2026-01-17T00:16:25.166260964Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.6-n-09ad808b32,Uid:05899a951d27afed2430cb3966b87e48,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:25.170856 kubelet[2110]: E0117 00:16:25.170794 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:25.171450 containerd[1483]: time="2026-01-17T00:16:25.171395175Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.6-n-09ad808b32,Uid:e2620ab1e85978569b630bfbb9aa0ed5,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:25.175395 kubelet[2110]: E0117 00:16:25.175359 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:25.175921 containerd[1483]: time="2026-01-17T00:16:25.175887663Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.6-n-09ad808b32,Uid:953358aaeeadfb7bfa465aefc2a50423,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:25.310353 kubelet[2110]: E0117 00:16:25.310267 2110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.57.57:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-09ad808b32?timeout=10s\": dial tcp 143.198.57.57:6443: connect: connection refused" interval="800ms" Jan 17 00:16:25.501735 kubelet[2110]: I0117 00:16:25.501578 2110 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:25.502620 kubelet[2110]: E0117 00:16:25.502576 2110 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://143.198.57.57:6443/api/v1/nodes\": dial tcp 143.198.57.57:6443: connect: connection refused" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:25.859630 kubelet[2110]: W0117 00:16:25.859432 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://143.198.57.57:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:25.859630 kubelet[2110]: E0117 00:16:25.859532 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://143.198.57.57:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:25.925203 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1858757625.mount: Deactivated successfully. Jan 17 00:16:25.931402 containerd[1483]: time="2026-01-17T00:16:25.931318894Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 00:16:25.932826 containerd[1483]: time="2026-01-17T00:16:25.932755648Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 00:16:25.934281 containerd[1483]: time="2026-01-17T00:16:25.934209480Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 17 00:16:25.934575 containerd[1483]: time="2026-01-17T00:16:25.934510333Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Jan 17 00:16:25.934745 containerd[1483]: time="2026-01-17T00:16:25.934709716Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 00:16:25.935588 containerd[1483]: time="2026-01-17T00:16:25.935526035Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 17 00:16:25.937180 containerd[1483]: time="2026-01-17T00:16:25.935895741Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 00:16:25.939812 containerd[1483]: time="2026-01-17T00:16:25.939712404Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 00:16:25.940965 containerd[1483]: time="2026-01-17T00:16:25.940627167Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 769.130029ms" Jan 17 00:16:25.944175 containerd[1483]: time="2026-01-17T00:16:25.943937103Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 767.975587ms" Jan 17 00:16:25.948078 containerd[1483]: time="2026-01-17T00:16:25.947809610Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 781.448872ms" Jan 17 00:16:25.949225 kubelet[2110]: W0117 00:16:25.948784 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://143.198.57.57:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:25.949225 kubelet[2110]: E0117 00:16:25.948959 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://143.198.57.57:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:26.077542 kubelet[2110]: W0117 00:16:26.066377 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://143.198.57.57:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-09ad808b32&limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:26.077542 kubelet[2110]: E0117 00:16:26.066464 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://143.198.57.57:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.6-n-09ad808b32&limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:26.111713 kubelet[2110]: E0117 00:16:26.111564 2110 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.57.57:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.6-n-09ad808b32?timeout=10s\": dial tcp 143.198.57.57:6443: connect: connection refused" interval="1.6s" Jan 17 00:16:26.139116 containerd[1483]: time="2026-01-17T00:16:26.138418241Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 00:16:26.139346 containerd[1483]: time="2026-01-17T00:16:26.139098676Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 00:16:26.139346 containerd[1483]: time="2026-01-17T00:16:26.139147902Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:26.139346 containerd[1483]: time="2026-01-17T00:16:26.139302613Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:26.156751 containerd[1483]: time="2026-01-17T00:16:26.156533145Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 00:16:26.158621 containerd[1483]: time="2026-01-17T00:16:26.158397228Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 00:16:26.158621 containerd[1483]: time="2026-01-17T00:16:26.158431772Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:26.159751 containerd[1483]: time="2026-01-17T00:16:26.158933292Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:26.163394 containerd[1483]: time="2026-01-17T00:16:26.162984007Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 00:16:26.165182 containerd[1483]: time="2026-01-17T00:16:26.164537163Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 00:16:26.165182 containerd[1483]: time="2026-01-17T00:16:26.164946183Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:26.165182 containerd[1483]: time="2026-01-17T00:16:26.165063588Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:26.178463 systemd[1]: Started cri-containerd-f4ac5e03aada736f6822b75bfde9b385038a922e932ba91bc39477bb76b1ec8e.scope - libcontainer container f4ac5e03aada736f6822b75bfde9b385038a922e932ba91bc39477bb76b1ec8e. Jan 17 00:16:26.213558 systemd[1]: Started cri-containerd-b67f26ea1212a1ad207bc95ff7c50d6b841ece6a92a02f429fcc61583755a0de.scope - libcontainer container b67f26ea1212a1ad207bc95ff7c50d6b841ece6a92a02f429fcc61583755a0de. Jan 17 00:16:26.220272 systemd[1]: Started cri-containerd-0b457cc2a4c3fdb13654c839c5f96171c83a8a5d255577124c186036d89dce2c.scope - libcontainer container 0b457cc2a4c3fdb13654c839c5f96171c83a8a5d255577124c186036d89dce2c. Jan 17 00:16:26.274102 containerd[1483]: time="2026-01-17T00:16:26.274017934Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.6-n-09ad808b32,Uid:e2620ab1e85978569b630bfbb9aa0ed5,Namespace:kube-system,Attempt:0,} returns sandbox id \"f4ac5e03aada736f6822b75bfde9b385038a922e932ba91bc39477bb76b1ec8e\"" Jan 17 00:16:26.279050 kubelet[2110]: E0117 00:16:26.278633 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:26.283049 containerd[1483]: time="2026-01-17T00:16:26.282915264Z" level=info msg="CreateContainer within sandbox \"f4ac5e03aada736f6822b75bfde9b385038a922e932ba91bc39477bb76b1ec8e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 17 00:16:26.298509 kubelet[2110]: W0117 00:16:26.298463 2110 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://143.198.57.57:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.57.57:6443: connect: connection refused Jan 17 00:16:26.299004 kubelet[2110]: E0117 00:16:26.298922 2110 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://143.198.57.57:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:26.301371 containerd[1483]: time="2026-01-17T00:16:26.301313619Z" level=info msg="CreateContainer within sandbox \"f4ac5e03aada736f6822b75bfde9b385038a922e932ba91bc39477bb76b1ec8e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c9dd87c28317f1a7a911b028793f655d1f6d7666b17e0c3209aba93b278bbef9\"" Jan 17 00:16:26.302588 containerd[1483]: time="2026-01-17T00:16:26.302280746Z" level=info msg="StartContainer for \"c9dd87c28317f1a7a911b028793f655d1f6d7666b17e0c3209aba93b278bbef9\"" Jan 17 00:16:26.304812 kubelet[2110]: I0117 00:16:26.304789 2110 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:26.306360 kubelet[2110]: E0117 00:16:26.305548 2110 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://143.198.57.57:6443/api/v1/nodes\": dial tcp 143.198.57.57:6443: connect: connection refused" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:26.324860 containerd[1483]: time="2026-01-17T00:16:26.324818462Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.6-n-09ad808b32,Uid:953358aaeeadfb7bfa465aefc2a50423,Namespace:kube-system,Attempt:0,} returns sandbox id \"0b457cc2a4c3fdb13654c839c5f96171c83a8a5d255577124c186036d89dce2c\"" Jan 17 00:16:26.326117 kubelet[2110]: E0117 00:16:26.326048 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:26.328906 containerd[1483]: time="2026-01-17T00:16:26.328854901Z" level=info msg="CreateContainer within sandbox \"0b457cc2a4c3fdb13654c839c5f96171c83a8a5d255577124c186036d89dce2c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 17 00:16:26.343839 containerd[1483]: time="2026-01-17T00:16:26.343781177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.6-n-09ad808b32,Uid:05899a951d27afed2430cb3966b87e48,Namespace:kube-system,Attempt:0,} returns sandbox id \"b67f26ea1212a1ad207bc95ff7c50d6b841ece6a92a02f429fcc61583755a0de\"" Jan 17 00:16:26.346029 kubelet[2110]: E0117 00:16:26.345996 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:26.348019 containerd[1483]: time="2026-01-17T00:16:26.347976382Z" level=info msg="CreateContainer within sandbox \"b67f26ea1212a1ad207bc95ff7c50d6b841ece6a92a02f429fcc61583755a0de\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 17 00:16:26.351780 containerd[1483]: time="2026-01-17T00:16:26.351744519Z" level=info msg="CreateContainer within sandbox \"0b457cc2a4c3fdb13654c839c5f96171c83a8a5d255577124c186036d89dce2c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"a8d6068ff03827abac76144c74fdbe365bdf1af24cc94e9369729315a9af79a6\"" Jan 17 00:16:26.353486 containerd[1483]: time="2026-01-17T00:16:26.353350109Z" level=info msg="StartContainer for \"a8d6068ff03827abac76144c74fdbe365bdf1af24cc94e9369729315a9af79a6\"" Jan 17 00:16:26.360538 systemd[1]: Started cri-containerd-c9dd87c28317f1a7a911b028793f655d1f6d7666b17e0c3209aba93b278bbef9.scope - libcontainer container c9dd87c28317f1a7a911b028793f655d1f6d7666b17e0c3209aba93b278bbef9. Jan 17 00:16:26.369156 containerd[1483]: time="2026-01-17T00:16:26.367681322Z" level=info msg="CreateContainer within sandbox \"b67f26ea1212a1ad207bc95ff7c50d6b841ece6a92a02f429fcc61583755a0de\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"899e62cec0a5dc57b344b47137d9fd3a81c9a6dcbe88c6e005f885e80beeb628\"" Jan 17 00:16:26.371024 containerd[1483]: time="2026-01-17T00:16:26.370965422Z" level=info msg="StartContainer for \"899e62cec0a5dc57b344b47137d9fd3a81c9a6dcbe88c6e005f885e80beeb628\"" Jan 17 00:16:26.424413 systemd[1]: Started cri-containerd-899e62cec0a5dc57b344b47137d9fd3a81c9a6dcbe88c6e005f885e80beeb628.scope - libcontainer container 899e62cec0a5dc57b344b47137d9fd3a81c9a6dcbe88c6e005f885e80beeb628. Jan 17 00:16:26.432617 systemd[1]: Started cri-containerd-a8d6068ff03827abac76144c74fdbe365bdf1af24cc94e9369729315a9af79a6.scope - libcontainer container a8d6068ff03827abac76144c74fdbe365bdf1af24cc94e9369729315a9af79a6. Jan 17 00:16:26.471057 containerd[1483]: time="2026-01-17T00:16:26.470137872Z" level=info msg="StartContainer for \"c9dd87c28317f1a7a911b028793f655d1f6d7666b17e0c3209aba93b278bbef9\" returns successfully" Jan 17 00:16:26.512792 containerd[1483]: time="2026-01-17T00:16:26.512560119Z" level=info msg="StartContainer for \"a8d6068ff03827abac76144c74fdbe365bdf1af24cc94e9369729315a9af79a6\" returns successfully" Jan 17 00:16:26.547221 containerd[1483]: time="2026-01-17T00:16:26.546506904Z" level=info msg="StartContainer for \"899e62cec0a5dc57b344b47137d9fd3a81c9a6dcbe88c6e005f885e80beeb628\" returns successfully" Jan 17 00:16:26.755915 kubelet[2110]: E0117 00:16:26.754890 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:26.755915 kubelet[2110]: E0117 00:16:26.755082 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:26.760914 kubelet[2110]: E0117 00:16:26.760240 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:26.760914 kubelet[2110]: E0117 00:16:26.760466 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:26.765502 kubelet[2110]: E0117 00:16:26.765185 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:26.765502 kubelet[2110]: E0117 00:16:26.765397 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:26.815188 kubelet[2110]: E0117 00:16:26.814004 2110 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://143.198.57.57:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 143.198.57.57:6443: connect: connection refused" logger="UnhandledError" Jan 17 00:16:27.767696 kubelet[2110]: E0117 00:16:27.767454 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:27.767696 kubelet[2110]: E0117 00:16:27.767512 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:27.767696 kubelet[2110]: E0117 00:16:27.767596 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:27.768489 kubelet[2110]: E0117 00:16:27.768411 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:27.908267 kubelet[2110]: I0117 00:16:27.907856 2110 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:28.771684 kubelet[2110]: E0117 00:16:28.770768 2110 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:28.772715 kubelet[2110]: E0117 00:16:28.772508 2110 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:29.028023 kubelet[2110]: E0117 00:16:29.027839 2110 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.6-n-09ad808b32\" not found" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.106199 kubelet[2110]: I0117 00:16:29.106160 2110 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.202882 kubelet[2110]: I0117 00:16:29.202503 2110 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.214679 kubelet[2110]: E0117 00:16:29.214633 2110 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.215007 kubelet[2110]: I0117 00:16:29.214868 2110 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.221044 kubelet[2110]: E0117 00:16:29.220928 2110 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.221044 kubelet[2110]: I0117 00:16:29.220984 2110 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.232173 kubelet[2110]: E0117 00:16:29.230272 2110 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-09ad808b32\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:29.663956 kubelet[2110]: I0117 00:16:29.663592 2110 apiserver.go:52] "Watching apiserver" Jan 17 00:16:29.704219 kubelet[2110]: I0117 00:16:29.704106 2110 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 17 00:16:31.313330 systemd[1]: Reloading requested from client PID 2380 ('systemctl') (unit session-5.scope)... Jan 17 00:16:31.313836 systemd[1]: Reloading... Jan 17 00:16:31.431179 zram_generator::config[2422]: No configuration found. Jan 17 00:16:31.612326 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 00:16:31.735768 systemd[1]: Reloading finished in 421 ms. Jan 17 00:16:31.794662 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 00:16:31.795678 kubelet[2110]: I0117 00:16:31.795048 2110 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 17 00:16:31.814251 systemd[1]: kubelet.service: Deactivated successfully. Jan 17 00:16:31.814742 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:16:31.814811 systemd[1]: kubelet.service: Consumed 1.363s CPU time, 129.3M memory peak, 0B memory swap peak. Jan 17 00:16:31.821610 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 00:16:32.015459 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 00:16:32.016654 (kubelet)[2470]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 17 00:16:32.112989 kubelet[2470]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 00:16:32.112989 kubelet[2470]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 17 00:16:32.112989 kubelet[2470]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 00:16:32.112989 kubelet[2470]: I0117 00:16:32.110751 2470 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 17 00:16:32.127425 kubelet[2470]: I0117 00:16:32.127340 2470 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Jan 17 00:16:32.127425 kubelet[2470]: I0117 00:16:32.127412 2470 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 17 00:16:32.127932 kubelet[2470]: I0117 00:16:32.127905 2470 server.go:954] "Client rotation is on, will bootstrap in background" Jan 17 00:16:32.132116 kubelet[2470]: I0117 00:16:32.132077 2470 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 17 00:16:32.139708 kubelet[2470]: I0117 00:16:32.139654 2470 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 17 00:16:32.151182 kubelet[2470]: E0117 00:16:32.150729 2470 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 17 00:16:32.151182 kubelet[2470]: I0117 00:16:32.150767 2470 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 17 00:16:32.154298 kubelet[2470]: I0117 00:16:32.154249 2470 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 17 00:16:32.154559 kubelet[2470]: I0117 00:16:32.154494 2470 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 17 00:16:32.154726 kubelet[2470]: I0117 00:16:32.154532 2470 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.6-n-09ad808b32","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 17 00:16:32.154830 kubelet[2470]: I0117 00:16:32.154729 2470 topology_manager.go:138] "Creating topology manager with none policy" Jan 17 00:16:32.154830 kubelet[2470]: I0117 00:16:32.154741 2470 container_manager_linux.go:304] "Creating device plugin manager" Jan 17 00:16:32.154830 kubelet[2470]: I0117 00:16:32.154819 2470 state_mem.go:36] "Initialized new in-memory state store" Jan 17 00:16:32.155225 kubelet[2470]: I0117 00:16:32.155198 2470 kubelet.go:446] "Attempting to sync node with API server" Jan 17 00:16:32.155225 kubelet[2470]: I0117 00:16:32.155225 2470 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 17 00:16:32.156390 kubelet[2470]: I0117 00:16:32.156358 2470 kubelet.go:352] "Adding apiserver pod source" Jan 17 00:16:32.156390 kubelet[2470]: I0117 00:16:32.156380 2470 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 17 00:16:32.171164 kubelet[2470]: I0117 00:16:32.170326 2470 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 17 00:16:32.171164 kubelet[2470]: I0117 00:16:32.170812 2470 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 17 00:16:32.171758 kubelet[2470]: I0117 00:16:32.171735 2470 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 17 00:16:32.171907 kubelet[2470]: I0117 00:16:32.171893 2470 server.go:1287] "Started kubelet" Jan 17 00:16:32.174500 kubelet[2470]: I0117 00:16:32.174469 2470 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 17 00:16:32.185075 kubelet[2470]: I0117 00:16:32.185026 2470 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Jan 17 00:16:32.185975 kubelet[2470]: I0117 00:16:32.185944 2470 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 17 00:16:32.187273 kubelet[2470]: I0117 00:16:32.187239 2470 server.go:479] "Adding debug handlers to kubelet server" Jan 17 00:16:32.191621 kubelet[2470]: I0117 00:16:32.190909 2470 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 17 00:16:32.191621 kubelet[2470]: I0117 00:16:32.191435 2470 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 17 00:16:32.192364 kubelet[2470]: I0117 00:16:32.192015 2470 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 17 00:16:32.193333 kubelet[2470]: I0117 00:16:32.193308 2470 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jan 17 00:16:32.193605 kubelet[2470]: I0117 00:16:32.193588 2470 reconciler.go:26] "Reconciler: start to sync state" Jan 17 00:16:32.196949 kubelet[2470]: I0117 00:16:32.196874 2470 factory.go:221] Registration of the systemd container factory successfully Jan 17 00:16:32.199058 kubelet[2470]: I0117 00:16:32.196995 2470 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 17 00:16:32.199058 kubelet[2470]: E0117 00:16:32.198198 2470 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 17 00:16:32.201416 kubelet[2470]: I0117 00:16:32.200624 2470 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 17 00:16:32.203010 kubelet[2470]: I0117 00:16:32.202092 2470 factory.go:221] Registration of the containerd container factory successfully Jan 17 00:16:32.205211 kubelet[2470]: I0117 00:16:32.203947 2470 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 17 00:16:32.205385 kubelet[2470]: I0117 00:16:32.205369 2470 status_manager.go:227] "Starting to sync pod status with apiserver" Jan 17 00:16:32.205462 kubelet[2470]: I0117 00:16:32.205452 2470 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 17 00:16:32.205516 kubelet[2470]: I0117 00:16:32.205509 2470 kubelet.go:2382] "Starting kubelet main sync loop" Jan 17 00:16:32.205663 kubelet[2470]: E0117 00:16:32.205633 2470 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 17 00:16:32.263691 kubelet[2470]: I0117 00:16:32.263652 2470 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 17 00:16:32.263691 kubelet[2470]: I0117 00:16:32.263680 2470 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 17 00:16:32.263891 kubelet[2470]: I0117 00:16:32.263707 2470 state_mem.go:36] "Initialized new in-memory state store" Jan 17 00:16:32.263972 kubelet[2470]: I0117 00:16:32.263940 2470 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 17 00:16:32.263972 kubelet[2470]: I0117 00:16:32.263954 2470 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 17 00:16:32.264065 kubelet[2470]: I0117 00:16:32.263979 2470 policy_none.go:49] "None policy: Start" Jan 17 00:16:32.264065 kubelet[2470]: I0117 00:16:32.263992 2470 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 17 00:16:32.264065 kubelet[2470]: I0117 00:16:32.264005 2470 state_mem.go:35] "Initializing new in-memory state store" Jan 17 00:16:32.264203 kubelet[2470]: I0117 00:16:32.264176 2470 state_mem.go:75] "Updated machine memory state" Jan 17 00:16:32.270804 kubelet[2470]: I0117 00:16:32.269281 2470 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 17 00:16:32.271225 kubelet[2470]: I0117 00:16:32.271122 2470 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 17 00:16:32.271841 kubelet[2470]: I0117 00:16:32.271535 2470 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 17 00:16:32.273474 kubelet[2470]: I0117 00:16:32.273010 2470 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 17 00:16:32.274843 kubelet[2470]: E0117 00:16:32.274822 2470 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 17 00:16:32.307383 kubelet[2470]: I0117 00:16:32.307345 2470 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.308805 kubelet[2470]: I0117 00:16:32.308691 2470 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.309122 kubelet[2470]: I0117 00:16:32.308780 2470 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.322814 kubelet[2470]: W0117 00:16:32.322720 2470 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 17 00:16:32.323042 kubelet[2470]: W0117 00:16:32.322896 2470 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 17 00:16:32.323637 kubelet[2470]: W0117 00:16:32.323421 2470 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 17 00:16:32.377107 kubelet[2470]: I0117 00:16:32.377051 2470 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.392198 kubelet[2470]: I0117 00:16:32.392161 2470 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.392386 kubelet[2470]: I0117 00:16:32.392274 2470 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395180 kubelet[2470]: I0117 00:16:32.394873 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-ca-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395180 kubelet[2470]: I0117 00:16:32.394920 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395180 kubelet[2470]: I0117 00:16:32.394952 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395180 kubelet[2470]: I0117 00:16:32.394972 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395180 kubelet[2470]: I0117 00:16:32.394995 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e2620ab1e85978569b630bfbb9aa0ed5-kubeconfig\") pod \"kube-scheduler-ci-4081.3.6-n-09ad808b32\" (UID: \"e2620ab1e85978569b630bfbb9aa0ed5\") " pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395514 kubelet[2470]: I0117 00:16:32.395014 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/05899a951d27afed2430cb3966b87e48-ca-certs\") pod \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" (UID: \"05899a951d27afed2430cb3966b87e48\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395514 kubelet[2470]: I0117 00:16:32.395029 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/05899a951d27afed2430cb3966b87e48-k8s-certs\") pod \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" (UID: \"05899a951d27afed2430cb3966b87e48\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395514 kubelet[2470]: I0117 00:16:32.395045 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/05899a951d27afed2430cb3966b87e48-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" (UID: \"05899a951d27afed2430cb3966b87e48\") " pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.395514 kubelet[2470]: I0117 00:16:32.395063 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/953358aaeeadfb7bfa465aefc2a50423-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.6-n-09ad808b32\" (UID: \"953358aaeeadfb7bfa465aefc2a50423\") " pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:32.627099 kubelet[2470]: E0117 00:16:32.624846 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:32.627099 kubelet[2470]: E0117 00:16:32.625473 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:32.627099 kubelet[2470]: E0117 00:16:32.625472 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:33.161032 kubelet[2470]: I0117 00:16:33.160967 2470 apiserver.go:52] "Watching apiserver" Jan 17 00:16:33.194120 kubelet[2470]: I0117 00:16:33.194060 2470 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jan 17 00:16:33.247997 kubelet[2470]: E0117 00:16:33.247683 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:33.251274 kubelet[2470]: I0117 00:16:33.249074 2470 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:33.251274 kubelet[2470]: I0117 00:16:33.248438 2470 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:33.281585 kubelet[2470]: W0117 00:16:33.281535 2470 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 17 00:16:33.281585 kubelet[2470]: W0117 00:16:33.281596 2470 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 17 00:16:33.281802 kubelet[2470]: E0117 00:16:33.281646 2470 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.6-n-09ad808b32\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:33.281857 kubelet[2470]: E0117 00:16:33.281827 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:33.284580 kubelet[2470]: E0117 00:16:33.283161 2470 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.6-n-09ad808b32\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" Jan 17 00:16:33.284580 kubelet[2470]: E0117 00:16:33.283345 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:33.348552 kubelet[2470]: I0117 00:16:33.348485 2470 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.6-n-09ad808b32" podStartSLOduration=1.348461795 podStartE2EDuration="1.348461795s" podCreationTimestamp="2026-01-17 00:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-17 00:16:33.326931926 +0000 UTC m=+1.299917343" watchObservedRunningTime="2026-01-17 00:16:33.348461795 +0000 UTC m=+1.321447194" Jan 17 00:16:33.368267 kubelet[2470]: I0117 00:16:33.368204 2470 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.6-n-09ad808b32" podStartSLOduration=1.368184095 podStartE2EDuration="1.368184095s" podCreationTimestamp="2026-01-17 00:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-17 00:16:33.350082099 +0000 UTC m=+1.323067505" watchObservedRunningTime="2026-01-17 00:16:33.368184095 +0000 UTC m=+1.341169503" Jan 17 00:16:33.387332 kubelet[2470]: I0117 00:16:33.387265 2470 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.6-n-09ad808b32" podStartSLOduration=1.387243365 podStartE2EDuration="1.387243365s" podCreationTimestamp="2026-01-17 00:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-17 00:16:33.368789353 +0000 UTC m=+1.341774761" watchObservedRunningTime="2026-01-17 00:16:33.387243365 +0000 UTC m=+1.360228764" Jan 17 00:16:33.575326 sudo[1613]: pam_unix(sudo:session): session closed for user root Jan 17 00:16:33.650956 sshd[1610]: pam_unix(sshd:session): session closed for user core Jan 17 00:16:33.654606 systemd[1]: sshd@4-143.198.57.57:22-4.153.228.146:51854.service: Deactivated successfully. Jan 17 00:16:33.657269 systemd[1]: session-5.scope: Deactivated successfully. Jan 17 00:16:33.657484 systemd[1]: session-5.scope: Consumed 4.163s CPU time, 144.6M memory peak, 0B memory swap peak. Jan 17 00:16:33.659665 systemd-logind[1452]: Session 5 logged out. Waiting for processes to exit. Jan 17 00:16:33.660884 systemd-logind[1452]: Removed session 5. Jan 17 00:16:34.249952 kubelet[2470]: E0117 00:16:34.249851 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:34.249952 kubelet[2470]: E0117 00:16:34.249851 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:35.697411 kubelet[2470]: I0117 00:16:35.697368 2470 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 17 00:16:35.697991 containerd[1483]: time="2026-01-17T00:16:35.697928954Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 17 00:16:35.698400 kubelet[2470]: I0117 00:16:35.698281 2470 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 17 00:16:36.381206 kubelet[2470]: W0117 00:16:36.381157 2470 reflector.go:569] object-"kube-flannel"/"kube-flannel-cfg": failed to list *v1.ConfigMap: configmaps "kube-flannel-cfg" is forbidden: User "system:node:ci-4081.3.6-n-09ad808b32" cannot list resource "configmaps" in API group "" in the namespace "kube-flannel": no relationship found between node 'ci-4081.3.6-n-09ad808b32' and this object Jan 17 00:16:36.381389 kubelet[2470]: E0117 00:16:36.381203 2470 reflector.go:166] "Unhandled Error" err="object-\"kube-flannel\"/\"kube-flannel-cfg\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-flannel-cfg\" is forbidden: User \"system:node:ci-4081.3.6-n-09ad808b32\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-flannel\": no relationship found between node 'ci-4081.3.6-n-09ad808b32' and this object" logger="UnhandledError" Jan 17 00:16:36.382491 systemd[1]: Created slice kubepods-besteffort-pod9e25585a_65dd_4556_8f65_c1aa47b36f6a.slice - libcontainer container kubepods-besteffort-pod9e25585a_65dd_4556_8f65_c1aa47b36f6a.slice. Jan 17 00:16:36.402141 systemd[1]: Created slice kubepods-burstable-pod5a369a50_610b_4a62_aa4b_07336f6ae1b3.slice - libcontainer container kubepods-burstable-pod5a369a50_610b_4a62_aa4b_07336f6ae1b3.slice. Jan 17 00:16:36.425643 kubelet[2470]: I0117 00:16:36.425568 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/5a369a50-610b-4a62-aa4b-07336f6ae1b3-xtables-lock\") pod \"kube-flannel-ds-d2cbv\" (UID: \"5a369a50-610b-4a62-aa4b-07336f6ae1b3\") " pod="kube-flannel/kube-flannel-ds-d2cbv" Jan 17 00:16:36.425643 kubelet[2470]: I0117 00:16:36.425646 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bktq5\" (UniqueName: \"kubernetes.io/projected/9e25585a-65dd-4556-8f65-c1aa47b36f6a-kube-api-access-bktq5\") pod \"kube-proxy-qrp95\" (UID: \"9e25585a-65dd-4556-8f65-c1aa47b36f6a\") " pod="kube-system/kube-proxy-qrp95" Jan 17 00:16:36.425856 kubelet[2470]: I0117 00:16:36.425685 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/5a369a50-610b-4a62-aa4b-07336f6ae1b3-cni-plugin\") pod \"kube-flannel-ds-d2cbv\" (UID: \"5a369a50-610b-4a62-aa4b-07336f6ae1b3\") " pod="kube-flannel/kube-flannel-ds-d2cbv" Jan 17 00:16:36.425856 kubelet[2470]: I0117 00:16:36.425718 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/5a369a50-610b-4a62-aa4b-07336f6ae1b3-cni\") pod \"kube-flannel-ds-d2cbv\" (UID: \"5a369a50-610b-4a62-aa4b-07336f6ae1b3\") " pod="kube-flannel/kube-flannel-ds-d2cbv" Jan 17 00:16:36.425856 kubelet[2470]: I0117 00:16:36.425745 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9e25585a-65dd-4556-8f65-c1aa47b36f6a-lib-modules\") pod \"kube-proxy-qrp95\" (UID: \"9e25585a-65dd-4556-8f65-c1aa47b36f6a\") " pod="kube-system/kube-proxy-qrp95" Jan 17 00:16:36.425856 kubelet[2470]: I0117 00:16:36.425796 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/5a369a50-610b-4a62-aa4b-07336f6ae1b3-flannel-cfg\") pod \"kube-flannel-ds-d2cbv\" (UID: \"5a369a50-610b-4a62-aa4b-07336f6ae1b3\") " pod="kube-flannel/kube-flannel-ds-d2cbv" Jan 17 00:16:36.425856 kubelet[2470]: I0117 00:16:36.425823 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9e25585a-65dd-4556-8f65-c1aa47b36f6a-xtables-lock\") pod \"kube-proxy-qrp95\" (UID: \"9e25585a-65dd-4556-8f65-c1aa47b36f6a\") " pod="kube-system/kube-proxy-qrp95" Jan 17 00:16:36.426002 kubelet[2470]: I0117 00:16:36.425865 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrdvf\" (UniqueName: \"kubernetes.io/projected/5a369a50-610b-4a62-aa4b-07336f6ae1b3-kube-api-access-vrdvf\") pod \"kube-flannel-ds-d2cbv\" (UID: \"5a369a50-610b-4a62-aa4b-07336f6ae1b3\") " pod="kube-flannel/kube-flannel-ds-d2cbv" Jan 17 00:16:36.426002 kubelet[2470]: I0117 00:16:36.425900 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/9e25585a-65dd-4556-8f65-c1aa47b36f6a-kube-proxy\") pod \"kube-proxy-qrp95\" (UID: \"9e25585a-65dd-4556-8f65-c1aa47b36f6a\") " pod="kube-system/kube-proxy-qrp95" Jan 17 00:16:36.426002 kubelet[2470]: I0117 00:16:36.425926 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a369a50-610b-4a62-aa4b-07336f6ae1b3-run\") pod \"kube-flannel-ds-d2cbv\" (UID: \"5a369a50-610b-4a62-aa4b-07336f6ae1b3\") " pod="kube-flannel/kube-flannel-ds-d2cbv" Jan 17 00:16:36.698073 kubelet[2470]: E0117 00:16:36.697423 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:36.699425 containerd[1483]: time="2026-01-17T00:16:36.698230204Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-qrp95,Uid:9e25585a-65dd-4556-8f65-c1aa47b36f6a,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:36.726323 containerd[1483]: time="2026-01-17T00:16:36.726013118Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 00:16:36.726323 containerd[1483]: time="2026-01-17T00:16:36.726091411Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 00:16:36.726323 containerd[1483]: time="2026-01-17T00:16:36.726106855Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:36.732781 containerd[1483]: time="2026-01-17T00:16:36.732449436Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:36.763439 systemd[1]: Started cri-containerd-5b19c2aadb2e2d33096c74a752acb991a13e1809dac44db711f692cd46d985f1.scope - libcontainer container 5b19c2aadb2e2d33096c74a752acb991a13e1809dac44db711f692cd46d985f1. Jan 17 00:16:36.810887 containerd[1483]: time="2026-01-17T00:16:36.810542639Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-qrp95,Uid:9e25585a-65dd-4556-8f65-c1aa47b36f6a,Namespace:kube-system,Attempt:0,} returns sandbox id \"5b19c2aadb2e2d33096c74a752acb991a13e1809dac44db711f692cd46d985f1\"" Jan 17 00:16:36.812907 kubelet[2470]: E0117 00:16:36.812588 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:36.820685 containerd[1483]: time="2026-01-17T00:16:36.820587244Z" level=info msg="CreateContainer within sandbox \"5b19c2aadb2e2d33096c74a752acb991a13e1809dac44db711f692cd46d985f1\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 17 00:16:36.841915 containerd[1483]: time="2026-01-17T00:16:36.841609460Z" level=info msg="CreateContainer within sandbox \"5b19c2aadb2e2d33096c74a752acb991a13e1809dac44db711f692cd46d985f1\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"c8d2cd94962b3e048d767bd5e25d3135d042182354a60003cedfcbb69b2882a0\"" Jan 17 00:16:36.842445 containerd[1483]: time="2026-01-17T00:16:36.842395169Z" level=info msg="StartContainer for \"c8d2cd94962b3e048d767bd5e25d3135d042182354a60003cedfcbb69b2882a0\"" Jan 17 00:16:36.880438 systemd[1]: Started cri-containerd-c8d2cd94962b3e048d767bd5e25d3135d042182354a60003cedfcbb69b2882a0.scope - libcontainer container c8d2cd94962b3e048d767bd5e25d3135d042182354a60003cedfcbb69b2882a0. Jan 17 00:16:36.923229 containerd[1483]: time="2026-01-17T00:16:36.923095202Z" level=info msg="StartContainer for \"c8d2cd94962b3e048d767bd5e25d3135d042182354a60003cedfcbb69b2882a0\" returns successfully" Jan 17 00:16:37.260942 kubelet[2470]: E0117 00:16:37.260893 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:37.530476 kubelet[2470]: E0117 00:16:37.529849 2470 configmap.go:193] Couldn't get configMap kube-flannel/kube-flannel-cfg: failed to sync configmap cache: timed out waiting for the condition Jan 17 00:16:37.530476 kubelet[2470]: E0117 00:16:37.529999 2470 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a369a50-610b-4a62-aa4b-07336f6ae1b3-flannel-cfg podName:5a369a50-610b-4a62-aa4b-07336f6ae1b3 nodeName:}" failed. No retries permitted until 2026-01-17 00:16:38.029969397 +0000 UTC m=+6.002954812 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "flannel-cfg" (UniqueName: "kubernetes.io/configmap/5a369a50-610b-4a62-aa4b-07336f6ae1b3-flannel-cfg") pod "kube-flannel-ds-d2cbv" (UID: "5a369a50-610b-4a62-aa4b-07336f6ae1b3") : failed to sync configmap cache: timed out waiting for the condition Jan 17 00:16:37.547666 systemd[1]: run-containerd-runc-k8s.io-5b19c2aadb2e2d33096c74a752acb991a13e1809dac44db711f692cd46d985f1-runc.OOXGBN.mount: Deactivated successfully. Jan 17 00:16:38.208885 kubelet[2470]: E0117 00:16:38.208652 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:38.209492 containerd[1483]: time="2026-01-17T00:16:38.209445471Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-d2cbv,Uid:5a369a50-610b-4a62-aa4b-07336f6ae1b3,Namespace:kube-flannel,Attempt:0,}" Jan 17 00:16:38.248179 containerd[1483]: time="2026-01-17T00:16:38.246347310Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 00:16:38.248179 containerd[1483]: time="2026-01-17T00:16:38.246539173Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 00:16:38.248179 containerd[1483]: time="2026-01-17T00:16:38.246574495Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:38.248179 containerd[1483]: time="2026-01-17T00:16:38.246860520Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:38.279402 systemd[1]: Started cri-containerd-577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af.scope - libcontainer container 577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af. Jan 17 00:16:38.337317 containerd[1483]: time="2026-01-17T00:16:38.337204275Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-d2cbv,Uid:5a369a50-610b-4a62-aa4b-07336f6ae1b3,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af\"" Jan 17 00:16:38.338430 kubelet[2470]: E0117 00:16:38.338399 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:38.341731 containerd[1483]: time="2026-01-17T00:16:38.341434678Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\"" Jan 17 00:16:40.163246 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1517108053.mount: Deactivated successfully. Jan 17 00:16:40.202232 containerd[1483]: time="2026-01-17T00:16:40.201585563Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:40.203314 containerd[1483]: time="2026-01-17T00:16:40.203030969Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3852937" Jan 17 00:16:40.204663 containerd[1483]: time="2026-01-17T00:16:40.204261668Z" level=info msg="ImageCreate event name:\"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:40.207931 containerd[1483]: time="2026-01-17T00:16:40.207882794Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:40.208752 containerd[1483]: time="2026-01-17T00:16:40.208690766Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3842055\" in 1.867198478s" Jan 17 00:16:40.208876 containerd[1483]: time="2026-01-17T00:16:40.208757411Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:7a2dcab94698c786e7e41360faf8cd0ea2b29952469be75becc34c61902240e0\"" Jan 17 00:16:40.212904 containerd[1483]: time="2026-01-17T00:16:40.212844163Z" level=info msg="CreateContainer within sandbox \"577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Jan 17 00:16:40.239202 containerd[1483]: time="2026-01-17T00:16:40.239033688Z" level=info msg="CreateContainer within sandbox \"577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749\"" Jan 17 00:16:40.240892 containerd[1483]: time="2026-01-17T00:16:40.240373254Z" level=info msg="StartContainer for \"46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749\"" Jan 17 00:16:40.278381 systemd[1]: Started cri-containerd-46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749.scope - libcontainer container 46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749. Jan 17 00:16:40.319294 containerd[1483]: time="2026-01-17T00:16:40.318526660Z" level=info msg="StartContainer for \"46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749\" returns successfully" Jan 17 00:16:40.318796 systemd[1]: cri-containerd-46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749.scope: Deactivated successfully. Jan 17 00:16:40.356900 containerd[1483]: time="2026-01-17T00:16:40.356796406Z" level=info msg="shim disconnected" id=46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749 namespace=k8s.io Jan 17 00:16:40.356900 containerd[1483]: time="2026-01-17T00:16:40.356889422Z" level=warning msg="cleaning up after shim disconnected" id=46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749 namespace=k8s.io Jan 17 00:16:40.356900 containerd[1483]: time="2026-01-17T00:16:40.356898585Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 17 00:16:40.825847 kubelet[2470]: E0117 00:16:40.825792 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:40.851981 kubelet[2470]: I0117 00:16:40.851881 2470 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-qrp95" podStartSLOduration=4.8518541200000005 podStartE2EDuration="4.85185412s" podCreationTimestamp="2026-01-17 00:16:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-17 00:16:37.277582659 +0000 UTC m=+5.250568070" watchObservedRunningTime="2026-01-17 00:16:40.85185412 +0000 UTC m=+8.824839529" Jan 17 00:16:41.056866 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-46c719641c5eb7cb90b1fada64ab21f300ad447a7302453233400b14b4cb7749-rootfs.mount: Deactivated successfully. Jan 17 00:16:41.281187 kubelet[2470]: E0117 00:16:41.279644 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:41.281187 kubelet[2470]: E0117 00:16:41.279648 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:41.282466 containerd[1483]: time="2026-01-17T00:16:41.282023114Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\"" Jan 17 00:16:41.697218 kubelet[2470]: E0117 00:16:41.696602 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:41.941230 kubelet[2470]: E0117 00:16:41.941056 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:42.283680 kubelet[2470]: E0117 00:16:42.283329 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:42.283680 kubelet[2470]: E0117 00:16:42.283329 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:42.361187 update_engine[1455]: I20260117 00:16:42.360615 1455 update_attempter.cc:509] Updating boot flags... Jan 17 00:16:42.400235 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (2844) Jan 17 00:16:42.480223 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (2845) Jan 17 00:16:42.552170 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 33 scanned by (udev-worker) (2845) Jan 17 00:16:43.286221 kubelet[2470]: E0117 00:16:43.286173 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:43.649980 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2254977768.mount: Deactivated successfully. Jan 17 00:16:45.397765 containerd[1483]: time="2026-01-17T00:16:45.397715677Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:45.400473 containerd[1483]: time="2026-01-17T00:16:45.400280791Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26866358" Jan 17 00:16:45.401086 containerd[1483]: time="2026-01-17T00:16:45.401050290Z" level=info msg="ImageCreate event name:\"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:45.404117 containerd[1483]: time="2026-01-17T00:16:45.404067875Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 00:16:45.406162 containerd[1483]: time="2026-01-17T00:16:45.405991115Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26855532\" in 4.123753443s" Jan 17 00:16:45.406162 containerd[1483]: time="2026-01-17T00:16:45.406042626Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:38c11b8f4aa1904512c0b3e93d34604de20ba24b38d4365d27fe05b7a4ce6f68\"" Jan 17 00:16:45.409561 containerd[1483]: time="2026-01-17T00:16:45.409260327Z" level=info msg="CreateContainer within sandbox \"577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 17 00:16:45.427298 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount966400563.mount: Deactivated successfully. Jan 17 00:16:45.431714 containerd[1483]: time="2026-01-17T00:16:45.431555201Z" level=info msg="CreateContainer within sandbox \"577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9\"" Jan 17 00:16:45.435554 containerd[1483]: time="2026-01-17T00:16:45.433286101Z" level=info msg="StartContainer for \"9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9\"" Jan 17 00:16:45.474436 systemd[1]: Started cri-containerd-9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9.scope - libcontainer container 9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9. Jan 17 00:16:45.511113 systemd[1]: cri-containerd-9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9.scope: Deactivated successfully. Jan 17 00:16:45.513751 containerd[1483]: time="2026-01-17T00:16:45.513360919Z" level=info msg="StartContainer for \"9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9\" returns successfully" Jan 17 00:16:45.543071 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9-rootfs.mount: Deactivated successfully. Jan 17 00:16:45.570361 containerd[1483]: time="2026-01-17T00:16:45.570238236Z" level=info msg="shim disconnected" id=9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9 namespace=k8s.io Jan 17 00:16:45.570361 containerd[1483]: time="2026-01-17T00:16:45.570357797Z" level=warning msg="cleaning up after shim disconnected" id=9e61f452d7c1d4ce9a0d1b9393e945f1fe561f535489f6903d0f4645634917f9 namespace=k8s.io Jan 17 00:16:45.570620 containerd[1483]: time="2026-01-17T00:16:45.570374901Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 17 00:16:45.607306 kubelet[2470]: I0117 00:16:45.606934 2470 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Jan 17 00:16:45.652344 systemd[1]: Created slice kubepods-burstable-podb950289f_cccf_4033_9caa_3145ebf1198b.slice - libcontainer container kubepods-burstable-podb950289f_cccf_4033_9caa_3145ebf1198b.slice. Jan 17 00:16:45.665336 systemd[1]: Created slice kubepods-burstable-pod8cc6f108_eadc_48fe_a878_07a846028871.slice - libcontainer container kubepods-burstable-pod8cc6f108_eadc_48fe_a878_07a846028871.slice. Jan 17 00:16:45.689303 kubelet[2470]: I0117 00:16:45.689242 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b950289f-cccf-4033-9caa-3145ebf1198b-config-volume\") pod \"coredns-668d6bf9bc-6s6hj\" (UID: \"b950289f-cccf-4033-9caa-3145ebf1198b\") " pod="kube-system/coredns-668d6bf9bc-6s6hj" Jan 17 00:16:45.689553 kubelet[2470]: I0117 00:16:45.689338 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cc6f108-eadc-48fe-a878-07a846028871-config-volume\") pod \"coredns-668d6bf9bc-cfdns\" (UID: \"8cc6f108-eadc-48fe-a878-07a846028871\") " pod="kube-system/coredns-668d6bf9bc-cfdns" Jan 17 00:16:45.689553 kubelet[2470]: I0117 00:16:45.689367 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vcqb\" (UniqueName: \"kubernetes.io/projected/b950289f-cccf-4033-9caa-3145ebf1198b-kube-api-access-2vcqb\") pod \"coredns-668d6bf9bc-6s6hj\" (UID: \"b950289f-cccf-4033-9caa-3145ebf1198b\") " pod="kube-system/coredns-668d6bf9bc-6s6hj" Jan 17 00:16:45.689553 kubelet[2470]: I0117 00:16:45.689412 2470 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r82q7\" (UniqueName: \"kubernetes.io/projected/8cc6f108-eadc-48fe-a878-07a846028871-kube-api-access-r82q7\") pod \"coredns-668d6bf9bc-cfdns\" (UID: \"8cc6f108-eadc-48fe-a878-07a846028871\") " pod="kube-system/coredns-668d6bf9bc-cfdns" Jan 17 00:16:45.956394 kubelet[2470]: E0117 00:16:45.956226 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:45.957563 containerd[1483]: time="2026-01-17T00:16:45.957495327Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6s6hj,Uid:b950289f-cccf-4033-9caa-3145ebf1198b,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:45.978526 kubelet[2470]: E0117 00:16:45.977178 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:45.984871 containerd[1483]: time="2026-01-17T00:16:45.984796728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-cfdns,Uid:8cc6f108-eadc-48fe-a878-07a846028871,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:46.007698 containerd[1483]: time="2026-01-17T00:16:46.007581779Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6s6hj,Uid:b950289f-cccf-4033-9caa-3145ebf1198b,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"db0a999d62550a169d8da022ae03f5692cf6d86a82bcdb52dda7cb819c78863c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 17 00:16:46.008514 kubelet[2470]: E0117 00:16:46.008452 2470 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db0a999d62550a169d8da022ae03f5692cf6d86a82bcdb52dda7cb819c78863c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 17 00:16:46.008678 kubelet[2470]: E0117 00:16:46.008562 2470 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db0a999d62550a169d8da022ae03f5692cf6d86a82bcdb52dda7cb819c78863c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-6s6hj" Jan 17 00:16:46.008678 kubelet[2470]: E0117 00:16:46.008602 2470 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"db0a999d62550a169d8da022ae03f5692cf6d86a82bcdb52dda7cb819c78863c\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-6s6hj" Jan 17 00:16:46.008678 kubelet[2470]: E0117 00:16:46.008657 2470 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-6s6hj_kube-system(b950289f-cccf-4033-9caa-3145ebf1198b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-6s6hj_kube-system(b950289f-cccf-4033-9caa-3145ebf1198b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"db0a999d62550a169d8da022ae03f5692cf6d86a82bcdb52dda7cb819c78863c\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-668d6bf9bc-6s6hj" podUID="b950289f-cccf-4033-9caa-3145ebf1198b" Jan 17 00:16:46.018409 containerd[1483]: time="2026-01-17T00:16:46.018356196Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-cfdns,Uid:8cc6f108-eadc-48fe-a878-07a846028871,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"189823230f520bda78698854d22ba366afc39e58a4f7293935ac9e397a6488be\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 17 00:16:46.019769 kubelet[2470]: E0117 00:16:46.018789 2470 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"189823230f520bda78698854d22ba366afc39e58a4f7293935ac9e397a6488be\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 17 00:16:46.019769 kubelet[2470]: E0117 00:16:46.018862 2470 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"189823230f520bda78698854d22ba366afc39e58a4f7293935ac9e397a6488be\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-cfdns" Jan 17 00:16:46.019769 kubelet[2470]: E0117 00:16:46.018883 2470 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"189823230f520bda78698854d22ba366afc39e58a4f7293935ac9e397a6488be\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-cfdns" Jan 17 00:16:46.019769 kubelet[2470]: E0117 00:16:46.018928 2470 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-cfdns_kube-system(8cc6f108-eadc-48fe-a878-07a846028871)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-cfdns_kube-system(8cc6f108-eadc-48fe-a878-07a846028871)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"189823230f520bda78698854d22ba366afc39e58a4f7293935ac9e397a6488be\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-668d6bf9bc-cfdns" podUID="8cc6f108-eadc-48fe-a878-07a846028871" Jan 17 00:16:46.294749 kubelet[2470]: E0117 00:16:46.294690 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:46.299067 containerd[1483]: time="2026-01-17T00:16:46.298746457Z" level=info msg="CreateContainer within sandbox \"577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Jan 17 00:16:46.309184 containerd[1483]: time="2026-01-17T00:16:46.309110794Z" level=info msg="CreateContainer within sandbox \"577cfe00ce964b6722c115f1c825cfb1dbef6d08a46898b466c00609518050af\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"27765b5ab154df2fcf6ebad19ce9310e2f0f68161cdce64228a7d4619bab7710\"" Jan 17 00:16:46.310732 containerd[1483]: time="2026-01-17T00:16:46.310654717Z" level=info msg="StartContainer for \"27765b5ab154df2fcf6ebad19ce9310e2f0f68161cdce64228a7d4619bab7710\"" Jan 17 00:16:46.341385 systemd[1]: Started cri-containerd-27765b5ab154df2fcf6ebad19ce9310e2f0f68161cdce64228a7d4619bab7710.scope - libcontainer container 27765b5ab154df2fcf6ebad19ce9310e2f0f68161cdce64228a7d4619bab7710. Jan 17 00:16:46.372906 containerd[1483]: time="2026-01-17T00:16:46.372779020Z" level=info msg="StartContainer for \"27765b5ab154df2fcf6ebad19ce9310e2f0f68161cdce64228a7d4619bab7710\" returns successfully" Jan 17 00:16:47.300800 kubelet[2470]: E0117 00:16:47.300755 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:47.317757 kubelet[2470]: I0117 00:16:47.317676 2470 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-d2cbv" podStartSLOduration=4.249121472 podStartE2EDuration="11.317367392s" podCreationTimestamp="2026-01-17 00:16:36 +0000 UTC" firstStartedPulling="2026-01-17 00:16:38.339147698 +0000 UTC m=+6.312133096" lastFinishedPulling="2026-01-17 00:16:45.407393616 +0000 UTC m=+13.380379016" observedRunningTime="2026-01-17 00:16:47.317323934 +0000 UTC m=+15.290309339" watchObservedRunningTime="2026-01-17 00:16:47.317367392 +0000 UTC m=+15.290352794" Jan 17 00:16:47.443512 systemd-networkd[1378]: flannel.1: Link UP Jan 17 00:16:47.443523 systemd-networkd[1378]: flannel.1: Gained carrier Jan 17 00:16:48.302983 kubelet[2470]: E0117 00:16:48.302944 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:48.855530 systemd-networkd[1378]: flannel.1: Gained IPv6LL Jan 17 00:16:57.207248 kubelet[2470]: E0117 00:16:57.207001 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:57.208613 containerd[1483]: time="2026-01-17T00:16:57.208171300Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6s6hj,Uid:b950289f-cccf-4033-9caa-3145ebf1198b,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:57.244636 systemd-networkd[1378]: cni0: Link UP Jan 17 00:16:57.244648 systemd-networkd[1378]: cni0: Gained carrier Jan 17 00:16:57.250305 systemd-networkd[1378]: cni0: Lost carrier Jan 17 00:16:57.258772 systemd-networkd[1378]: veth5646f5df: Link UP Jan 17 00:16:57.260319 kernel: cni0: port 1(veth5646f5df) entered blocking state Jan 17 00:16:57.260425 kernel: cni0: port 1(veth5646f5df) entered disabled state Jan 17 00:16:57.262175 kernel: veth5646f5df: entered allmulticast mode Jan 17 00:16:57.265376 kernel: veth5646f5df: entered promiscuous mode Jan 17 00:16:57.265480 kernel: cni0: port 1(veth5646f5df) entered blocking state Jan 17 00:16:57.265499 kernel: cni0: port 1(veth5646f5df) entered forwarding state Jan 17 00:16:57.267334 kernel: cni0: port 1(veth5646f5df) entered disabled state Jan 17 00:16:57.283590 kernel: cni0: port 1(veth5646f5df) entered blocking state Jan 17 00:16:57.283876 kernel: cni0: port 1(veth5646f5df) entered forwarding state Jan 17 00:16:57.287960 systemd-networkd[1378]: veth5646f5df: Gained carrier Jan 17 00:16:57.288756 systemd-networkd[1378]: cni0: Gained carrier Jan 17 00:16:57.302607 containerd[1483]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc0000908d8), "name":"cbr0", "type":"bridge"} Jan 17 00:16:57.302607 containerd[1483]: delegateAdd: netconf sent to delegate plugin: Jan 17 00:16:57.343445 containerd[1483]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2026-01-17T00:16:57.343205990Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 00:16:57.343445 containerd[1483]: time="2026-01-17T00:16:57.343287934Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 00:16:57.343445 containerd[1483]: time="2026-01-17T00:16:57.343304825Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:57.343838 containerd[1483]: time="2026-01-17T00:16:57.343588935Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:57.383551 systemd[1]: Started cri-containerd-64a3b4e3971a3b19ad4ba4f59f6a6115627eeaa6e99d1ead850376e1ad40eca8.scope - libcontainer container 64a3b4e3971a3b19ad4ba4f59f6a6115627eeaa6e99d1ead850376e1ad40eca8. Jan 17 00:16:57.445928 containerd[1483]: time="2026-01-17T00:16:57.445856069Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6s6hj,Uid:b950289f-cccf-4033-9caa-3145ebf1198b,Namespace:kube-system,Attempt:0,} returns sandbox id \"64a3b4e3971a3b19ad4ba4f59f6a6115627eeaa6e99d1ead850376e1ad40eca8\"" Jan 17 00:16:57.447508 kubelet[2470]: E0117 00:16:57.447477 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:57.454744 containerd[1483]: time="2026-01-17T00:16:57.454649535Z" level=info msg="CreateContainer within sandbox \"64a3b4e3971a3b19ad4ba4f59f6a6115627eeaa6e99d1ead850376e1ad40eca8\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 17 00:16:57.478416 containerd[1483]: time="2026-01-17T00:16:57.474543205Z" level=info msg="CreateContainer within sandbox \"64a3b4e3971a3b19ad4ba4f59f6a6115627eeaa6e99d1ead850376e1ad40eca8\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"303992d6ed9af5b4926b25b870159e6e8530b59f3e3ecf628e27ab6e46990840\"" Jan 17 00:16:57.479423 containerd[1483]: time="2026-01-17T00:16:57.478772686Z" level=info msg="StartContainer for \"303992d6ed9af5b4926b25b870159e6e8530b59f3e3ecf628e27ab6e46990840\"" Jan 17 00:16:57.481101 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount271188750.mount: Deactivated successfully. Jan 17 00:16:57.524502 systemd[1]: Started cri-containerd-303992d6ed9af5b4926b25b870159e6e8530b59f3e3ecf628e27ab6e46990840.scope - libcontainer container 303992d6ed9af5b4926b25b870159e6e8530b59f3e3ecf628e27ab6e46990840. Jan 17 00:16:57.571387 containerd[1483]: time="2026-01-17T00:16:57.571298263Z" level=info msg="StartContainer for \"303992d6ed9af5b4926b25b870159e6e8530b59f3e3ecf628e27ab6e46990840\" returns successfully" Jan 17 00:16:58.208218 kubelet[2470]: E0117 00:16:58.207963 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:58.209398 containerd[1483]: time="2026-01-17T00:16:58.209071275Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-cfdns,Uid:8cc6f108-eadc-48fe-a878-07a846028871,Namespace:kube-system,Attempt:0,}" Jan 17 00:16:58.251472 kernel: cni0: port 2(veth24b3cea0) entered blocking state Jan 17 00:16:58.251592 kernel: cni0: port 2(veth24b3cea0) entered disabled state Jan 17 00:16:58.251628 kernel: veth24b3cea0: entered allmulticast mode Jan 17 00:16:58.250304 systemd-networkd[1378]: veth24b3cea0: Link UP Jan 17 00:16:58.253497 kernel: veth24b3cea0: entered promiscuous mode Jan 17 00:16:58.256580 kernel: cni0: port 2(veth24b3cea0) entered blocking state Jan 17 00:16:58.256699 kernel: cni0: port 2(veth24b3cea0) entered forwarding state Jan 17 00:16:58.257298 kernel: cni0: port 2(veth24b3cea0) entered disabled state Jan 17 00:16:58.268319 kernel: cni0: port 2(veth24b3cea0) entered blocking state Jan 17 00:16:58.268437 kernel: cni0: port 2(veth24b3cea0) entered forwarding state Jan 17 00:16:58.268861 systemd-networkd[1378]: veth24b3cea0: Gained carrier Jan 17 00:16:58.272975 containerd[1483]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0xc000018938), "name":"cbr0", "type":"bridge"} Jan 17 00:16:58.272975 containerd[1483]: delegateAdd: netconf sent to delegate plugin: Jan 17 00:16:58.303911 containerd[1483]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2026-01-17T00:16:58.303553545Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 00:16:58.303911 containerd[1483]: time="2026-01-17T00:16:58.303627867Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 00:16:58.303911 containerd[1483]: time="2026-01-17T00:16:58.303643110Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:58.303911 containerd[1483]: time="2026-01-17T00:16:58.303743803Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 00:16:58.327428 systemd[1]: run-containerd-runc-k8s.io-8c88f86be55cb1c0b4c51a5cd3663da49e953bac2dcd96abf29d11d5cc1c6082-runc.R6XDDv.mount: Deactivated successfully. Jan 17 00:16:58.330333 kubelet[2470]: E0117 00:16:58.330295 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:58.342035 systemd[1]: Started cri-containerd-8c88f86be55cb1c0b4c51a5cd3663da49e953bac2dcd96abf29d11d5cc1c6082.scope - libcontainer container 8c88f86be55cb1c0b4c51a5cd3663da49e953bac2dcd96abf29d11d5cc1c6082. Jan 17 00:16:58.391978 kubelet[2470]: I0117 00:16:58.391920 2470 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-6s6hj" podStartSLOduration=22.391902181 podStartE2EDuration="22.391902181s" podCreationTimestamp="2026-01-17 00:16:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-17 00:16:58.367881024 +0000 UTC m=+26.340866430" watchObservedRunningTime="2026-01-17 00:16:58.391902181 +0000 UTC m=+26.364887588" Jan 17 00:16:58.392686 systemd-networkd[1378]: veth5646f5df: Gained IPv6LL Jan 17 00:16:58.419529 containerd[1483]: time="2026-01-17T00:16:58.419454151Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-cfdns,Uid:8cc6f108-eadc-48fe-a878-07a846028871,Namespace:kube-system,Attempt:0,} returns sandbox id \"8c88f86be55cb1c0b4c51a5cd3663da49e953bac2dcd96abf29d11d5cc1c6082\"" Jan 17 00:16:58.421827 kubelet[2470]: E0117 00:16:58.421781 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:58.425836 containerd[1483]: time="2026-01-17T00:16:58.425773848Z" level=info msg="CreateContainer within sandbox \"8c88f86be55cb1c0b4c51a5cd3663da49e953bac2dcd96abf29d11d5cc1c6082\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 17 00:16:58.441403 containerd[1483]: time="2026-01-17T00:16:58.441337071Z" level=info msg="CreateContainer within sandbox \"8c88f86be55cb1c0b4c51a5cd3663da49e953bac2dcd96abf29d11d5cc1c6082\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ba015f8b9b3b9a3a4fbc0823820627cd9f0335c02a94dfb47869f2c52c24f438\"" Jan 17 00:16:58.443970 containerd[1483]: time="2026-01-17T00:16:58.442433842Z" level=info msg="StartContainer for \"ba015f8b9b3b9a3a4fbc0823820627cd9f0335c02a94dfb47869f2c52c24f438\"" Jan 17 00:16:58.455599 systemd-networkd[1378]: cni0: Gained IPv6LL Jan 17 00:16:58.479376 systemd[1]: Started cri-containerd-ba015f8b9b3b9a3a4fbc0823820627cd9f0335c02a94dfb47869f2c52c24f438.scope - libcontainer container ba015f8b9b3b9a3a4fbc0823820627cd9f0335c02a94dfb47869f2c52c24f438. Jan 17 00:16:58.512672 containerd[1483]: time="2026-01-17T00:16:58.512534100Z" level=info msg="StartContainer for \"ba015f8b9b3b9a3a4fbc0823820627cd9f0335c02a94dfb47869f2c52c24f438\" returns successfully" Jan 17 00:16:59.335410 kubelet[2470]: E0117 00:16:59.335079 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:59.335410 kubelet[2470]: E0117 00:16:59.335260 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:16:59.351137 kubelet[2470]: I0117 00:16:59.350591 2470 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-cfdns" podStartSLOduration=23.35056878 podStartE2EDuration="23.35056878s" podCreationTimestamp="2026-01-17 00:16:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-17 00:16:59.350411972 +0000 UTC m=+27.323397381" watchObservedRunningTime="2026-01-17 00:16:59.35056878 +0000 UTC m=+27.323554180" Jan 17 00:16:59.735490 systemd-networkd[1378]: veth24b3cea0: Gained IPv6LL Jan 17 00:17:00.337294 kubelet[2470]: E0117 00:17:00.336812 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:17:09.335774 kubelet[2470]: E0117 00:17:09.335563 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:17:09.357163 kubelet[2470]: E0117 00:17:09.355846 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:17:14.281999 systemd[1]: Started sshd@5-143.198.57.57:22-4.153.228.146:57752.service - OpenSSH per-connection server daemon (4.153.228.146:57752). Jan 17 00:17:14.714724 sshd[3452]: Accepted publickey for core from 4.153.228.146 port 57752 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:14.718254 sshd[3452]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:14.725473 systemd-logind[1452]: New session 6 of user core. Jan 17 00:17:14.730416 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 17 00:17:15.143037 sshd[3452]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:15.146923 systemd[1]: sshd@5-143.198.57.57:22-4.153.228.146:57752.service: Deactivated successfully. Jan 17 00:17:15.148831 systemd[1]: session-6.scope: Deactivated successfully. Jan 17 00:17:15.149888 systemd-logind[1452]: Session 6 logged out. Waiting for processes to exit. Jan 17 00:17:15.151452 systemd-logind[1452]: Removed session 6. Jan 17 00:17:20.211518 systemd[1]: Started sshd@6-143.198.57.57:22-4.153.228.146:43220.service - OpenSSH per-connection server daemon (4.153.228.146:43220). Jan 17 00:17:20.597832 sshd[3487]: Accepted publickey for core from 4.153.228.146 port 43220 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:20.600117 sshd[3487]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:20.608411 systemd-logind[1452]: New session 7 of user core. Jan 17 00:17:20.613463 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 17 00:17:20.975065 sshd[3487]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:20.981674 systemd-logind[1452]: Session 7 logged out. Waiting for processes to exit. Jan 17 00:17:20.982485 systemd[1]: sshd@6-143.198.57.57:22-4.153.228.146:43220.service: Deactivated successfully. Jan 17 00:17:20.985347 systemd[1]: session-7.scope: Deactivated successfully. Jan 17 00:17:20.988027 systemd-logind[1452]: Removed session 7. Jan 17 00:17:26.051583 systemd[1]: Started sshd@7-143.198.57.57:22-4.153.228.146:55892.service - OpenSSH per-connection server daemon (4.153.228.146:55892). Jan 17 00:17:26.451244 sshd[3522]: Accepted publickey for core from 4.153.228.146 port 55892 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:26.453366 sshd[3522]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:26.459363 systemd-logind[1452]: New session 8 of user core. Jan 17 00:17:26.465460 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 17 00:17:26.811793 sshd[3522]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:26.816266 systemd[1]: sshd@7-143.198.57.57:22-4.153.228.146:55892.service: Deactivated successfully. Jan 17 00:17:26.819142 systemd[1]: session-8.scope: Deactivated successfully. Jan 17 00:17:26.820587 systemd-logind[1452]: Session 8 logged out. Waiting for processes to exit. Jan 17 00:17:26.821786 systemd-logind[1452]: Removed session 8. Jan 17 00:17:26.886589 systemd[1]: Started sshd@8-143.198.57.57:22-4.153.228.146:55908.service - OpenSSH per-connection server daemon (4.153.228.146:55908). Jan 17 00:17:27.282178 sshd[3536]: Accepted publickey for core from 4.153.228.146 port 55908 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:27.284084 sshd[3536]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:27.290221 systemd-logind[1452]: New session 9 of user core. Jan 17 00:17:27.302558 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 17 00:17:27.716479 sshd[3536]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:27.722905 systemd[1]: sshd@8-143.198.57.57:22-4.153.228.146:55908.service: Deactivated successfully. Jan 17 00:17:27.727052 systemd[1]: session-9.scope: Deactivated successfully. Jan 17 00:17:27.728766 systemd-logind[1452]: Session 9 logged out. Waiting for processes to exit. Jan 17 00:17:27.730158 systemd-logind[1452]: Removed session 9. Jan 17 00:17:27.809316 systemd[1]: Started sshd@9-143.198.57.57:22-4.153.228.146:55918.service - OpenSSH per-connection server daemon (4.153.228.146:55918). Jan 17 00:17:28.263784 sshd[3566]: Accepted publickey for core from 4.153.228.146 port 55918 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:28.267328 sshd[3566]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:28.277665 systemd-logind[1452]: New session 10 of user core. Jan 17 00:17:28.284449 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 17 00:17:28.666729 sshd[3566]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:28.671292 systemd-logind[1452]: Session 10 logged out. Waiting for processes to exit. Jan 17 00:17:28.671496 systemd[1]: sshd@9-143.198.57.57:22-4.153.228.146:55918.service: Deactivated successfully. Jan 17 00:17:28.674688 systemd[1]: session-10.scope: Deactivated successfully. Jan 17 00:17:28.677963 systemd-logind[1452]: Removed session 10. Jan 17 00:17:33.738568 systemd[1]: Started sshd@10-143.198.57.57:22-4.153.228.146:55926.service - OpenSSH per-connection server daemon (4.153.228.146:55926). Jan 17 00:17:34.124675 sshd[3604]: Accepted publickey for core from 4.153.228.146 port 55926 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:34.127243 sshd[3604]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:34.134468 systemd-logind[1452]: New session 11 of user core. Jan 17 00:17:34.139351 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 17 00:17:34.479150 sshd[3604]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:34.483467 systemd[1]: sshd@10-143.198.57.57:22-4.153.228.146:55926.service: Deactivated successfully. Jan 17 00:17:34.485352 systemd[1]: session-11.scope: Deactivated successfully. Jan 17 00:17:34.486210 systemd-logind[1452]: Session 11 logged out. Waiting for processes to exit. Jan 17 00:17:34.487549 systemd-logind[1452]: Removed session 11. Jan 17 00:17:39.206896 kubelet[2470]: E0117 00:17:39.206844 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:17:39.557005 systemd[1]: Started sshd@11-143.198.57.57:22-4.153.228.146:35436.service - OpenSSH per-connection server daemon (4.153.228.146:35436). Jan 17 00:17:39.996021 sshd[3639]: Accepted publickey for core from 4.153.228.146 port 35436 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:39.997875 sshd[3639]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:40.005226 systemd-logind[1452]: New session 12 of user core. Jan 17 00:17:40.010486 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 17 00:17:40.378247 sshd[3639]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:40.382310 systemd[1]: sshd@11-143.198.57.57:22-4.153.228.146:35436.service: Deactivated successfully. Jan 17 00:17:40.384570 systemd[1]: session-12.scope: Deactivated successfully. Jan 17 00:17:40.385572 systemd-logind[1452]: Session 12 logged out. Waiting for processes to exit. Jan 17 00:17:40.386701 systemd-logind[1452]: Removed session 12. Jan 17 00:17:40.474639 systemd[1]: Started sshd@12-143.198.57.57:22-4.153.228.146:35448.service - OpenSSH per-connection server daemon (4.153.228.146:35448). Jan 17 00:17:40.937352 sshd[3652]: Accepted publickey for core from 4.153.228.146 port 35448 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:40.939717 sshd[3652]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:40.946380 systemd-logind[1452]: New session 13 of user core. Jan 17 00:17:40.953500 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 17 00:17:41.613383 sshd[3652]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:41.618350 systemd-logind[1452]: Session 13 logged out. Waiting for processes to exit. Jan 17 00:17:41.618989 systemd[1]: sshd@12-143.198.57.57:22-4.153.228.146:35448.service: Deactivated successfully. Jan 17 00:17:41.623714 systemd[1]: session-13.scope: Deactivated successfully. Jan 17 00:17:41.625449 systemd-logind[1452]: Removed session 13. Jan 17 00:17:41.706826 systemd[1]: Started sshd@13-143.198.57.57:22-4.153.228.146:35452.service - OpenSSH per-connection server daemon (4.153.228.146:35452). Jan 17 00:17:42.170728 sshd[3662]: Accepted publickey for core from 4.153.228.146 port 35452 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:42.173591 sshd[3662]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:42.180466 systemd-logind[1452]: New session 14 of user core. Jan 17 00:17:42.186441 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 17 00:17:43.118732 sshd[3662]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:43.122973 systemd[1]: sshd@13-143.198.57.57:22-4.153.228.146:35452.service: Deactivated successfully. Jan 17 00:17:43.125624 systemd[1]: session-14.scope: Deactivated successfully. Jan 17 00:17:43.126745 systemd-logind[1452]: Session 14 logged out. Waiting for processes to exit. Jan 17 00:17:43.128356 systemd-logind[1452]: Removed session 14. Jan 17 00:17:43.206359 systemd[1]: Started sshd@14-143.198.57.57:22-4.153.228.146:35456.service - OpenSSH per-connection server daemon (4.153.228.146:35456). Jan 17 00:17:43.629302 sshd[3703]: Accepted publickey for core from 4.153.228.146 port 35456 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:43.632157 sshd[3703]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:43.639657 systemd-logind[1452]: New session 15 of user core. Jan 17 00:17:43.642403 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 17 00:17:44.160358 sshd[3703]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:44.164840 systemd-logind[1452]: Session 15 logged out. Waiting for processes to exit. Jan 17 00:17:44.165371 systemd[1]: sshd@14-143.198.57.57:22-4.153.228.146:35456.service: Deactivated successfully. Jan 17 00:17:44.168685 systemd[1]: session-15.scope: Deactivated successfully. Jan 17 00:17:44.171645 systemd-logind[1452]: Removed session 15. Jan 17 00:17:44.235630 systemd[1]: Started sshd@15-143.198.57.57:22-4.153.228.146:35470.service - OpenSSH per-connection server daemon (4.153.228.146:35470). Jan 17 00:17:44.627212 sshd[3714]: Accepted publickey for core from 4.153.228.146 port 35470 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:44.629071 sshd[3714]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:44.635609 systemd-logind[1452]: New session 16 of user core. Jan 17 00:17:44.650490 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 17 00:17:44.983987 sshd[3714]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:44.989798 systemd[1]: sshd@15-143.198.57.57:22-4.153.228.146:35470.service: Deactivated successfully. Jan 17 00:17:44.989850 systemd-logind[1452]: Session 16 logged out. Waiting for processes to exit. Jan 17 00:17:44.993285 systemd[1]: session-16.scope: Deactivated successfully. Jan 17 00:17:44.994926 systemd-logind[1452]: Removed session 16. Jan 17 00:17:50.072683 systemd[1]: Started sshd@16-143.198.57.57:22-4.153.228.146:45012.service - OpenSSH per-connection server daemon (4.153.228.146:45012). Jan 17 00:17:50.499957 sshd[3750]: Accepted publickey for core from 4.153.228.146 port 45012 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:50.502086 sshd[3750]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:50.508589 systemd-logind[1452]: New session 17 of user core. Jan 17 00:17:50.513458 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 17 00:17:50.873865 sshd[3750]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:50.878600 systemd[1]: sshd@16-143.198.57.57:22-4.153.228.146:45012.service: Deactivated successfully. Jan 17 00:17:50.880567 systemd[1]: session-17.scope: Deactivated successfully. Jan 17 00:17:50.881395 systemd-logind[1452]: Session 17 logged out. Waiting for processes to exit. Jan 17 00:17:50.883088 systemd-logind[1452]: Removed session 17. Jan 17 00:17:51.207722 kubelet[2470]: E0117 00:17:51.207181 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:17:51.207722 kubelet[2470]: E0117 00:17:51.207337 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:17:51.207722 kubelet[2470]: E0117 00:17:51.207681 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Jan 17 00:17:55.944512 systemd[1]: Started sshd@17-143.198.57.57:22-4.153.228.146:58042.service - OpenSSH per-connection server daemon (4.153.228.146:58042). Jan 17 00:17:56.340967 sshd[3785]: Accepted publickey for core from 4.153.228.146 port 58042 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:17:56.342979 sshd[3785]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:17:56.348235 systemd-logind[1452]: New session 18 of user core. Jan 17 00:17:56.353427 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 17 00:17:56.693587 sshd[3785]: pam_unix(sshd:session): session closed for user core Jan 17 00:17:56.698178 systemd[1]: sshd@17-143.198.57.57:22-4.153.228.146:58042.service: Deactivated successfully. Jan 17 00:17:56.702679 systemd[1]: session-18.scope: Deactivated successfully. Jan 17 00:17:56.706703 systemd-logind[1452]: Session 18 logged out. Waiting for processes to exit. Jan 17 00:17:56.708368 systemd-logind[1452]: Removed session 18. Jan 17 00:18:01.778763 systemd[1]: Started sshd@18-143.198.57.57:22-4.153.228.146:58046.service - OpenSSH per-connection server daemon (4.153.228.146:58046). Jan 17 00:18:02.216879 sshd[3818]: Accepted publickey for core from 4.153.228.146 port 58046 ssh2: RSA SHA256:d1xssXCxZ7/RICQNTzGJeDFE6NneBADHoj85LlPFNm8 Jan 17 00:18:02.220013 sshd[3818]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 00:18:02.228295 systemd-logind[1452]: New session 19 of user core. Jan 17 00:18:02.235460 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 17 00:18:02.618591 sshd[3818]: pam_unix(sshd:session): session closed for user core Jan 17 00:18:02.623946 systemd[1]: sshd@18-143.198.57.57:22-4.153.228.146:58046.service: Deactivated successfully. Jan 17 00:18:02.626554 systemd[1]: session-19.scope: Deactivated successfully. Jan 17 00:18:02.628179 systemd-logind[1452]: Session 19 logged out. Waiting for processes to exit. Jan 17 00:18:02.630084 systemd-logind[1452]: Removed session 19. Jan 17 00:18:04.207939 kubelet[2470]: E0117 00:18:04.207893 2470 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3"