May 17 00:21:56.973468 kernel: Linux version 6.6.90-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri May 16 22:44:56 -00 2025 May 17 00:21:56.973513 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=6b60288baeea1613a76a6f06a8f0e8edc178eae4857ce00eac42d48e92ed015e May 17 00:21:56.973530 kernel: BIOS-provided physical RAM map: May 17 00:21:56.973537 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable May 17 00:21:56.973543 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved May 17 00:21:56.973550 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved May 17 00:21:56.973558 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable May 17 00:21:56.973565 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved May 17 00:21:56.973571 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 17 00:21:56.973582 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved May 17 00:21:56.973590 kernel: NX (Execute Disable) protection: active May 17 00:21:56.973597 kernel: APIC: Static calls initialized May 17 00:21:56.973610 kernel: SMBIOS 2.8 present. May 17 00:21:56.973617 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 May 17 00:21:56.973627 kernel: Hypervisor detected: KVM May 17 00:21:56.973638 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 17 00:21:56.973649 kernel: kvm-clock: using sched offset of 3038208623 cycles May 17 00:21:56.973658 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 17 00:21:56.973667 kernel: tsc: Detected 2494.136 MHz processor May 17 00:21:56.973675 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 17 00:21:56.973684 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 17 00:21:56.973692 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 May 17 00:21:56.973700 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs May 17 00:21:56.973708 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 17 00:21:56.973719 kernel: ACPI: Early table checksum verification disabled May 17 00:21:56.973727 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) May 17 00:21:56.973735 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 17 00:21:56.973743 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 17 00:21:56.973751 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) May 17 00:21:56.973759 kernel: ACPI: FACS 0x000000007FFE0000 000040 May 17 00:21:56.973767 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 17 00:21:56.973774 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 17 00:21:56.973783 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 17 00:21:56.973794 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 17 00:21:56.973802 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] May 17 00:21:56.973810 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] May 17 00:21:56.973818 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] May 17 00:21:56.973826 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] May 17 00:21:56.973834 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] May 17 00:21:56.973842 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] May 17 00:21:56.973858 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] May 17 00:21:56.973866 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 May 17 00:21:56.973874 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 May 17 00:21:56.973883 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] May 17 00:21:56.973891 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] May 17 00:21:56.973901 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] May 17 00:21:56.973910 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] May 17 00:21:56.973922 kernel: Zone ranges: May 17 00:21:56.973930 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 17 00:21:56.973938 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] May 17 00:21:56.973946 kernel: Normal empty May 17 00:21:56.973954 kernel: Movable zone start for each node May 17 00:21:56.973963 kernel: Early memory node ranges May 17 00:21:56.973971 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] May 17 00:21:56.973979 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] May 17 00:21:56.973988 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] May 17 00:21:56.974000 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 17 00:21:56.974008 kernel: On node 0, zone DMA: 97 pages in unavailable ranges May 17 00:21:56.974018 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges May 17 00:21:56.974026 kernel: ACPI: PM-Timer IO Port: 0x608 May 17 00:21:56.974035 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 17 00:21:56.974043 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 17 00:21:56.974051 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 17 00:21:56.974059 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 17 00:21:56.974068 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 17 00:21:56.974079 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 17 00:21:56.974088 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 17 00:21:56.974096 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 17 00:21:56.974104 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 17 00:21:56.974112 kernel: TSC deadline timer available May 17 00:21:56.974120 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs May 17 00:21:56.974129 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 17 00:21:56.974137 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices May 17 00:21:56.974147 kernel: Booting paravirtualized kernel on KVM May 17 00:21:56.974156 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 17 00:21:56.974168 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 May 17 00:21:56.974177 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 May 17 00:21:56.974185 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 May 17 00:21:56.974193 kernel: pcpu-alloc: [0] 0 1 May 17 00:21:56.974201 kernel: kvm-guest: PV spinlocks disabled, no host support May 17 00:21:56.974210 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=6b60288baeea1613a76a6f06a8f0e8edc178eae4857ce00eac42d48e92ed015e May 17 00:21:56.974219 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 17 00:21:56.974230 kernel: random: crng init done May 17 00:21:56.974239 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 17 00:21:56.974247 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 17 00:21:56.974255 kernel: Fallback order for Node 0: 0 May 17 00:21:56.974263 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 May 17 00:21:56.974271 kernel: Policy zone: DMA32 May 17 00:21:56.974279 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 17 00:21:56.974288 kernel: Memory: 1971204K/2096612K available (12288K kernel code, 2295K rwdata, 22740K rodata, 42872K init, 2320K bss, 125148K reserved, 0K cma-reserved) May 17 00:21:56.974296 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 17 00:21:56.974326 kernel: Kernel/User page tables isolation: enabled May 17 00:21:56.974335 kernel: ftrace: allocating 37948 entries in 149 pages May 17 00:21:56.974343 kernel: ftrace: allocated 149 pages with 4 groups May 17 00:21:56.974351 kernel: Dynamic Preempt: voluntary May 17 00:21:56.974359 kernel: rcu: Preemptible hierarchical RCU implementation. May 17 00:21:56.974370 kernel: rcu: RCU event tracing is enabled. May 17 00:21:56.974378 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 17 00:21:56.974387 kernel: Trampoline variant of Tasks RCU enabled. May 17 00:21:56.974395 kernel: Rude variant of Tasks RCU enabled. May 17 00:21:56.974408 kernel: Tracing variant of Tasks RCU enabled. May 17 00:21:56.974416 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 17 00:21:56.974424 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 17 00:21:56.974432 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 17 00:21:56.974440 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 17 00:21:56.974451 kernel: Console: colour VGA+ 80x25 May 17 00:21:56.974459 kernel: printk: console [tty0] enabled May 17 00:21:56.974468 kernel: printk: console [ttyS0] enabled May 17 00:21:56.974476 kernel: ACPI: Core revision 20230628 May 17 00:21:56.974489 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns May 17 00:21:56.974506 kernel: APIC: Switch to symmetric I/O mode setup May 17 00:21:56.974519 kernel: x2apic enabled May 17 00:21:56.974532 kernel: APIC: Switched APIC routing to: physical x2apic May 17 00:21:56.974545 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 17 00:21:56.974558 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 17 00:21:56.974572 kernel: Calibrating delay loop (skipped) preset value.. 4988.27 BogoMIPS (lpj=2494136) May 17 00:21:56.974585 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 May 17 00:21:56.974600 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 May 17 00:21:56.974632 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 17 00:21:56.974646 kernel: Spectre V2 : Mitigation: Retpolines May 17 00:21:56.974661 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT May 17 00:21:56.974680 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls May 17 00:21:56.974693 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 17 00:21:56.974708 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl May 17 00:21:56.974718 kernel: MDS: Mitigation: Clear CPU buffers May 17 00:21:56.974726 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 17 00:21:56.974740 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 17 00:21:56.974754 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 17 00:21:56.974763 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 17 00:21:56.974771 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 17 00:21:56.974780 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. May 17 00:21:56.974789 kernel: Freeing SMP alternatives memory: 32K May 17 00:21:56.974798 kernel: pid_max: default: 32768 minimum: 301 May 17 00:21:56.974807 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity May 17 00:21:56.974816 kernel: landlock: Up and running. May 17 00:21:56.974828 kernel: SELinux: Initializing. May 17 00:21:56.974837 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 17 00:21:56.974846 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 17 00:21:56.974855 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) May 17 00:21:56.974864 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 17 00:21:56.974872 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 17 00:21:56.974881 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 17 00:21:56.974890 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. May 17 00:21:56.974902 kernel: signal: max sigframe size: 1776 May 17 00:21:56.974910 kernel: rcu: Hierarchical SRCU implementation. May 17 00:21:56.974919 kernel: rcu: Max phase no-delay instances is 400. May 17 00:21:56.974928 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 17 00:21:56.974937 kernel: smp: Bringing up secondary CPUs ... May 17 00:21:56.974945 kernel: smpboot: x86: Booting SMP configuration: May 17 00:21:56.974954 kernel: .... node #0, CPUs: #1 May 17 00:21:56.974963 kernel: smp: Brought up 1 node, 2 CPUs May 17 00:21:56.974972 kernel: smpboot: Max logical packages: 1 May 17 00:21:56.974983 kernel: smpboot: Total of 2 processors activated (9976.54 BogoMIPS) May 17 00:21:56.974996 kernel: devtmpfs: initialized May 17 00:21:56.975004 kernel: x86/mm: Memory block size: 128MB May 17 00:21:56.975014 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 17 00:21:56.975022 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 17 00:21:56.975031 kernel: pinctrl core: initialized pinctrl subsystem May 17 00:21:56.975040 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 17 00:21:56.975049 kernel: audit: initializing netlink subsys (disabled) May 17 00:21:56.975057 kernel: audit: type=2000 audit(1747441315.423:1): state=initialized audit_enabled=0 res=1 May 17 00:21:56.975066 kernel: thermal_sys: Registered thermal governor 'step_wise' May 17 00:21:56.975078 kernel: thermal_sys: Registered thermal governor 'user_space' May 17 00:21:56.975087 kernel: cpuidle: using governor menu May 17 00:21:56.975096 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 17 00:21:56.975104 kernel: dca service started, version 1.12.1 May 17 00:21:56.975113 kernel: PCI: Using configuration type 1 for base access May 17 00:21:56.975122 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 17 00:21:56.975131 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 17 00:21:56.975140 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 17 00:21:56.975149 kernel: ACPI: Added _OSI(Module Device) May 17 00:21:56.975161 kernel: ACPI: Added _OSI(Processor Device) May 17 00:21:56.975170 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 17 00:21:56.975179 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 17 00:21:56.975187 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 17 00:21:56.975196 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC May 17 00:21:56.975205 kernel: ACPI: Interpreter enabled May 17 00:21:56.975213 kernel: ACPI: PM: (supports S0 S5) May 17 00:21:56.975222 kernel: ACPI: Using IOAPIC for interrupt routing May 17 00:21:56.975231 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 17 00:21:56.975243 kernel: PCI: Using E820 reservations for host bridge windows May 17 00:21:56.975252 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F May 17 00:21:56.975260 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 17 00:21:56.975906 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 17 00:21:56.976023 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] May 17 00:21:56.976119 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge May 17 00:21:56.976131 kernel: acpiphp: Slot [3] registered May 17 00:21:56.976147 kernel: acpiphp: Slot [4] registered May 17 00:21:56.976156 kernel: acpiphp: Slot [5] registered May 17 00:21:56.976165 kernel: acpiphp: Slot [6] registered May 17 00:21:56.976174 kernel: acpiphp: Slot [7] registered May 17 00:21:56.976183 kernel: acpiphp: Slot [8] registered May 17 00:21:56.976191 kernel: acpiphp: Slot [9] registered May 17 00:21:56.976200 kernel: acpiphp: Slot [10] registered May 17 00:21:56.976210 kernel: acpiphp: Slot [11] registered May 17 00:21:56.976218 kernel: acpiphp: Slot [12] registered May 17 00:21:56.976231 kernel: acpiphp: Slot [13] registered May 17 00:21:56.976240 kernel: acpiphp: Slot [14] registered May 17 00:21:56.976249 kernel: acpiphp: Slot [15] registered May 17 00:21:56.976258 kernel: acpiphp: Slot [16] registered May 17 00:21:56.976267 kernel: acpiphp: Slot [17] registered May 17 00:21:56.976275 kernel: acpiphp: Slot [18] registered May 17 00:21:56.976284 kernel: acpiphp: Slot [19] registered May 17 00:21:56.976293 kernel: acpiphp: Slot [20] registered May 17 00:21:56.976302 kernel: acpiphp: Slot [21] registered May 17 00:21:56.976342 kernel: acpiphp: Slot [22] registered May 17 00:21:56.976355 kernel: acpiphp: Slot [23] registered May 17 00:21:56.976364 kernel: acpiphp: Slot [24] registered May 17 00:21:56.976373 kernel: acpiphp: Slot [25] registered May 17 00:21:56.976382 kernel: acpiphp: Slot [26] registered May 17 00:21:56.976391 kernel: acpiphp: Slot [27] registered May 17 00:21:56.976399 kernel: acpiphp: Slot [28] registered May 17 00:21:56.976408 kernel: acpiphp: Slot [29] registered May 17 00:21:56.976417 kernel: acpiphp: Slot [30] registered May 17 00:21:56.976426 kernel: acpiphp: Slot [31] registered May 17 00:21:56.976438 kernel: PCI host bridge to bus 0000:00 May 17 00:21:56.976597 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 17 00:21:56.976686 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 17 00:21:56.976770 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 17 00:21:56.976853 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] May 17 00:21:56.976936 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] May 17 00:21:56.977019 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 17 00:21:56.977151 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 May 17 00:21:56.977265 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 May 17 00:21:56.977388 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 May 17 00:21:56.977486 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] May 17 00:21:56.977582 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] May 17 00:21:56.977674 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] May 17 00:21:56.977768 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] May 17 00:21:56.977866 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] May 17 00:21:56.977970 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 May 17 00:21:56.978064 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] May 17 00:21:56.978175 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 May 17 00:21:56.978269 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI May 17 00:21:56.978372 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB May 17 00:21:56.978487 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 May 17 00:21:56.978585 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] May 17 00:21:56.978680 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] May 17 00:21:56.978773 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] May 17 00:21:56.978867 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] May 17 00:21:56.978960 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 17 00:21:56.979068 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 May 17 00:21:56.979169 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] May 17 00:21:56.979261 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] May 17 00:21:56.979368 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] May 17 00:21:56.979502 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 May 17 00:21:56.979617 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] May 17 00:21:56.979714 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] May 17 00:21:56.979825 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] May 17 00:21:56.979931 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 May 17 00:21:56.980027 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] May 17 00:21:56.980120 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] May 17 00:21:56.980215 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] May 17 00:21:56.980348 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 May 17 00:21:56.980480 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] May 17 00:21:56.980580 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] May 17 00:21:56.980699 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] May 17 00:21:56.980840 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 May 17 00:21:56.980941 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] May 17 00:21:56.981037 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] May 17 00:21:56.981142 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] May 17 00:21:56.981252 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 May 17 00:21:56.981440 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] May 17 00:21:56.981544 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] May 17 00:21:56.981558 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 17 00:21:56.981568 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 17 00:21:56.981577 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 17 00:21:56.981586 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 17 00:21:56.981596 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 17 00:21:56.981613 kernel: iommu: Default domain type: Translated May 17 00:21:56.981623 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 17 00:21:56.981632 kernel: PCI: Using ACPI for IRQ routing May 17 00:21:56.981642 kernel: PCI: pci_cache_line_size set to 64 bytes May 17 00:21:56.981652 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] May 17 00:21:56.981661 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] May 17 00:21:56.981836 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device May 17 00:21:56.981989 kernel: pci 0000:00:02.0: vgaarb: bridge control possible May 17 00:21:56.982092 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 17 00:21:56.982127 kernel: vgaarb: loaded May 17 00:21:56.982147 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 May 17 00:21:56.982156 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter May 17 00:21:56.982165 kernel: clocksource: Switched to clocksource kvm-clock May 17 00:21:56.982175 kernel: VFS: Disk quotas dquot_6.6.0 May 17 00:21:56.982184 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 17 00:21:56.982194 kernel: pnp: PnP ACPI init May 17 00:21:56.982203 kernel: pnp: PnP ACPI: found 4 devices May 17 00:21:56.982212 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 17 00:21:56.982227 kernel: NET: Registered PF_INET protocol family May 17 00:21:56.982236 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) May 17 00:21:56.982245 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) May 17 00:21:56.982258 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 17 00:21:56.982269 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) May 17 00:21:56.982278 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) May 17 00:21:56.982287 kernel: TCP: Hash tables configured (established 16384 bind 16384) May 17 00:21:56.982296 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) May 17 00:21:56.982305 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) May 17 00:21:56.982330 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 17 00:21:56.982339 kernel: NET: Registered PF_XDP protocol family May 17 00:21:56.982448 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 17 00:21:56.982537 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 17 00:21:56.982622 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 17 00:21:56.982717 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] May 17 00:21:56.982854 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] May 17 00:21:56.983010 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release May 17 00:21:56.983174 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 17 00:21:56.983198 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 May 17 00:21:56.983639 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x7b0 took 27993 usecs May 17 00:21:56.983668 kernel: PCI: CLS 0 bytes, default 64 May 17 00:21:56.983686 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 17 00:21:56.983702 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39654230, max_idle_ns: 440795207432 ns May 17 00:21:56.983720 kernel: Initialise system trusted keyrings May 17 00:21:56.983734 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 May 17 00:21:56.983757 kernel: Key type asymmetric registered May 17 00:21:56.983769 kernel: Asymmetric key parser 'x509' registered May 17 00:21:56.983781 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) May 17 00:21:56.983794 kernel: io scheduler mq-deadline registered May 17 00:21:56.983807 kernel: io scheduler kyber registered May 17 00:21:56.983820 kernel: io scheduler bfq registered May 17 00:21:56.983833 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 17 00:21:56.983846 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 May 17 00:21:56.983859 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 May 17 00:21:56.983873 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 May 17 00:21:56.983894 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 17 00:21:56.983907 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 17 00:21:56.983920 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 17 00:21:56.983934 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 17 00:21:56.983949 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 17 00:21:56.983964 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 17 00:21:56.984188 kernel: rtc_cmos 00:03: RTC can wake from S4 May 17 00:21:56.984361 kernel: rtc_cmos 00:03: registered as rtc0 May 17 00:21:56.984481 kernel: rtc_cmos 00:03: setting system clock to 2025-05-17T00:21:56 UTC (1747441316) May 17 00:21:56.984628 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram May 17 00:21:56.984642 kernel: intel_pstate: CPU model not supported May 17 00:21:56.984652 kernel: NET: Registered PF_INET6 protocol family May 17 00:21:56.984662 kernel: Segment Routing with IPv6 May 17 00:21:56.984681 kernel: In-situ OAM (IOAM) with IPv6 May 17 00:21:56.984699 kernel: NET: Registered PF_PACKET protocol family May 17 00:21:56.984712 kernel: Key type dns_resolver registered May 17 00:21:56.984736 kernel: IPI shorthand broadcast: enabled May 17 00:21:56.984752 kernel: sched_clock: Marking stable (920005068, 92411061)->(1117104578, -104688449) May 17 00:21:56.984764 kernel: registered taskstats version 1 May 17 00:21:56.984773 kernel: Loading compiled-in X.509 certificates May 17 00:21:56.984782 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.90-flatcar: 85b8d1234ceca483cb3defc2030d93f7792663c9' May 17 00:21:56.984791 kernel: Key type .fscrypt registered May 17 00:21:56.984800 kernel: Key type fscrypt-provisioning registered May 17 00:21:56.984809 kernel: ima: No TPM chip found, activating TPM-bypass! May 17 00:21:56.984818 kernel: ima: Allocated hash algorithm: sha1 May 17 00:21:56.984831 kernel: ima: No architecture policies found May 17 00:21:56.984840 kernel: clk: Disabling unused clocks May 17 00:21:56.984849 kernel: Freeing unused kernel image (initmem) memory: 42872K May 17 00:21:56.984858 kernel: Write protecting the kernel read-only data: 36864k May 17 00:21:56.984868 kernel: Freeing unused kernel image (rodata/data gap) memory: 1836K May 17 00:21:56.984904 kernel: Run /init as init process May 17 00:21:56.984917 kernel: with arguments: May 17 00:21:56.984927 kernel: /init May 17 00:21:56.984936 kernel: with environment: May 17 00:21:56.984949 kernel: HOME=/ May 17 00:21:56.984958 kernel: TERM=linux May 17 00:21:56.984967 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 17 00:21:56.984980 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) May 17 00:21:56.984992 systemd[1]: Detected virtualization kvm. May 17 00:21:56.985002 systemd[1]: Detected architecture x86-64. May 17 00:21:56.985011 systemd[1]: Running in initrd. May 17 00:21:56.985021 systemd[1]: No hostname configured, using default hostname. May 17 00:21:56.985033 systemd[1]: Hostname set to . May 17 00:21:56.985043 systemd[1]: Initializing machine ID from VM UUID. May 17 00:21:56.985053 systemd[1]: Queued start job for default target initrd.target. May 17 00:21:56.985063 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 17 00:21:56.985073 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 17 00:21:56.985084 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 17 00:21:56.985093 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 17 00:21:56.985103 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 17 00:21:56.985116 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 17 00:21:56.985127 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 17 00:21:56.985137 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 17 00:21:56.985147 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 17 00:21:56.985157 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 17 00:21:56.985166 systemd[1]: Reached target paths.target - Path Units. May 17 00:21:56.985180 systemd[1]: Reached target slices.target - Slice Units. May 17 00:21:56.985189 systemd[1]: Reached target swap.target - Swaps. May 17 00:21:56.985199 systemd[1]: Reached target timers.target - Timer Units. May 17 00:21:56.985212 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 17 00:21:56.985222 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 17 00:21:56.985232 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 17 00:21:56.985245 systemd[1]: Listening on systemd-journald.socket - Journal Socket. May 17 00:21:56.985256 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 17 00:21:56.985266 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 17 00:21:56.985275 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 17 00:21:56.985285 systemd[1]: Reached target sockets.target - Socket Units. May 17 00:21:56.985295 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 17 00:21:56.985304 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 17 00:21:56.985391 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 17 00:21:56.985406 systemd[1]: Starting systemd-fsck-usr.service... May 17 00:21:56.985416 systemd[1]: Starting systemd-journald.service - Journal Service... May 17 00:21:56.985426 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 17 00:21:56.985436 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 17 00:21:56.985446 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 17 00:21:56.985456 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 17 00:21:56.985465 systemd[1]: Finished systemd-fsck-usr.service. May 17 00:21:56.985480 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 17 00:21:56.985525 systemd-journald[183]: Collecting audit messages is disabled. May 17 00:21:56.985556 systemd-journald[183]: Journal started May 17 00:21:56.985579 systemd-journald[183]: Runtime Journal (/run/log/journal/e34fe4d6fcc144df8cda12379c9efcd3) is 4.9M, max 39.3M, 34.4M free. May 17 00:21:56.980819 systemd-modules-load[184]: Inserted module 'overlay' May 17 00:21:57.010235 systemd[1]: Started systemd-journald.service - Journal Service. May 17 00:21:57.014350 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 17 00:21:57.016466 systemd-modules-load[184]: Inserted module 'br_netfilter' May 17 00:21:57.017608 kernel: Bridge firewalling registered May 17 00:21:57.018539 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 17 00:21:57.020990 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 17 00:21:57.026901 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 17 00:21:57.027507 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 17 00:21:57.047672 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 17 00:21:57.049698 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 17 00:21:57.051879 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 17 00:21:57.056714 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 17 00:21:57.073818 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 17 00:21:57.080383 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 17 00:21:57.081093 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 17 00:21:57.082840 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 17 00:21:57.089515 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 17 00:21:57.102523 dracut-cmdline[215]: dracut-dracut-053 May 17 00:21:57.106720 dracut-cmdline[215]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=6b60288baeea1613a76a6f06a8f0e8edc178eae4857ce00eac42d48e92ed015e May 17 00:21:57.136396 systemd-resolved[219]: Positive Trust Anchors: May 17 00:21:57.136414 systemd-resolved[219]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 17 00:21:57.136451 systemd-resolved[219]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 17 00:21:57.143028 systemd-resolved[219]: Defaulting to hostname 'linux'. May 17 00:21:57.145458 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 17 00:21:57.146789 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 17 00:21:57.206353 kernel: SCSI subsystem initialized May 17 00:21:57.217357 kernel: Loading iSCSI transport class v2.0-870. May 17 00:21:57.231341 kernel: iscsi: registered transport (tcp) May 17 00:21:57.254453 kernel: iscsi: registered transport (qla4xxx) May 17 00:21:57.254559 kernel: QLogic iSCSI HBA Driver May 17 00:21:57.308654 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 17 00:21:57.314725 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 17 00:21:57.342722 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 17 00:21:57.342826 kernel: device-mapper: uevent: version 1.0.3 May 17 00:21:57.342842 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com May 17 00:21:57.390384 kernel: raid6: avx2x4 gen() 19952 MB/s May 17 00:21:57.407359 kernel: raid6: avx2x2 gen() 23090 MB/s May 17 00:21:57.424405 kernel: raid6: avx2x1 gen() 19630 MB/s May 17 00:21:57.424517 kernel: raid6: using algorithm avx2x2 gen() 23090 MB/s May 17 00:21:57.442618 kernel: raid6: .... xor() 17797 MB/s, rmw enabled May 17 00:21:57.442749 kernel: raid6: using avx2x2 recovery algorithm May 17 00:21:57.466374 kernel: xor: automatically using best checksumming function avx May 17 00:21:57.631356 kernel: Btrfs loaded, zoned=no, fsverity=no May 17 00:21:57.647024 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 17 00:21:57.653627 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 17 00:21:57.680606 systemd-udevd[402]: Using default interface naming scheme 'v255'. May 17 00:21:57.686509 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 17 00:21:57.697637 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 17 00:21:57.712870 dracut-pre-trigger[407]: rd.md=0: removing MD RAID activation May 17 00:21:57.752191 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 17 00:21:57.757644 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 17 00:21:57.838510 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 17 00:21:57.849510 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 17 00:21:57.880637 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 17 00:21:57.883158 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 17 00:21:57.884899 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 17 00:21:57.885833 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 17 00:21:57.895131 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 17 00:21:57.916647 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 17 00:21:57.936362 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues May 17 00:21:57.941356 kernel: scsi host0: Virtio SCSI HBA May 17 00:21:57.941624 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) May 17 00:21:57.953610 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 17 00:21:57.953722 kernel: GPT:9289727 != 125829119 May 17 00:21:57.953744 kernel: GPT:Alternate GPT header not at the end of the disk. May 17 00:21:57.955061 kernel: GPT:9289727 != 125829119 May 17 00:21:57.956707 kernel: GPT: Use GNU Parted to correct GPT errors. May 17 00:21:57.956758 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 17 00:21:57.975355 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues May 17 00:21:57.977662 kernel: virtio_blk virtio5: [vdb] 980 512-byte logical blocks (502 kB/490 KiB) May 17 00:21:57.987346 kernel: cryptd: max_cpu_qlen set to 1000 May 17 00:21:58.001417 kernel: ACPI: bus type USB registered May 17 00:21:58.008667 kernel: usbcore: registered new interface driver usbfs May 17 00:21:58.008743 kernel: usbcore: registered new interface driver hub May 17 00:21:58.021339 kernel: usbcore: registered new device driver usb May 17 00:21:58.034342 kernel: AVX2 version of gcm_enc/dec engaged. May 17 00:21:58.034458 kernel: AES CTR mode by8 optimization enabled May 17 00:21:58.053220 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 17 00:21:58.053415 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 17 00:21:58.055130 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 17 00:21:58.056392 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 17 00:21:58.056548 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 17 00:21:58.057301 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 17 00:21:58.064649 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 17 00:21:58.088349 kernel: libata version 3.00 loaded. May 17 00:21:58.091756 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (451) May 17 00:21:58.091852 kernel: ata_piix 0000:00:01.1: version 2.13 May 17 00:21:58.093330 kernel: scsi host1: ata_piix May 17 00:21:58.096487 kernel: scsi host2: ata_piix May 17 00:21:58.096770 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 May 17 00:21:58.096787 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 May 17 00:21:58.102438 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 17 00:21:58.111388 kernel: BTRFS: device fsid 7f88d479-6686-439c-8052-b96f0a9d77bc devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (446) May 17 00:21:58.113933 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 17 00:21:58.150126 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 17 00:21:58.156700 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 17 00:21:58.161005 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 17 00:21:58.161569 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 17 00:21:58.167718 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 17 00:21:58.170528 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 17 00:21:58.177744 disk-uuid[537]: Primary Header is updated. May 17 00:21:58.177744 disk-uuid[537]: Secondary Entries is updated. May 17 00:21:58.177744 disk-uuid[537]: Secondary Header is updated. May 17 00:21:58.190332 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 17 00:21:58.204166 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 17 00:21:58.211387 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 17 00:21:58.297385 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller May 17 00:21:58.297671 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 May 17 00:21:58.297799 kernel: uhci_hcd 0000:00:01.2: detected 2 ports May 17 00:21:58.298873 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 May 17 00:21:58.300738 kernel: hub 1-0:1.0: USB hub found May 17 00:21:58.301006 kernel: hub 1-0:1.0: 2 ports detected May 17 00:21:59.201354 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 17 00:21:59.201659 disk-uuid[538]: The operation has completed successfully. May 17 00:21:59.256045 systemd[1]: disk-uuid.service: Deactivated successfully. May 17 00:21:59.256217 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 17 00:21:59.261659 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 17 00:21:59.270036 sh[557]: Success May 17 00:21:59.287333 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" May 17 00:21:59.352864 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 17 00:21:59.367561 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 17 00:21:59.370952 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 17 00:21:59.407880 kernel: BTRFS info (device dm-0): first mount of filesystem 7f88d479-6686-439c-8052-b96f0a9d77bc May 17 00:21:59.407978 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 17 00:21:59.407992 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead May 17 00:21:59.409556 kernel: BTRFS info (device dm-0): disabling log replay at mount time May 17 00:21:59.410454 kernel: BTRFS info (device dm-0): using free space tree May 17 00:21:59.421719 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 17 00:21:59.423269 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 17 00:21:59.430628 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 17 00:21:59.434853 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 17 00:21:59.447342 kernel: BTRFS info (device vda6): first mount of filesystem a013fe34-315a-4c90-9ca1-aace1df6c4ac May 17 00:21:59.447527 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 17 00:21:59.447556 kernel: BTRFS info (device vda6): using free space tree May 17 00:21:59.454345 kernel: BTRFS info (device vda6): auto enabling async discard May 17 00:21:59.469656 systemd[1]: mnt-oem.mount: Deactivated successfully. May 17 00:21:59.470343 kernel: BTRFS info (device vda6): last unmount of filesystem a013fe34-315a-4c90-9ca1-aace1df6c4ac May 17 00:21:59.478640 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 17 00:21:59.486779 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 17 00:21:59.622719 ignition[646]: Ignition 2.19.0 May 17 00:21:59.623639 ignition[646]: Stage: fetch-offline May 17 00:21:59.623701 ignition[646]: no configs at "/usr/lib/ignition/base.d" May 17 00:21:59.624698 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 17 00:21:59.623712 ignition[646]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 17 00:21:59.623828 ignition[646]: parsed url from cmdline: "" May 17 00:21:59.623832 ignition[646]: no config URL provided May 17 00:21:59.623838 ignition[646]: reading system config file "/usr/lib/ignition/user.ign" May 17 00:21:59.623846 ignition[646]: no config at "/usr/lib/ignition/user.ign" May 17 00:21:59.623852 ignition[646]: failed to fetch config: resource requires networking May 17 00:21:59.627759 ignition[646]: Ignition finished successfully May 17 00:21:59.634819 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 17 00:21:59.637504 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 17 00:21:59.678507 systemd-networkd[747]: lo: Link UP May 17 00:21:59.678517 systemd-networkd[747]: lo: Gained carrier May 17 00:21:59.681301 systemd-networkd[747]: Enumeration completed May 17 00:21:59.681864 systemd-networkd[747]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 17 00:21:59.681870 systemd-networkd[747]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. May 17 00:21:59.681885 systemd[1]: Started systemd-networkd.service - Network Configuration. May 17 00:21:59.682404 systemd[1]: Reached target network.target - Network. May 17 00:21:59.683146 systemd-networkd[747]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 17 00:21:59.683150 systemd-networkd[747]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. May 17 00:21:59.684258 systemd-networkd[747]: eth0: Link UP May 17 00:21:59.684262 systemd-networkd[747]: eth0: Gained carrier May 17 00:21:59.684274 systemd-networkd[747]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. May 17 00:21:59.688181 systemd-networkd[747]: eth1: Link UP May 17 00:21:59.688185 systemd-networkd[747]: eth1: Gained carrier May 17 00:21:59.688199 systemd-networkd[747]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 17 00:21:59.690173 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 17 00:21:59.706433 systemd-networkd[747]: eth0: DHCPv4 address 64.23.167.227/20, gateway 64.23.160.1 acquired from 169.254.169.253 May 17 00:21:59.712462 systemd-networkd[747]: eth1: DHCPv4 address 10.124.0.29/20 acquired from 169.254.169.253 May 17 00:21:59.725386 ignition[750]: Ignition 2.19.0 May 17 00:21:59.725398 ignition[750]: Stage: fetch May 17 00:21:59.725605 ignition[750]: no configs at "/usr/lib/ignition/base.d" May 17 00:21:59.725616 ignition[750]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 17 00:21:59.725760 ignition[750]: parsed url from cmdline: "" May 17 00:21:59.725765 ignition[750]: no config URL provided May 17 00:21:59.725773 ignition[750]: reading system config file "/usr/lib/ignition/user.ign" May 17 00:21:59.725787 ignition[750]: no config at "/usr/lib/ignition/user.ign" May 17 00:21:59.725816 ignition[750]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 May 17 00:21:59.758134 ignition[750]: GET result: OK May 17 00:21:59.758894 ignition[750]: parsing config with SHA512: 5c3091e150535a5bb4925c51d77841236242c9840248064a246d8f42d0d991e14f7cb4a1ae2d31357430bf6b769b6a3a3101233f7dd3e2d33cdf98398d64f30a May 17 00:21:59.764802 unknown[750]: fetched base config from "system" May 17 00:21:59.764817 unknown[750]: fetched base config from "system" May 17 00:21:59.764825 unknown[750]: fetched user config from "digitalocean" May 17 00:21:59.766075 ignition[750]: fetch: fetch complete May 17 00:21:59.766083 ignition[750]: fetch: fetch passed May 17 00:21:59.766176 ignition[750]: Ignition finished successfully May 17 00:21:59.769850 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 17 00:21:59.775646 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 17 00:21:59.807967 ignition[758]: Ignition 2.19.0 May 17 00:21:59.807978 ignition[758]: Stage: kargs May 17 00:21:59.808199 ignition[758]: no configs at "/usr/lib/ignition/base.d" May 17 00:21:59.808210 ignition[758]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 17 00:21:59.809328 ignition[758]: kargs: kargs passed May 17 00:21:59.809409 ignition[758]: Ignition finished successfully May 17 00:21:59.811645 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 17 00:21:59.815700 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 17 00:21:59.845848 ignition[764]: Ignition 2.19.0 May 17 00:21:59.845863 ignition[764]: Stage: disks May 17 00:21:59.846165 ignition[764]: no configs at "/usr/lib/ignition/base.d" May 17 00:21:59.846183 ignition[764]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 17 00:21:59.847551 ignition[764]: disks: disks passed May 17 00:21:59.847641 ignition[764]: Ignition finished successfully May 17 00:21:59.849004 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 17 00:21:59.852865 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 17 00:21:59.853286 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 17 00:21:59.854233 systemd[1]: Reached target local-fs.target - Local File Systems. May 17 00:21:59.854912 systemd[1]: Reached target sysinit.target - System Initialization. May 17 00:21:59.855618 systemd[1]: Reached target basic.target - Basic System. May 17 00:21:59.862617 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 17 00:21:59.895071 systemd-fsck[773]: ROOT: clean, 14/553520 files, 52654/553472 blocks May 17 00:21:59.897669 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 17 00:21:59.903932 systemd[1]: Mounting sysroot.mount - /sysroot... May 17 00:22:00.041343 kernel: EXT4-fs (vda9): mounted filesystem 278698a4-82b6-49b4-b6df-f7999ed4e35e r/w with ordered data mode. Quota mode: none. May 17 00:22:00.042835 systemd[1]: Mounted sysroot.mount - /sysroot. May 17 00:22:00.044170 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 17 00:22:00.057590 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 17 00:22:00.061158 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 17 00:22:00.063496 systemd[1]: Starting flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent... May 17 00:22:00.072392 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (781) May 17 00:22:00.073629 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... May 17 00:22:00.079547 kernel: BTRFS info (device vda6): first mount of filesystem a013fe34-315a-4c90-9ca1-aace1df6c4ac May 17 00:22:00.079601 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 17 00:22:00.079616 kernel: BTRFS info (device vda6): using free space tree May 17 00:22:00.080509 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 17 00:22:00.085084 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 17 00:22:00.089968 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 17 00:22:00.103612 kernel: BTRFS info (device vda6): auto enabling async discard May 17 00:22:00.099914 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 17 00:22:00.136995 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 17 00:22:00.163824 initrd-setup-root[812]: cut: /sysroot/etc/passwd: No such file or directory May 17 00:22:00.188378 initrd-setup-root[819]: cut: /sysroot/etc/group: No such file or directory May 17 00:22:00.194254 coreos-metadata[783]: May 17 00:22:00.193 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 17 00:22:00.196338 coreos-metadata[784]: May 17 00:22:00.195 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 17 00:22:00.199560 initrd-setup-root[826]: cut: /sysroot/etc/shadow: No such file or directory May 17 00:22:00.207230 initrd-setup-root[833]: cut: /sysroot/etc/gshadow: No such file or directory May 17 00:22:00.209144 coreos-metadata[784]: May 17 00:22:00.209 INFO Fetch successful May 17 00:22:00.210880 coreos-metadata[783]: May 17 00:22:00.209 INFO Fetch successful May 17 00:22:00.218478 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. May 17 00:22:00.218696 systemd[1]: Finished flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent. May 17 00:22:00.223224 coreos-metadata[784]: May 17 00:22:00.223 INFO wrote hostname ci-4081.3.3-n-0e116cd3eb to /sysroot/etc/hostname May 17 00:22:00.225525 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 17 00:22:00.357370 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 17 00:22:00.367598 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 17 00:22:00.371705 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 17 00:22:00.384367 kernel: BTRFS info (device vda6): last unmount of filesystem a013fe34-315a-4c90-9ca1-aace1df6c4ac May 17 00:22:00.405695 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 17 00:22:00.422623 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 17 00:22:00.426981 ignition[902]: INFO : Ignition 2.19.0 May 17 00:22:00.426981 ignition[902]: INFO : Stage: mount May 17 00:22:00.428481 ignition[902]: INFO : no configs at "/usr/lib/ignition/base.d" May 17 00:22:00.428481 ignition[902]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 17 00:22:00.429733 ignition[902]: INFO : mount: mount passed May 17 00:22:00.429733 ignition[902]: INFO : Ignition finished successfully May 17 00:22:00.431408 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 17 00:22:00.438601 systemd[1]: Starting ignition-files.service - Ignition (files)... May 17 00:22:00.473861 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 17 00:22:00.489364 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (915) May 17 00:22:00.492706 kernel: BTRFS info (device vda6): first mount of filesystem a013fe34-315a-4c90-9ca1-aace1df6c4ac May 17 00:22:00.492818 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 17 00:22:00.492834 kernel: BTRFS info (device vda6): using free space tree May 17 00:22:00.498357 kernel: BTRFS info (device vda6): auto enabling async discard May 17 00:22:00.501896 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 17 00:22:00.545041 ignition[932]: INFO : Ignition 2.19.0 May 17 00:22:00.545041 ignition[932]: INFO : Stage: files May 17 00:22:00.545041 ignition[932]: INFO : no configs at "/usr/lib/ignition/base.d" May 17 00:22:00.545041 ignition[932]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 17 00:22:00.547999 ignition[932]: DEBUG : files: compiled without relabeling support, skipping May 17 00:22:00.548538 ignition[932]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 17 00:22:00.548538 ignition[932]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 17 00:22:00.553651 ignition[932]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 17 00:22:00.554818 ignition[932]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 17 00:22:00.556301 unknown[932]: wrote ssh authorized keys file for user: core May 17 00:22:00.557422 ignition[932]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 17 00:22:00.558726 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" May 17 00:22:00.559784 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" May 17 00:22:00.559784 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 17 00:22:00.559784 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 May 17 00:22:00.603230 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 17 00:22:00.691058 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 17 00:22:00.691058 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 17 00:22:00.693112 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 May 17 00:22:00.983806 systemd-networkd[747]: eth1: Gained IPv6LL May 17 00:22:01.112109 systemd-networkd[747]: eth0: Gained IPv6LL May 17 00:22:01.172017 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK May 17 00:22:01.263769 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 17 00:22:01.263769 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" May 17 00:22:01.265586 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" May 17 00:22:01.265586 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" May 17 00:22:01.265586 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" May 17 00:22:01.265586 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 17 00:22:01.265586 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 17 00:22:01.265586 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 17 00:22:01.265586 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 17 00:22:01.274485 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" May 17 00:22:01.274485 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 17 00:22:01.274485 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:22:01.274485 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:22:01.274485 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:22:01.274485 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 May 17 00:22:02.018442 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK May 17 00:22:02.403989 ignition[932]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:22:02.403989 ignition[932]: INFO : files: op(d): [started] processing unit "containerd.service" May 17 00:22:02.405592 ignition[932]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 17 00:22:02.405592 ignition[932]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 17 00:22:02.405592 ignition[932]: INFO : files: op(d): [finished] processing unit "containerd.service" May 17 00:22:02.405592 ignition[932]: INFO : files: op(f): [started] processing unit "prepare-helm.service" May 17 00:22:02.405592 ignition[932]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 17 00:22:02.410830 ignition[932]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 17 00:22:02.410830 ignition[932]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" May 17 00:22:02.410830 ignition[932]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" May 17 00:22:02.410830 ignition[932]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" May 17 00:22:02.410830 ignition[932]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" May 17 00:22:02.410830 ignition[932]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" May 17 00:22:02.410830 ignition[932]: INFO : files: files passed May 17 00:22:02.410830 ignition[932]: INFO : Ignition finished successfully May 17 00:22:02.408258 systemd[1]: Finished ignition-files.service - Ignition (files). May 17 00:22:02.417762 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 17 00:22:02.423907 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 17 00:22:02.428078 systemd[1]: ignition-quench.service: Deactivated successfully. May 17 00:22:02.428214 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 17 00:22:02.449926 initrd-setup-root-after-ignition[960]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 17 00:22:02.449926 initrd-setup-root-after-ignition[960]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 17 00:22:02.453222 initrd-setup-root-after-ignition[964]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 17 00:22:02.456476 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 17 00:22:02.458053 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 17 00:22:02.473412 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 17 00:22:02.525531 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 17 00:22:02.525717 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 17 00:22:02.526990 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 17 00:22:02.527585 systemd[1]: Reached target initrd.target - Initrd Default Target. May 17 00:22:02.528544 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 17 00:22:02.530164 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 17 00:22:02.563717 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 17 00:22:02.571847 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 17 00:22:02.597745 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 17 00:22:02.599120 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 17 00:22:02.600376 systemd[1]: Stopped target timers.target - Timer Units. May 17 00:22:02.601054 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 17 00:22:02.601254 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 17 00:22:02.602444 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 17 00:22:02.602864 systemd[1]: Stopped target basic.target - Basic System. May 17 00:22:02.604026 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 17 00:22:02.604900 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 17 00:22:02.605834 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 17 00:22:02.606854 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 17 00:22:02.607841 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 17 00:22:02.608780 systemd[1]: Stopped target sysinit.target - System Initialization. May 17 00:22:02.609772 systemd[1]: Stopped target local-fs.target - Local File Systems. May 17 00:22:02.610702 systemd[1]: Stopped target swap.target - Swaps. May 17 00:22:02.611626 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 17 00:22:02.611842 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 17 00:22:02.612890 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 17 00:22:02.613837 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 17 00:22:02.614776 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 17 00:22:02.614967 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 17 00:22:02.615727 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 17 00:22:02.615933 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 17 00:22:02.617154 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 17 00:22:02.617377 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 17 00:22:02.618417 systemd[1]: ignition-files.service: Deactivated successfully. May 17 00:22:02.618581 systemd[1]: Stopped ignition-files.service - Ignition (files). May 17 00:22:02.619217 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. May 17 00:22:02.619394 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 17 00:22:02.626903 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 17 00:22:02.628349 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 17 00:22:02.629277 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 17 00:22:02.640782 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 17 00:22:02.641680 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 17 00:22:02.641911 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 17 00:22:02.646034 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 17 00:22:02.646228 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 17 00:22:02.662646 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 17 00:22:02.662826 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 17 00:22:02.666074 ignition[984]: INFO : Ignition 2.19.0 May 17 00:22:02.666074 ignition[984]: INFO : Stage: umount May 17 00:22:02.666074 ignition[984]: INFO : no configs at "/usr/lib/ignition/base.d" May 17 00:22:02.666074 ignition[984]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" May 17 00:22:02.671262 ignition[984]: INFO : umount: umount passed May 17 00:22:02.671262 ignition[984]: INFO : Ignition finished successfully May 17 00:22:02.669937 systemd[1]: ignition-mount.service: Deactivated successfully. May 17 00:22:02.670079 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 17 00:22:02.671806 systemd[1]: ignition-disks.service: Deactivated successfully. May 17 00:22:02.671910 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 17 00:22:02.672999 systemd[1]: ignition-kargs.service: Deactivated successfully. May 17 00:22:02.673069 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 17 00:22:02.673657 systemd[1]: ignition-fetch.service: Deactivated successfully. May 17 00:22:02.673721 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 17 00:22:02.676317 systemd[1]: Stopped target network.target - Network. May 17 00:22:02.676818 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 17 00:22:02.676906 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 17 00:22:02.678035 systemd[1]: Stopped target paths.target - Path Units. May 17 00:22:02.679935 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 17 00:22:02.680178 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 17 00:22:02.680830 systemd[1]: Stopped target slices.target - Slice Units. May 17 00:22:02.683900 systemd[1]: Stopped target sockets.target - Socket Units. May 17 00:22:02.684669 systemd[1]: iscsid.socket: Deactivated successfully. May 17 00:22:02.684752 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 17 00:22:02.685501 systemd[1]: iscsiuio.socket: Deactivated successfully. May 17 00:22:02.685556 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 17 00:22:02.689142 systemd[1]: ignition-setup.service: Deactivated successfully. May 17 00:22:02.689264 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 17 00:22:02.689884 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 17 00:22:02.689978 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 17 00:22:02.692168 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 17 00:22:02.694061 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 17 00:22:02.706726 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 17 00:22:02.712612 systemd-networkd[747]: eth0: DHCPv6 lease lost May 17 00:22:02.733480 systemd-networkd[747]: eth1: DHCPv6 lease lost May 17 00:22:02.735930 systemd[1]: systemd-resolved.service: Deactivated successfully. May 17 00:22:02.736104 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 17 00:22:02.776833 systemd[1]: systemd-networkd.service: Deactivated successfully. May 17 00:22:02.777118 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 17 00:22:02.781233 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 17 00:22:02.782557 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 17 00:22:02.798735 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 17 00:22:02.799394 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 17 00:22:02.799607 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 17 00:22:02.803744 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 17 00:22:02.803861 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 17 00:22:02.805277 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 17 00:22:02.805446 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 17 00:22:02.806157 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 17 00:22:02.806224 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 17 00:22:02.807232 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 17 00:22:02.810815 systemd[1]: sysroot-boot.service: Deactivated successfully. May 17 00:22:02.810950 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 17 00:22:02.822059 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 17 00:22:02.822294 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 17 00:22:02.828778 systemd[1]: network-cleanup.service: Deactivated successfully. May 17 00:22:02.828992 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 17 00:22:02.830510 systemd[1]: systemd-udevd.service: Deactivated successfully. May 17 00:22:02.830756 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 17 00:22:02.833761 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 17 00:22:02.833872 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 17 00:22:02.834888 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 17 00:22:02.834951 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 17 00:22:02.835879 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 17 00:22:02.835974 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 17 00:22:02.837149 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 17 00:22:02.837238 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 17 00:22:02.838520 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 17 00:22:02.838601 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 17 00:22:02.845693 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 17 00:22:02.846225 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 17 00:22:02.846379 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 17 00:22:02.846890 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 17 00:22:02.846947 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 17 00:22:02.869049 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 17 00:22:02.869279 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 17 00:22:02.870831 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 17 00:22:02.875841 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 17 00:22:02.906384 systemd[1]: Switching root. May 17 00:22:02.985662 systemd-journald[183]: Received SIGTERM from PID 1 (systemd). May 17 00:22:02.985845 systemd-journald[183]: Journal stopped May 17 00:22:04.341946 kernel: SELinux: policy capability network_peer_controls=1 May 17 00:22:04.342110 kernel: SELinux: policy capability open_perms=1 May 17 00:22:04.342134 kernel: SELinux: policy capability extended_socket_class=1 May 17 00:22:04.342154 kernel: SELinux: policy capability always_check_network=0 May 17 00:22:04.342202 kernel: SELinux: policy capability cgroup_seclabel=1 May 17 00:22:04.342226 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 17 00:22:04.342245 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 17 00:22:04.342273 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 17 00:22:04.342293 kernel: audit: type=1403 audit(1747441323.188:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 17 00:22:04.342338 systemd[1]: Successfully loaded SELinux policy in 46.388ms. May 17 00:22:04.342376 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 13.840ms. May 17 00:22:04.342398 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) May 17 00:22:04.342419 systemd[1]: Detected virtualization kvm. May 17 00:22:04.342439 systemd[1]: Detected architecture x86-64. May 17 00:22:04.342459 systemd[1]: Detected first boot. May 17 00:22:04.342477 systemd[1]: Hostname set to . May 17 00:22:04.342496 systemd[1]: Initializing machine ID from VM UUID. May 17 00:22:04.342522 zram_generator::config[1044]: No configuration found. May 17 00:22:04.342546 systemd[1]: Populated /etc with preset unit settings. May 17 00:22:04.342567 systemd[1]: Queued start job for default target multi-user.target. May 17 00:22:04.342587 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 17 00:22:04.342608 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 17 00:22:04.342628 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 17 00:22:04.342647 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 17 00:22:04.342666 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 17 00:22:04.342685 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 17 00:22:04.342717 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 17 00:22:04.342748 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 17 00:22:04.342768 systemd[1]: Created slice user.slice - User and Session Slice. May 17 00:22:04.342790 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 17 00:22:04.342808 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 17 00:22:04.342827 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 17 00:22:04.342847 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 17 00:22:04.342867 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 17 00:22:04.342893 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 17 00:22:04.342913 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 17 00:22:04.342932 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 17 00:22:04.342952 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 17 00:22:04.342971 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 17 00:22:04.343013 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 17 00:22:04.343034 systemd[1]: Reached target slices.target - Slice Units. May 17 00:22:04.343062 systemd[1]: Reached target swap.target - Swaps. May 17 00:22:04.343084 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 17 00:22:04.343103 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 17 00:22:04.343124 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 17 00:22:04.343143 systemd[1]: Listening on systemd-journald.socket - Journal Socket. May 17 00:22:04.343163 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 17 00:22:04.343188 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 17 00:22:04.343209 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 17 00:22:04.343229 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 17 00:22:04.343252 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 17 00:22:04.343279 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 17 00:22:04.343302 systemd[1]: Mounting media.mount - External Media Directory... May 17 00:22:04.362226 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:04.362262 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 17 00:22:04.362282 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 17 00:22:04.362303 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 17 00:22:04.362351 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 17 00:22:04.362371 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 17 00:22:04.362397 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 17 00:22:04.362416 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 17 00:22:04.362434 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 17 00:22:04.362452 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 17 00:22:04.362471 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 17 00:22:04.362489 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 17 00:22:04.362517 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 17 00:22:04.362535 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 17 00:22:04.362559 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. May 17 00:22:04.362580 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) May 17 00:22:04.362598 systemd[1]: Starting systemd-journald.service - Journal Service... May 17 00:22:04.362617 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 17 00:22:04.362635 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 17 00:22:04.362654 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 17 00:22:04.362672 kernel: fuse: init (API version 7.39) May 17 00:22:04.362699 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 17 00:22:04.362718 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:04.362742 kernel: loop: module loaded May 17 00:22:04.362760 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 17 00:22:04.362829 systemd-journald[1131]: Collecting audit messages is disabled. May 17 00:22:04.362872 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 17 00:22:04.362891 systemd-journald[1131]: Journal started May 17 00:22:04.362925 systemd-journald[1131]: Runtime Journal (/run/log/journal/e34fe4d6fcc144df8cda12379c9efcd3) is 4.9M, max 39.3M, 34.4M free. May 17 00:22:04.376195 kernel: ACPI: bus type drm_connector registered May 17 00:22:04.380350 systemd[1]: Mounted media.mount - External Media Directory. May 17 00:22:04.380469 systemd[1]: Started systemd-journald.service - Journal Service. May 17 00:22:04.383290 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 17 00:22:04.386685 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 17 00:22:04.388789 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 17 00:22:04.389932 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 17 00:22:04.391269 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 17 00:22:04.391876 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 17 00:22:04.392997 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:22:04.393635 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 17 00:22:04.394814 systemd[1]: modprobe@drm.service: Deactivated successfully. May 17 00:22:04.395201 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 17 00:22:04.396886 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:22:04.397152 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 17 00:22:04.398785 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 17 00:22:04.399718 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 17 00:22:04.401356 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:22:04.402369 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 17 00:22:04.403908 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 17 00:22:04.407161 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 17 00:22:04.409144 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 17 00:22:04.434435 systemd[1]: Reached target network-pre.target - Preparation for Network. May 17 00:22:04.442575 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 17 00:22:04.454528 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 17 00:22:04.455409 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 17 00:22:04.471703 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 17 00:22:04.484694 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 17 00:22:04.486728 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:22:04.498673 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 17 00:22:04.499762 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 17 00:22:04.511629 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 17 00:22:04.523835 systemd-journald[1131]: Time spent on flushing to /var/log/journal/e34fe4d6fcc144df8cda12379c9efcd3 is 41.333ms for 973 entries. May 17 00:22:04.523835 systemd-journald[1131]: System Journal (/var/log/journal/e34fe4d6fcc144df8cda12379c9efcd3) is 8.0M, max 195.6M, 187.6M free. May 17 00:22:04.575957 systemd-journald[1131]: Received client request to flush runtime journal. May 17 00:22:04.525122 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 17 00:22:04.543887 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 17 00:22:04.544913 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 17 00:22:04.545685 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 17 00:22:04.546693 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 17 00:22:04.555681 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 17 00:22:04.585028 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 17 00:22:04.631982 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 17 00:22:04.654527 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 17 00:22:04.656730 systemd-tmpfiles[1185]: ACLs are not supported, ignoring. May 17 00:22:04.656754 systemd-tmpfiles[1185]: ACLs are not supported, ignoring. May 17 00:22:04.667781 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... May 17 00:22:04.679902 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 17 00:22:04.692688 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 17 00:22:04.711150 udevadm[1201]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. May 17 00:22:04.748991 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 17 00:22:04.761727 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 17 00:22:04.799055 systemd-tmpfiles[1208]: ACLs are not supported, ignoring. May 17 00:22:04.799693 systemd-tmpfiles[1208]: ACLs are not supported, ignoring. May 17 00:22:04.812886 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 17 00:22:05.481194 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 17 00:22:05.489686 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 17 00:22:05.529351 systemd-udevd[1214]: Using default interface naming scheme 'v255'. May 17 00:22:05.561450 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 17 00:22:05.571579 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 17 00:22:05.601577 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 17 00:22:05.647337 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1226) May 17 00:22:05.700842 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 17 00:22:05.702066 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:05.702242 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 17 00:22:05.709543 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 17 00:22:05.717619 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 17 00:22:05.721602 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 17 00:22:05.722300 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 17 00:22:05.722394 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 17 00:22:05.722444 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:05.723259 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. May 17 00:22:05.755014 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:22:05.755230 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 17 00:22:05.769921 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:22:05.770207 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 17 00:22:05.770973 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:22:05.775971 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:22:05.778428 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 17 00:22:05.786781 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 17 00:22:05.852244 systemd-networkd[1219]: lo: Link UP May 17 00:22:05.852626 systemd-networkd[1219]: lo: Gained carrier May 17 00:22:05.856197 systemd-networkd[1219]: Enumeration completed May 17 00:22:05.856432 systemd[1]: Started systemd-networkd.service - Network Configuration. May 17 00:22:05.856590 systemd-networkd[1219]: eth0: Configuring with /run/systemd/network/10-da:57:52:2d:7f:8c.network. May 17 00:22:05.857533 systemd-networkd[1219]: eth1: Configuring with /run/systemd/network/10-96:24:43:bb:88:ea.network. May 17 00:22:05.858199 systemd-networkd[1219]: eth0: Link UP May 17 00:22:05.858287 systemd-networkd[1219]: eth0: Gained carrier May 17 00:22:05.863522 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 17 00:22:05.865995 systemd-networkd[1219]: eth1: Link UP May 17 00:22:05.866005 systemd-networkd[1219]: eth1: Gained carrier May 17 00:22:05.886428 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 May 17 00:22:05.897338 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 May 17 00:22:05.897644 kernel: ACPI: button: Power Button [PWRF] May 17 00:22:05.910529 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 17 00:22:05.927345 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 May 17 00:22:05.983198 kernel: mousedev: PS/2 mouse device common for all mice May 17 00:22:06.003864 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 17 00:22:06.081363 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 May 17 00:22:06.081490 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console May 17 00:22:06.088279 kernel: Console: switching to colour dummy device 80x25 May 17 00:22:06.088380 kernel: [drm] features: -virgl +edid -resource_blob -host_visible May 17 00:22:06.088396 kernel: [drm] features: -context_init May 17 00:22:06.088410 kernel: [drm] number of scanouts: 1 May 17 00:22:06.091345 kernel: [drm] number of cap sets: 0 May 17 00:22:06.096420 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 May 17 00:22:06.108262 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device May 17 00:22:06.108352 kernel: Console: switching to colour frame buffer device 128x48 May 17 00:22:06.120077 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device May 17 00:22:06.121646 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 17 00:22:06.121954 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 17 00:22:06.141750 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 17 00:22:06.152584 kernel: EDAC MC: Ver: 3.0.0 May 17 00:22:06.158237 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 17 00:22:06.159377 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 17 00:22:06.169698 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 17 00:22:06.182925 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. May 17 00:22:06.191638 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... May 17 00:22:06.215360 lvm[1274]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 17 00:22:06.236059 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 17 00:22:06.250500 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. May 17 00:22:06.252182 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 17 00:22:06.262820 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... May 17 00:22:06.269156 lvm[1282]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 17 00:22:06.301048 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. May 17 00:22:06.302592 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 17 00:22:06.311705 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... May 17 00:22:06.314848 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 17 00:22:06.314911 systemd[1]: Reached target machines.target - Containers. May 17 00:22:06.317539 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 17 00:22:06.335981 kernel: ISO 9660 Extensions: RRIP_1991A May 17 00:22:06.333512 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. May 17 00:22:06.334585 systemd[1]: Reached target local-fs.target - Local File Systems. May 17 00:22:06.338302 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). May 17 00:22:06.345731 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 17 00:22:06.358578 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 17 00:22:06.361066 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 17 00:22:06.364070 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... May 17 00:22:06.377615 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 17 00:22:06.381908 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 17 00:22:06.390326 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 17 00:22:06.410639 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 17 00:22:06.413895 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. May 17 00:22:06.427195 kernel: loop0: detected capacity change from 0 to 221472 May 17 00:22:06.451419 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 17 00:22:06.488361 kernel: loop1: detected capacity change from 0 to 140768 May 17 00:22:06.532682 kernel: loop2: detected capacity change from 0 to 142488 May 17 00:22:06.583361 kernel: loop3: detected capacity change from 0 to 8 May 17 00:22:06.612955 kernel: loop4: detected capacity change from 0 to 221472 May 17 00:22:06.631418 kernel: loop5: detected capacity change from 0 to 140768 May 17 00:22:06.648777 kernel: loop6: detected capacity change from 0 to 142488 May 17 00:22:06.664427 kernel: loop7: detected capacity change from 0 to 8 May 17 00:22:06.665730 (sd-merge)[1307]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. May 17 00:22:06.667903 (sd-merge)[1307]: Merged extensions into '/usr'. May 17 00:22:06.682429 systemd[1]: Reloading requested from client PID 1296 ('systemd-sysext') (unit systemd-sysext.service)... May 17 00:22:06.682449 systemd[1]: Reloading... May 17 00:22:06.811325 zram_generator::config[1332]: No configuration found. May 17 00:22:06.911255 ldconfig[1294]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 17 00:22:06.964695 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:22:07.042124 systemd[1]: Reloading finished in 359 ms. May 17 00:22:07.057260 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 17 00:22:07.060972 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 17 00:22:07.072754 systemd[1]: Starting ensure-sysext.service... May 17 00:22:07.079662 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 17 00:22:07.089610 systemd[1]: Reloading requested from client PID 1385 ('systemctl') (unit ensure-sysext.service)... May 17 00:22:07.089632 systemd[1]: Reloading... May 17 00:22:07.132457 systemd-tmpfiles[1386]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 17 00:22:07.132818 systemd-tmpfiles[1386]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 17 00:22:07.133794 systemd-tmpfiles[1386]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 17 00:22:07.134107 systemd-tmpfiles[1386]: ACLs are not supported, ignoring. May 17 00:22:07.134171 systemd-tmpfiles[1386]: ACLs are not supported, ignoring. May 17 00:22:07.138814 systemd-tmpfiles[1386]: Detected autofs mount point /boot during canonicalization of boot. May 17 00:22:07.138827 systemd-tmpfiles[1386]: Skipping /boot May 17 00:22:07.170270 systemd-tmpfiles[1386]: Detected autofs mount point /boot during canonicalization of boot. May 17 00:22:07.172392 systemd-tmpfiles[1386]: Skipping /boot May 17 00:22:07.188351 zram_generator::config[1413]: No configuration found. May 17 00:22:07.365127 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:22:07.437879 systemd[1]: Reloading finished in 347 ms. May 17 00:22:07.458934 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 17 00:22:07.478701 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... May 17 00:22:07.487403 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 17 00:22:07.500639 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 17 00:22:07.507934 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 17 00:22:07.512456 systemd-networkd[1219]: eth0: Gained IPv6LL May 17 00:22:07.512980 systemd-networkd[1219]: eth1: Gained IPv6LL May 17 00:22:07.526573 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 17 00:22:07.529199 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 17 00:22:07.571011 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:07.571494 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 17 00:22:07.579300 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 17 00:22:07.595232 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 17 00:22:07.612860 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 17 00:22:07.613889 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 17 00:22:07.614205 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:07.628006 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 17 00:22:07.653709 augenrules[1491]: No rules May 17 00:22:07.638779 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 17 00:22:07.645621 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. May 17 00:22:07.650551 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:22:07.650835 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 17 00:22:07.660194 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:22:07.661624 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 17 00:22:07.680455 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:22:07.681453 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 17 00:22:07.704097 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 17 00:22:07.713440 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:07.714047 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 17 00:22:07.720686 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 17 00:22:07.732069 systemd-resolved[1473]: Positive Trust Anchors: May 17 00:22:07.732089 systemd-resolved[1473]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 17 00:22:07.732139 systemd-resolved[1473]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 17 00:22:07.736451 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 17 00:22:07.740249 systemd-resolved[1473]: Using system hostname 'ci-4081.3.3-n-0e116cd3eb'. May 17 00:22:07.745678 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 17 00:22:07.757654 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 17 00:22:07.760133 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 17 00:22:07.764544 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 17 00:22:07.767081 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:22:07.767192 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:22:07.767911 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 17 00:22:07.774096 systemd[1]: Finished ensure-sysext.service. May 17 00:22:07.778247 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:22:07.780513 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 17 00:22:07.781612 systemd[1]: modprobe@drm.service: Deactivated successfully. May 17 00:22:07.781816 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 17 00:22:07.789648 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:22:07.789877 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 17 00:22:07.790990 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:22:07.791194 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 17 00:22:07.804861 systemd[1]: Reached target network.target - Network. May 17 00:22:07.809349 systemd[1]: Reached target network-online.target - Network is Online. May 17 00:22:07.809911 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 17 00:22:07.811499 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:22:07.811607 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 17 00:22:07.824709 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 17 00:22:07.828817 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 17 00:22:07.899865 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 17 00:22:07.901532 systemd[1]: Reached target sysinit.target - System Initialization. May 17 00:22:07.902149 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 17 00:22:07.904854 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 17 00:22:07.905479 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 17 00:22:07.906063 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 17 00:22:07.906098 systemd[1]: Reached target paths.target - Path Units. May 17 00:22:07.906970 systemd[1]: Reached target time-set.target - System Time Set. May 17 00:22:07.908390 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 17 00:22:07.909181 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 17 00:22:07.909891 systemd[1]: Reached target timers.target - Timer Units. May 17 00:22:07.911557 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 17 00:22:07.914497 systemd[1]: Starting docker.socket - Docker Socket for the API... May 17 00:22:07.922442 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 17 00:22:07.925489 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 17 00:22:07.926220 systemd[1]: Reached target sockets.target - Socket Units. May 17 00:22:07.928316 systemd[1]: Reached target basic.target - Basic System. May 17 00:22:07.928999 systemd[1]: System is tainted: cgroupsv1 May 17 00:22:07.929055 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 17 00:22:07.929080 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 17 00:22:07.940518 systemd[1]: Starting containerd.service - containerd container runtime... May 17 00:22:07.948598 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 17 00:22:07.957721 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 17 00:22:07.969222 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 17 00:22:07.974763 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 17 00:22:07.975438 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 17 00:22:07.994538 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 17 00:22:08.004242 dbus-daemon[1536]: [system] SELinux support is enabled May 17 00:22:08.017513 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 17 00:22:08.025091 coreos-metadata[1535]: May 17 00:22:08.024 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 17 00:22:08.030615 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 17 00:22:08.043498 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 17 00:22:08.054431 extend-filesystems[1540]: Found loop4 May 17 00:22:08.054431 extend-filesystems[1540]: Found loop5 May 17 00:22:08.054431 extend-filesystems[1540]: Found loop6 May 17 00:22:08.054431 extend-filesystems[1540]: Found loop7 May 17 00:22:08.054431 extend-filesystems[1540]: Found vda May 17 00:22:08.054431 extend-filesystems[1540]: Found vda1 May 17 00:22:08.054431 extend-filesystems[1540]: Found vda2 May 17 00:22:08.054431 extend-filesystems[1540]: Found vda3 May 17 00:22:08.054431 extend-filesystems[1540]: Found usr May 17 00:22:08.054431 extend-filesystems[1540]: Found vda4 May 17 00:22:08.054431 extend-filesystems[1540]: Found vda6 May 17 00:22:08.054431 extend-filesystems[1540]: Found vda7 May 17 00:22:08.054431 extend-filesystems[1540]: Found vda9 May 17 00:22:08.122252 extend-filesystems[1540]: Checking size of /dev/vda9 May 17 00:22:08.055775 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 17 00:22:08.123203 coreos-metadata[1535]: May 17 00:22:08.047 INFO Fetch successful May 17 00:22:08.123250 jq[1539]: false May 17 00:22:08.107157 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 17 00:22:08.126373 extend-filesystems[1540]: Resized partition /dev/vda9 May 17 00:22:08.131929 systemd[1]: Starting systemd-logind.service - User Login Management... May 17 00:22:08.132900 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 17 00:22:08.138847 extend-filesystems[1572]: resize2fs 1.47.1 (20-May-2024) May 17 00:22:08.144516 systemd[1]: Starting update-engine.service - Update Engine... May 17 00:22:08.152579 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks May 17 00:22:08.162479 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 17 00:22:08.165570 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 17 00:22:08.184820 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1226) May 17 00:22:08.189721 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 17 00:22:08.190230 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 17 00:22:08.194962 systemd-timesyncd[1528]: Contacted time server 129.250.35.251:123 (0.flatcar.pool.ntp.org). May 17 00:22:08.195363 systemd-timesyncd[1528]: Initial clock synchronization to Sat 2025-05-17 00:22:08.115335 UTC. May 17 00:22:08.205941 systemd[1]: motdgen.service: Deactivated successfully. May 17 00:22:08.206243 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 17 00:22:08.207150 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 17 00:22:08.231150 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 17 00:22:08.231474 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 17 00:22:08.247510 jq[1574]: true May 17 00:22:08.273747 kernel: EXT4-fs (vda9): resized filesystem to 15121403 May 17 00:22:08.292792 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 17 00:22:08.292841 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 17 00:22:08.298752 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 17 00:22:08.298896 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). May 17 00:22:08.298935 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 17 00:22:08.313846 extend-filesystems[1572]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 17 00:22:08.313846 extend-filesystems[1572]: old_desc_blocks = 1, new_desc_blocks = 8 May 17 00:22:08.313846 extend-filesystems[1572]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. May 17 00:22:08.336232 extend-filesystems[1540]: Resized filesystem in /dev/vda9 May 17 00:22:08.336232 extend-filesystems[1540]: Found vdb May 17 00:22:08.317026 systemd[1]: extend-filesystems.service: Deactivated successfully. May 17 00:22:08.317427 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 17 00:22:08.365025 jq[1586]: true May 17 00:22:08.324114 (ntainerd)[1595]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 17 00:22:08.373999 update_engine[1573]: I20250517 00:22:08.369935 1573 main.cc:92] Flatcar Update Engine starting May 17 00:22:08.398371 update_engine[1573]: I20250517 00:22:08.397300 1573 update_check_scheduler.cc:74] Next update check in 3m9s May 17 00:22:08.401016 systemd[1]: Started update-engine.service - Update Engine. May 17 00:22:08.418589 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 17 00:22:08.426594 tar[1582]: linux-amd64/helm May 17 00:22:08.426207 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 17 00:22:08.427261 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 17 00:22:08.434665 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 17 00:22:08.592657 bash[1623]: Updated "/home/core/.ssh/authorized_keys" May 17 00:22:08.595777 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 17 00:22:08.600539 systemd-logind[1571]: New seat seat0. May 17 00:22:08.604151 systemd-logind[1571]: Watching system buttons on /dev/input/event1 (Power Button) May 17 00:22:08.604190 systemd-logind[1571]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 17 00:22:08.611661 systemd[1]: Starting sshkeys.service... May 17 00:22:08.612815 systemd[1]: Started systemd-logind.service - User Login Management. May 17 00:22:08.675484 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. May 17 00:22:08.685988 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... May 17 00:22:08.884098 coreos-metadata[1627]: May 17 00:22:08.883 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 May 17 00:22:08.902660 coreos-metadata[1627]: May 17 00:22:08.898 INFO Fetch successful May 17 00:22:08.924061 unknown[1627]: wrote ssh authorized keys file for user: core May 17 00:22:08.933871 locksmithd[1609]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 17 00:22:08.980361 update-ssh-keys[1649]: Updated "/home/core/.ssh/authorized_keys" May 17 00:22:08.980719 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). May 17 00:22:08.990234 systemd[1]: Finished sshkeys.service. May 17 00:22:09.006354 containerd[1595]: time="2025-05-17T00:22:09.005163787Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 May 17 00:22:09.057211 containerd[1595]: time="2025-05-17T00:22:09.057111260Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 May 17 00:22:09.063331 containerd[1595]: time="2025-05-17T00:22:09.063225819Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.90-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 May 17 00:22:09.063331 containerd[1595]: time="2025-05-17T00:22:09.063295725Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 May 17 00:22:09.063750 containerd[1595]: time="2025-05-17T00:22:09.063541658Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 May 17 00:22:09.063977 containerd[1595]: time="2025-05-17T00:22:09.063866991Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 May 17 00:22:09.063977 containerd[1595]: time="2025-05-17T00:22:09.063908559Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 May 17 00:22:09.064244 containerd[1595]: time="2025-05-17T00:22:09.064112490Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:22:09.064244 containerd[1595]: time="2025-05-17T00:22:09.064134834Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 May 17 00:22:09.064790 containerd[1595]: time="2025-05-17T00:22:09.064664171Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:22:09.064790 containerd[1595]: time="2025-05-17T00:22:09.064691815Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 May 17 00:22:09.064790 containerd[1595]: time="2025-05-17T00:22:09.064723708Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:22:09.064790 containerd[1595]: time="2025-05-17T00:22:09.064739754Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 May 17 00:22:09.065368 containerd[1595]: time="2025-05-17T00:22:09.065084359Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 May 17 00:22:09.065887 containerd[1595]: time="2025-05-17T00:22:09.065720866Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 May 17 00:22:09.066322 containerd[1595]: time="2025-05-17T00:22:09.066220011Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:22:09.066322 containerd[1595]: time="2025-05-17T00:22:09.066264316Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 May 17 00:22:09.066957 containerd[1595]: time="2025-05-17T00:22:09.066652057Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 May 17 00:22:09.066957 containerd[1595]: time="2025-05-17T00:22:09.066774728Z" level=info msg="metadata content store policy set" policy=shared May 17 00:22:09.069980 sshd_keygen[1581]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 17 00:22:09.079457 containerd[1595]: time="2025-05-17T00:22:09.076405720Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 May 17 00:22:09.079457 containerd[1595]: time="2025-05-17T00:22:09.076531842Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 May 17 00:22:09.079457 containerd[1595]: time="2025-05-17T00:22:09.076558480Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 May 17 00:22:09.079457 containerd[1595]: time="2025-05-17T00:22:09.076628016Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 May 17 00:22:09.079457 containerd[1595]: time="2025-05-17T00:22:09.076649531Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 May 17 00:22:09.079457 containerd[1595]: time="2025-05-17T00:22:09.076942469Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 May 17 00:22:09.079457 containerd[1595]: time="2025-05-17T00:22:09.078982136Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082416633Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082469218Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082490765Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082514998Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082532700Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082549771Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082570895Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082606388Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082623588Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082639549Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082655982Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082686864Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082704146Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 May 17 00:22:09.083930 containerd[1595]: time="2025-05-17T00:22:09.082719843Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082763306Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082782640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082801376Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082817908Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082835135Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082851328Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082869030Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082884528Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082902616Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082921155Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082943405Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.082974865Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.083004114Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 May 17 00:22:09.084510 containerd[1595]: time="2025-05-17T00:22:09.083025869Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083107117Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083132481Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083148018Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083165131Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083181150Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083197210Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083220074Z" level=info msg="NRI interface is disabled by configuration." May 17 00:22:09.085005 containerd[1595]: time="2025-05-17T00:22:09.083242248Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 May 17 00:22:09.085296 containerd[1595]: time="2025-05-17T00:22:09.083652313Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" May 17 00:22:09.085296 containerd[1595]: time="2025-05-17T00:22:09.083745923Z" level=info msg="Connect containerd service" May 17 00:22:09.085296 containerd[1595]: time="2025-05-17T00:22:09.083805300Z" level=info msg="using legacy CRI server" May 17 00:22:09.085296 containerd[1595]: time="2025-05-17T00:22:09.083820382Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 17 00:22:09.096698 containerd[1595]: time="2025-05-17T00:22:09.092549482Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" May 17 00:22:09.100798 containerd[1595]: time="2025-05-17T00:22:09.100740963Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 17 00:22:09.102031 containerd[1595]: time="2025-05-17T00:22:09.101654818Z" level=info msg="Start subscribing containerd event" May 17 00:22:09.102031 containerd[1595]: time="2025-05-17T00:22:09.101738236Z" level=info msg="Start recovering state" May 17 00:22:09.102031 containerd[1595]: time="2025-05-17T00:22:09.101835374Z" level=info msg="Start event monitor" May 17 00:22:09.102031 containerd[1595]: time="2025-05-17T00:22:09.101857238Z" level=info msg="Start snapshots syncer" May 17 00:22:09.102031 containerd[1595]: time="2025-05-17T00:22:09.101870022Z" level=info msg="Start cni network conf syncer for default" May 17 00:22:09.102031 containerd[1595]: time="2025-05-17T00:22:09.101879870Z" level=info msg="Start streaming server" May 17 00:22:09.103749 containerd[1595]: time="2025-05-17T00:22:09.103712101Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 17 00:22:09.105348 containerd[1595]: time="2025-05-17T00:22:09.103928001Z" level=info msg=serving... address=/run/containerd/containerd.sock May 17 00:22:09.105348 containerd[1595]: time="2025-05-17T00:22:09.104012138Z" level=info msg="containerd successfully booted in 0.100111s" May 17 00:22:09.104755 systemd[1]: Started containerd.service - containerd container runtime. May 17 00:22:09.168831 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 17 00:22:09.179337 systemd[1]: Starting issuegen.service - Generate /run/issue... May 17 00:22:09.223653 systemd[1]: issuegen.service: Deactivated successfully. May 17 00:22:09.226016 systemd[1]: Finished issuegen.service - Generate /run/issue. May 17 00:22:09.246501 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 17 00:22:09.291723 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 17 00:22:09.312372 systemd[1]: Started getty@tty1.service - Getty on tty1. May 17 00:22:09.330413 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 17 00:22:09.334695 systemd[1]: Reached target getty.target - Login Prompts. May 17 00:22:09.620129 tar[1582]: linux-amd64/LICENSE May 17 00:22:09.620129 tar[1582]: linux-amd64/README.md May 17 00:22:09.643078 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 17 00:22:10.041618 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 17 00:22:10.046438 systemd[1]: Reached target multi-user.target - Multi-User System. May 17 00:22:10.046719 (kubelet)[1694]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 17 00:22:10.050449 systemd[1]: Startup finished in 7.539s (kernel) + 6.907s (userspace) = 14.447s. May 17 00:22:10.461585 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 17 00:22:10.468845 systemd[1]: Started sshd@0-64.23.167.227:22-139.178.68.195:59560.service - OpenSSH per-connection server daemon (139.178.68.195:59560). May 17 00:22:10.559437 sshd[1703]: Accepted publickey for core from 139.178.68.195 port 59560 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:22:10.562906 sshd[1703]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:22:10.576265 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 17 00:22:10.583986 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 17 00:22:10.589999 systemd-logind[1571]: New session 1 of user core. May 17 00:22:10.624039 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 17 00:22:10.632210 systemd[1]: Starting user@500.service - User Manager for UID 500... May 17 00:22:10.646944 (systemd)[1709]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 17 00:22:10.730203 kubelet[1694]: E0517 00:22:10.730000 1694 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 17 00:22:10.735061 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 17 00:22:10.735449 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 17 00:22:10.810409 systemd[1709]: Queued start job for default target default.target. May 17 00:22:10.811515 systemd[1709]: Created slice app.slice - User Application Slice. May 17 00:22:10.811745 systemd[1709]: Reached target paths.target - Paths. May 17 00:22:10.811844 systemd[1709]: Reached target timers.target - Timers. May 17 00:22:10.821563 systemd[1709]: Starting dbus.socket - D-Bus User Message Bus Socket... May 17 00:22:10.833530 systemd[1709]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 17 00:22:10.833914 systemd[1709]: Reached target sockets.target - Sockets. May 17 00:22:10.834065 systemd[1709]: Reached target basic.target - Basic System. May 17 00:22:10.834440 systemd[1709]: Reached target default.target - Main User Target. May 17 00:22:10.834790 systemd[1]: Started user@500.service - User Manager for UID 500. May 17 00:22:10.836416 systemd[1709]: Startup finished in 175ms. May 17 00:22:10.839928 systemd[1]: Started session-1.scope - Session 1 of User core. May 17 00:22:10.909791 systemd[1]: Started sshd@1-64.23.167.227:22-139.178.68.195:59566.service - OpenSSH per-connection server daemon (139.178.68.195:59566). May 17 00:22:10.958142 sshd[1724]: Accepted publickey for core from 139.178.68.195 port 59566 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:22:10.960454 sshd[1724]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:22:10.968231 systemd-logind[1571]: New session 2 of user core. May 17 00:22:10.971868 systemd[1]: Started session-2.scope - Session 2 of User core. May 17 00:22:11.041796 sshd[1724]: pam_unix(sshd:session): session closed for user core May 17 00:22:11.053780 systemd[1]: Started sshd@2-64.23.167.227:22-139.178.68.195:59578.service - OpenSSH per-connection server daemon (139.178.68.195:59578). May 17 00:22:11.055056 systemd[1]: sshd@1-64.23.167.227:22-139.178.68.195:59566.service: Deactivated successfully. May 17 00:22:11.059025 systemd[1]: session-2.scope: Deactivated successfully. May 17 00:22:11.061492 systemd-logind[1571]: Session 2 logged out. Waiting for processes to exit. May 17 00:22:11.063661 systemd-logind[1571]: Removed session 2. May 17 00:22:11.104080 sshd[1730]: Accepted publickey for core from 139.178.68.195 port 59578 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:22:11.105937 sshd[1730]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:22:11.113793 systemd-logind[1571]: New session 3 of user core. May 17 00:22:11.119919 systemd[1]: Started session-3.scope - Session 3 of User core. May 17 00:22:11.182448 sshd[1730]: pam_unix(sshd:session): session closed for user core May 17 00:22:11.192786 systemd[1]: Started sshd@3-64.23.167.227:22-139.178.68.195:59588.service - OpenSSH per-connection server daemon (139.178.68.195:59588). May 17 00:22:11.193976 systemd[1]: sshd@2-64.23.167.227:22-139.178.68.195:59578.service: Deactivated successfully. May 17 00:22:11.198388 systemd[1]: session-3.scope: Deactivated successfully. May 17 00:22:11.199887 systemd-logind[1571]: Session 3 logged out. Waiting for processes to exit. May 17 00:22:11.202291 systemd-logind[1571]: Removed session 3. May 17 00:22:11.249501 sshd[1738]: Accepted publickey for core from 139.178.68.195 port 59588 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:22:11.252262 sshd[1738]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:22:11.259141 systemd-logind[1571]: New session 4 of user core. May 17 00:22:11.268919 systemd[1]: Started session-4.scope - Session 4 of User core. May 17 00:22:11.334620 sshd[1738]: pam_unix(sshd:session): session closed for user core May 17 00:22:11.344790 systemd[1]: Started sshd@4-64.23.167.227:22-139.178.68.195:59594.service - OpenSSH per-connection server daemon (139.178.68.195:59594). May 17 00:22:11.345366 systemd[1]: sshd@3-64.23.167.227:22-139.178.68.195:59588.service: Deactivated successfully. May 17 00:22:11.354646 systemd[1]: session-4.scope: Deactivated successfully. May 17 00:22:11.356812 systemd-logind[1571]: Session 4 logged out. Waiting for processes to exit. May 17 00:22:11.358247 systemd-logind[1571]: Removed session 4. May 17 00:22:11.388246 sshd[1745]: Accepted publickey for core from 139.178.68.195 port 59594 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:22:11.389559 sshd[1745]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:22:11.396722 systemd-logind[1571]: New session 5 of user core. May 17 00:22:11.402796 systemd[1]: Started session-5.scope - Session 5 of User core. May 17 00:22:11.474503 sudo[1752]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 17 00:22:11.475673 sudo[1752]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 17 00:22:11.490415 sudo[1752]: pam_unix(sudo:session): session closed for user root May 17 00:22:11.497095 sshd[1745]: pam_unix(sshd:session): session closed for user core May 17 00:22:11.507981 systemd[1]: Started sshd@5-64.23.167.227:22-139.178.68.195:59606.service - OpenSSH per-connection server daemon (139.178.68.195:59606). May 17 00:22:11.508761 systemd[1]: sshd@4-64.23.167.227:22-139.178.68.195:59594.service: Deactivated successfully. May 17 00:22:11.512730 systemd[1]: session-5.scope: Deactivated successfully. May 17 00:22:11.514771 systemd-logind[1571]: Session 5 logged out. Waiting for processes to exit. May 17 00:22:11.518476 systemd-logind[1571]: Removed session 5. May 17 00:22:11.553791 sshd[1754]: Accepted publickey for core from 139.178.68.195 port 59606 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:22:11.555753 sshd[1754]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:22:11.562404 systemd-logind[1571]: New session 6 of user core. May 17 00:22:11.565821 systemd[1]: Started session-6.scope - Session 6 of User core. May 17 00:22:11.629837 sudo[1762]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 17 00:22:11.630173 sudo[1762]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 17 00:22:11.636202 sudo[1762]: pam_unix(sudo:session): session closed for user root May 17 00:22:11.645043 sudo[1761]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules May 17 00:22:11.645416 sudo[1761]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 17 00:22:11.665725 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... May 17 00:22:11.668351 auditctl[1765]: No rules May 17 00:22:11.668804 systemd[1]: audit-rules.service: Deactivated successfully. May 17 00:22:11.669129 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. May 17 00:22:11.681249 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... May 17 00:22:11.713476 augenrules[1784]: No rules May 17 00:22:11.716087 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. May 17 00:22:11.719041 sudo[1761]: pam_unix(sudo:session): session closed for user root May 17 00:22:11.722434 sshd[1754]: pam_unix(sshd:session): session closed for user core May 17 00:22:11.734675 systemd[1]: Started sshd@6-64.23.167.227:22-139.178.68.195:59616.service - OpenSSH per-connection server daemon (139.178.68.195:59616). May 17 00:22:11.735305 systemd[1]: sshd@5-64.23.167.227:22-139.178.68.195:59606.service: Deactivated successfully. May 17 00:22:11.737160 systemd[1]: session-6.scope: Deactivated successfully. May 17 00:22:11.738816 systemd-logind[1571]: Session 6 logged out. Waiting for processes to exit. May 17 00:22:11.741088 systemd-logind[1571]: Removed session 6. May 17 00:22:11.783100 sshd[1791]: Accepted publickey for core from 139.178.68.195 port 59616 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:22:11.785617 sshd[1791]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:22:11.791591 systemd-logind[1571]: New session 7 of user core. May 17 00:22:11.797999 systemd[1]: Started session-7.scope - Session 7 of User core. May 17 00:22:11.859835 sudo[1797]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 17 00:22:11.860223 sudo[1797]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 17 00:22:12.390053 (dockerd)[1813]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 17 00:22:12.390619 systemd[1]: Starting docker.service - Docker Application Container Engine... May 17 00:22:12.920362 dockerd[1813]: time="2025-05-17T00:22:12.920206470Z" level=info msg="Starting up" May 17 00:22:13.059391 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport3732982324-merged.mount: Deactivated successfully. May 17 00:22:13.162067 dockerd[1813]: time="2025-05-17T00:22:13.161989195Z" level=info msg="Loading containers: start." May 17 00:22:13.311463 kernel: Initializing XFRM netlink socket May 17 00:22:13.419118 systemd-networkd[1219]: docker0: Link UP May 17 00:22:13.435813 dockerd[1813]: time="2025-05-17T00:22:13.435668612Z" level=info msg="Loading containers: done." May 17 00:22:13.459091 dockerd[1813]: time="2025-05-17T00:22:13.458969940Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 17 00:22:13.459273 dockerd[1813]: time="2025-05-17T00:22:13.459217412Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 May 17 00:22:13.459808 dockerd[1813]: time="2025-05-17T00:22:13.459408221Z" level=info msg="Daemon has completed initialization" May 17 00:22:13.500088 dockerd[1813]: time="2025-05-17T00:22:13.499882896Z" level=info msg="API listen on /run/docker.sock" May 17 00:22:13.500661 systemd[1]: Started docker.service - Docker Application Container Engine. May 17 00:22:14.339826 containerd[1595]: time="2025-05-17T00:22:14.339751370Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.9\"" May 17 00:22:14.897888 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount616227369.mount: Deactivated successfully. May 17 00:22:16.103738 containerd[1595]: time="2025-05-17T00:22:16.103655028Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:16.105172 containerd[1595]: time="2025-05-17T00:22:16.105094387Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.9: active requests=0, bytes read=28078845" May 17 00:22:16.106347 containerd[1595]: time="2025-05-17T00:22:16.105409519Z" level=info msg="ImageCreate event name:\"sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:16.109351 containerd[1595]: time="2025-05-17T00:22:16.108611829Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:5b68f0df22013422dc8fb9ddfcff513eb6fc92f9dbf8aae41555c895efef5a20\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:16.111352 containerd[1595]: time="2025-05-17T00:22:16.110421401Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.9\" with image id \"sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:5b68f0df22013422dc8fb9ddfcff513eb6fc92f9dbf8aae41555c895efef5a20\", size \"28075645\" in 1.770587135s" May 17 00:22:16.111352 containerd[1595]: time="2025-05-17T00:22:16.110491369Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.9\" returns image reference \"sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1\"" May 17 00:22:16.111882 containerd[1595]: time="2025-05-17T00:22:16.111833188Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.9\"" May 17 00:22:17.435614 containerd[1595]: time="2025-05-17T00:22:17.435510671Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:17.437094 containerd[1595]: time="2025-05-17T00:22:17.436773512Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.9: active requests=0, bytes read=24713522" May 17 00:22:17.438034 containerd[1595]: time="2025-05-17T00:22:17.437985827Z" level=info msg="ImageCreate event name:\"sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:17.442237 containerd[1595]: time="2025-05-17T00:22:17.442162816Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:be9e7987d323b38a12e28436cff6d6ec6fc31ffdd3ea11eaa9d74852e9d31248\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:17.443838 containerd[1595]: time="2025-05-17T00:22:17.443571847Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.9\" with image id \"sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:be9e7987d323b38a12e28436cff6d6ec6fc31ffdd3ea11eaa9d74852e9d31248\", size \"26315362\" in 1.331548493s" May 17 00:22:17.443838 containerd[1595]: time="2025-05-17T00:22:17.443626131Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.9\" returns image reference \"sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2\"" May 17 00:22:17.445206 containerd[1595]: time="2025-05-17T00:22:17.445148223Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.9\"" May 17 00:22:18.668079 containerd[1595]: time="2025-05-17T00:22:18.667436602Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:18.669872 containerd[1595]: time="2025-05-17T00:22:18.669319412Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.9: active requests=0, bytes read=18784311" May 17 00:22:18.671381 containerd[1595]: time="2025-05-17T00:22:18.670836898Z" level=info msg="ImageCreate event name:\"sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:18.675031 containerd[1595]: time="2025-05-17T00:22:18.674955738Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:eb358c7346bb17ab2c639c3ff8ab76a147dec7ae609f5c0c2800233e42253ed1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:18.680978 containerd[1595]: time="2025-05-17T00:22:18.680897047Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.9\" with image id \"sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:eb358c7346bb17ab2c639c3ff8ab76a147dec7ae609f5c0c2800233e42253ed1\", size \"20386169\" in 1.235583414s" May 17 00:22:18.681268 containerd[1595]: time="2025-05-17T00:22:18.681240824Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.9\" returns image reference \"sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3\"" May 17 00:22:18.686391 containerd[1595]: time="2025-05-17T00:22:18.685846810Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.9\"" May 17 00:22:19.786136 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3956527641.mount: Deactivated successfully. May 17 00:22:20.323832 containerd[1595]: time="2025-05-17T00:22:20.323777342Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:20.326367 containerd[1595]: time="2025-05-17T00:22:20.326201791Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.9: active requests=0, bytes read=30355623" May 17 00:22:20.327363 containerd[1595]: time="2025-05-17T00:22:20.326986835Z" level=info msg="ImageCreate event name:\"sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:20.330493 containerd[1595]: time="2025-05-17T00:22:20.329429116Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:fdf026cf2434537e499e9c739d189ca8fc57101d929ac5ccd8e24f979a9738c1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:20.330493 containerd[1595]: time="2025-05-17T00:22:20.330297116Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.9\" with image id \"sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578\", repo tag \"registry.k8s.io/kube-proxy:v1.31.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:fdf026cf2434537e499e9c739d189ca8fc57101d929ac5ccd8e24f979a9738c1\", size \"30354642\" in 1.644386295s" May 17 00:22:20.330493 containerd[1595]: time="2025-05-17T00:22:20.330364947Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.9\" returns image reference \"sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578\"" May 17 00:22:20.331695 containerd[1595]: time="2025-05-17T00:22:20.331661192Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" May 17 00:22:20.638350 systemd-resolved[1473]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. May 17 00:22:20.767145 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 17 00:22:20.773208 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 17 00:22:20.800245 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2993825470.mount: Deactivated successfully. May 17 00:22:20.990563 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 17 00:22:21.000911 (kubelet)[2051]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 17 00:22:21.104143 kubelet[2051]: E0517 00:22:21.104029 2051 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 17 00:22:21.108245 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 17 00:22:21.108588 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 17 00:22:22.246778 containerd[1595]: time="2025-05-17T00:22:22.246588919Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:22.249367 containerd[1595]: time="2025-05-17T00:22:22.248810596Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" May 17 00:22:22.251348 containerd[1595]: time="2025-05-17T00:22:22.250003878Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:22.255649 containerd[1595]: time="2025-05-17T00:22:22.255562186Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:22.256476 containerd[1595]: time="2025-05-17T00:22:22.256421527Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.92457488s" May 17 00:22:22.256568 containerd[1595]: time="2025-05-17T00:22:22.256510540Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" May 17 00:22:22.258344 containerd[1595]: time="2025-05-17T00:22:22.258289815Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 17 00:22:22.695916 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3899458432.mount: Deactivated successfully. May 17 00:22:22.700934 containerd[1595]: time="2025-05-17T00:22:22.700862162Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:22.702357 containerd[1595]: time="2025-05-17T00:22:22.701967955Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" May 17 00:22:22.703091 containerd[1595]: time="2025-05-17T00:22:22.703035776Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:22.705902 containerd[1595]: time="2025-05-17T00:22:22.705837595Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:22.707235 containerd[1595]: time="2025-05-17T00:22:22.706725209Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 448.196164ms" May 17 00:22:22.707235 containerd[1595]: time="2025-05-17T00:22:22.706762658Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" May 17 00:22:22.707494 containerd[1595]: time="2025-05-17T00:22:22.707471481Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" May 17 00:22:23.215022 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2122027218.mount: Deactivated successfully. May 17 00:22:23.959711 systemd-resolved[1473]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. May 17 00:22:25.075415 containerd[1595]: time="2025-05-17T00:22:25.075015849Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56780013" May 17 00:22:25.077590 containerd[1595]: time="2025-05-17T00:22:25.076581530Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:25.079926 containerd[1595]: time="2025-05-17T00:22:25.079871207Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:25.081858 containerd[1595]: time="2025-05-17T00:22:25.081779467Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 2.374271174s" May 17 00:22:25.081988 containerd[1595]: time="2025-05-17T00:22:25.081864555Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" May 17 00:22:25.085651 containerd[1595]: time="2025-05-17T00:22:25.084941538Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:27.873506 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 17 00:22:27.881696 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 17 00:22:27.928224 systemd[1]: Reloading requested from client PID 2186 ('systemctl') (unit session-7.scope)... May 17 00:22:27.928557 systemd[1]: Reloading... May 17 00:22:28.100342 zram_generator::config[2231]: No configuration found. May 17 00:22:28.254787 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:22:28.357806 systemd[1]: Reloading finished in 428 ms. May 17 00:22:28.400045 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 17 00:22:28.400138 systemd[1]: kubelet.service: Failed with result 'signal'. May 17 00:22:28.400643 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 17 00:22:28.406072 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 17 00:22:28.579726 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 17 00:22:28.588283 (kubelet)[2288]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 17 00:22:28.641271 kubelet[2288]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:22:28.641271 kubelet[2288]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 17 00:22:28.641271 kubelet[2288]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:22:28.641826 kubelet[2288]: I0517 00:22:28.641387 2288 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 17 00:22:29.081665 kubelet[2288]: I0517 00:22:29.079688 2288 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" May 17 00:22:29.081665 kubelet[2288]: I0517 00:22:29.079736 2288 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 17 00:22:29.081665 kubelet[2288]: I0517 00:22:29.080051 2288 server.go:934] "Client rotation is on, will bootstrap in background" May 17 00:22:29.107801 kubelet[2288]: E0517 00:22:29.107758 2288 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://64.23.167.227:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:29.116276 kubelet[2288]: I0517 00:22:29.116217 2288 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 17 00:22:29.125535 kubelet[2288]: E0517 00:22:29.125497 2288 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" May 17 00:22:29.125776 kubelet[2288]: I0517 00:22:29.125754 2288 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." May 17 00:22:29.132348 kubelet[2288]: I0517 00:22:29.132279 2288 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 17 00:22:29.133081 kubelet[2288]: I0517 00:22:29.133054 2288 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" May 17 00:22:29.133459 kubelet[2288]: I0517 00:22:29.133412 2288 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 17 00:22:29.133869 kubelet[2288]: I0517 00:22:29.133570 2288 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.3-n-0e116cd3eb","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} May 17 00:22:29.134162 kubelet[2288]: I0517 00:22:29.134145 2288 topology_manager.go:138] "Creating topology manager with none policy" May 17 00:22:29.134244 kubelet[2288]: I0517 00:22:29.134236 2288 container_manager_linux.go:300] "Creating device plugin manager" May 17 00:22:29.134527 kubelet[2288]: I0517 00:22:29.134509 2288 state_mem.go:36] "Initialized new in-memory state store" May 17 00:22:29.139990 kubelet[2288]: I0517 00:22:29.139947 2288 kubelet.go:408] "Attempting to sync node with API server" May 17 00:22:29.140208 kubelet[2288]: I0517 00:22:29.140190 2288 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" May 17 00:22:29.140385 kubelet[2288]: I0517 00:22:29.140371 2288 kubelet.go:314] "Adding apiserver pod source" May 17 00:22:29.140521 kubelet[2288]: I0517 00:22:29.140507 2288 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 17 00:22:29.141649 kubelet[2288]: W0517 00:22:29.141562 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://64.23.167.227:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-n-0e116cd3eb&limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:29.141758 kubelet[2288]: E0517 00:22:29.141657 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://64.23.167.227:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-n-0e116cd3eb&limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:29.143929 kubelet[2288]: I0517 00:22:29.143905 2288 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" May 17 00:22:29.148117 kubelet[2288]: I0517 00:22:29.147940 2288 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 17 00:22:29.149251 kubelet[2288]: W0517 00:22:29.148500 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://64.23.167.227:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:29.149251 kubelet[2288]: E0517 00:22:29.148567 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://64.23.167.227:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:29.149251 kubelet[2288]: W0517 00:22:29.148883 2288 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 17 00:22:29.150250 kubelet[2288]: I0517 00:22:29.150231 2288 server.go:1274] "Started kubelet" May 17 00:22:29.151555 kubelet[2288]: I0517 00:22:29.150909 2288 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 17 00:22:29.152684 kubelet[2288]: I0517 00:22:29.152096 2288 server.go:449] "Adding debug handlers to kubelet server" May 17 00:22:29.154498 kubelet[2288]: I0517 00:22:29.154454 2288 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 17 00:22:29.154930 kubelet[2288]: I0517 00:22:29.154908 2288 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 17 00:22:29.156672 kubelet[2288]: E0517 00:22:29.155226 2288 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://64.23.167.227:6443/api/v1/namespaces/default/events\": dial tcp 64.23.167.227:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.3-n-0e116cd3eb.184028a611ed3e6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.3-n-0e116cd3eb,UID:ci-4081.3.3-n-0e116cd3eb,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.3-n-0e116cd3eb,},FirstTimestamp:2025-05-17 00:22:29.150195311 +0000 UTC m=+0.556264887,LastTimestamp:2025-05-17 00:22:29.150195311 +0000 UTC m=+0.556264887,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.3-n-0e116cd3eb,}" May 17 00:22:29.158790 kubelet[2288]: I0517 00:22:29.158177 2288 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 17 00:22:29.161654 kubelet[2288]: I0517 00:22:29.159122 2288 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 17 00:22:29.167601 kubelet[2288]: I0517 00:22:29.166847 2288 volume_manager.go:289] "Starting Kubelet Volume Manager" May 17 00:22:29.167601 kubelet[2288]: E0517 00:22:29.167395 2288 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:29.171260 kubelet[2288]: I0517 00:22:29.170813 2288 desired_state_of_world_populator.go:147] "Desired state populator starts to run" May 17 00:22:29.171260 kubelet[2288]: I0517 00:22:29.170927 2288 reconciler.go:26] "Reconciler: start to sync state" May 17 00:22:29.172275 kubelet[2288]: E0517 00:22:29.172233 2288 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.167.227:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-n-0e116cd3eb?timeout=10s\": dial tcp 64.23.167.227:6443: connect: connection refused" interval="200ms" May 17 00:22:29.172736 kubelet[2288]: I0517 00:22:29.172712 2288 factory.go:221] Registration of the systemd container factory successfully May 17 00:22:29.172991 kubelet[2288]: I0517 00:22:29.172968 2288 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 17 00:22:29.177333 kubelet[2288]: I0517 00:22:29.177287 2288 factory.go:221] Registration of the containerd container factory successfully May 17 00:22:29.181663 kubelet[2288]: W0517 00:22:29.181581 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://64.23.167.227:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:29.182735 kubelet[2288]: E0517 00:22:29.182699 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://64.23.167.227:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:29.186728 kubelet[2288]: I0517 00:22:29.186674 2288 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 17 00:22:29.191569 kubelet[2288]: I0517 00:22:29.191530 2288 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 17 00:22:29.191706 kubelet[2288]: I0517 00:22:29.191592 2288 status_manager.go:217] "Starting to sync pod status with apiserver" May 17 00:22:29.191706 kubelet[2288]: I0517 00:22:29.191623 2288 kubelet.go:2321] "Starting kubelet main sync loop" May 17 00:22:29.191706 kubelet[2288]: E0517 00:22:29.191677 2288 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 17 00:22:29.194652 kubelet[2288]: E0517 00:22:29.178181 2288 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 17 00:22:29.206485 kubelet[2288]: W0517 00:22:29.206407 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://64.23.167.227:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:29.206485 kubelet[2288]: E0517 00:22:29.206486 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://64.23.167.227:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:29.222361 kubelet[2288]: I0517 00:22:29.222328 2288 cpu_manager.go:214] "Starting CPU manager" policy="none" May 17 00:22:29.222796 kubelet[2288]: I0517 00:22:29.222687 2288 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 17 00:22:29.222796 kubelet[2288]: I0517 00:22:29.222750 2288 state_mem.go:36] "Initialized new in-memory state store" May 17 00:22:29.226722 kubelet[2288]: I0517 00:22:29.226553 2288 policy_none.go:49] "None policy: Start" May 17 00:22:29.227592 kubelet[2288]: I0517 00:22:29.227566 2288 memory_manager.go:170] "Starting memorymanager" policy="None" May 17 00:22:29.227713 kubelet[2288]: I0517 00:22:29.227603 2288 state_mem.go:35] "Initializing new in-memory state store" May 17 00:22:29.233416 kubelet[2288]: I0517 00:22:29.233327 2288 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 17 00:22:29.233805 kubelet[2288]: I0517 00:22:29.233622 2288 eviction_manager.go:189] "Eviction manager: starting control loop" May 17 00:22:29.233805 kubelet[2288]: I0517 00:22:29.233655 2288 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 17 00:22:29.235934 kubelet[2288]: I0517 00:22:29.235834 2288 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 17 00:22:29.238780 kubelet[2288]: E0517 00:22:29.238726 2288 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:29.336112 kubelet[2288]: I0517 00:22:29.335927 2288 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.337851 kubelet[2288]: E0517 00:22:29.337801 2288 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.167.227:6443/api/v1/nodes\": dial tcp 64.23.167.227:6443: connect: connection refused" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.374409 kubelet[2288]: E0517 00:22:29.374355 2288 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.167.227:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-n-0e116cd3eb?timeout=10s\": dial tcp 64.23.167.227:6443: connect: connection refused" interval="400ms" May 17 00:22:29.471702 kubelet[2288]: I0517 00:22:29.471615 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/379bfc971db3b4ae016ab6b991edb0cc-ca-certs\") pod \"kube-apiserver-ci-4081.3.3-n-0e116cd3eb\" (UID: \"379bfc971db3b4ae016ab6b991edb0cc\") " pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.471702 kubelet[2288]: I0517 00:22:29.471668 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/379bfc971db3b4ae016ab6b991edb0cc-k8s-certs\") pod \"kube-apiserver-ci-4081.3.3-n-0e116cd3eb\" (UID: \"379bfc971db3b4ae016ab6b991edb0cc\") " pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.471702 kubelet[2288]: I0517 00:22:29.471686 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-ca-certs\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.471702 kubelet[2288]: I0517 00:22:29.471709 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.471702 kubelet[2288]: I0517 00:22:29.471726 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.472077 kubelet[2288]: I0517 00:22:29.471742 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/34fd926d0d45564074f8fc2aed24959b-kubeconfig\") pod \"kube-scheduler-ci-4081.3.3-n-0e116cd3eb\" (UID: \"34fd926d0d45564074f8fc2aed24959b\") " pod="kube-system/kube-scheduler-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.472077 kubelet[2288]: I0517 00:22:29.471757 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/379bfc971db3b4ae016ab6b991edb0cc-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.3-n-0e116cd3eb\" (UID: \"379bfc971db3b4ae016ab6b991edb0cc\") " pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.472077 kubelet[2288]: I0517 00:22:29.471771 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.472077 kubelet[2288]: I0517 00:22:29.471788 2288 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.539283 kubelet[2288]: I0517 00:22:29.539231 2288 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.539652 kubelet[2288]: E0517 00:22:29.539585 2288 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.167.227:6443/api/v1/nodes\": dial tcp 64.23.167.227:6443: connect: connection refused" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.603755 kubelet[2288]: E0517 00:22:29.603519 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:29.603755 kubelet[2288]: E0517 00:22:29.603589 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:29.605695 containerd[1595]: time="2025-05-17T00:22:29.605266510Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.3-n-0e116cd3eb,Uid:379bfc971db3b4ae016ab6b991edb0cc,Namespace:kube-system,Attempt:0,}" May 17 00:22:29.605695 containerd[1595]: time="2025-05-17T00:22:29.605604212Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.3-n-0e116cd3eb,Uid:b237a751c3a466b82870764fcca590b0,Namespace:kube-system,Attempt:0,}" May 17 00:22:29.607940 systemd-resolved[1473]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.3. May 17 00:22:29.608378 kubelet[2288]: E0517 00:22:29.608352 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:29.612259 containerd[1595]: time="2025-05-17T00:22:29.612188105Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.3-n-0e116cd3eb,Uid:34fd926d0d45564074f8fc2aed24959b,Namespace:kube-system,Attempt:0,}" May 17 00:22:29.775077 kubelet[2288]: E0517 00:22:29.775000 2288 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.167.227:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-n-0e116cd3eb?timeout=10s\": dial tcp 64.23.167.227:6443: connect: connection refused" interval="800ms" May 17 00:22:29.941929 kubelet[2288]: I0517 00:22:29.941764 2288 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:29.942572 kubelet[2288]: E0517 00:22:29.942532 2288 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.167.227:6443/api/v1/nodes\": dial tcp 64.23.167.227:6443: connect: connection refused" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:30.054967 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2096029874.mount: Deactivated successfully. May 17 00:22:30.059098 containerd[1595]: time="2025-05-17T00:22:30.059047159Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 17 00:22:30.060475 containerd[1595]: time="2025-05-17T00:22:30.060421438Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" May 17 00:22:30.061522 containerd[1595]: time="2025-05-17T00:22:30.061473411Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 17 00:22:30.067458 containerd[1595]: time="2025-05-17T00:22:30.067383471Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 17 00:22:30.068226 containerd[1595]: time="2025-05-17T00:22:30.068194205Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 462.496303ms" May 17 00:22:30.071589 containerd[1595]: time="2025-05-17T00:22:30.071397767Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" May 17 00:22:30.072777 containerd[1595]: time="2025-05-17T00:22:30.072729101Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" May 17 00:22:30.073630 containerd[1595]: time="2025-05-17T00:22:30.073594785Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 17 00:22:30.075127 containerd[1595]: time="2025-05-17T00:22:30.075092801Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 17 00:22:30.079000 containerd[1595]: time="2025-05-17T00:22:30.078954128Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 466.648259ms" May 17 00:22:30.108393 containerd[1595]: time="2025-05-17T00:22:30.108043631Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 502.639849ms" May 17 00:22:30.193859 kubelet[2288]: W0517 00:22:30.193653 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://64.23.167.227:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:30.194202 kubelet[2288]: E0517 00:22:30.194152 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://64.23.167.227:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:30.248754 containerd[1595]: time="2025-05-17T00:22:30.248130894Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:22:30.249300 containerd[1595]: time="2025-05-17T00:22:30.249132829Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:22:30.249727 containerd[1595]: time="2025-05-17T00:22:30.249261344Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:30.251195 containerd[1595]: time="2025-05-17T00:22:30.251014425Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:30.251195 containerd[1595]: time="2025-05-17T00:22:30.249604367Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:22:30.251195 containerd[1595]: time="2025-05-17T00:22:30.250927749Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:22:30.251195 containerd[1595]: time="2025-05-17T00:22:30.250950026Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:30.252138 containerd[1595]: time="2025-05-17T00:22:30.252014077Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:30.258682 containerd[1595]: time="2025-05-17T00:22:30.258527666Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:22:30.258682 containerd[1595]: time="2025-05-17T00:22:30.258610995Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:22:30.258682 containerd[1595]: time="2025-05-17T00:22:30.258632934Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:30.258985 containerd[1595]: time="2025-05-17T00:22:30.258759815Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:30.321849 kubelet[2288]: W0517 00:22:30.321780 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://64.23.167.227:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-n-0e116cd3eb&limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:30.322328 kubelet[2288]: E0517 00:22:30.322143 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://64.23.167.227:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.3-n-0e116cd3eb&limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:30.372954 containerd[1595]: time="2025-05-17T00:22:30.372905823Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.3-n-0e116cd3eb,Uid:34fd926d0d45564074f8fc2aed24959b,Namespace:kube-system,Attempt:0,} returns sandbox id \"157bd6e4c8e8e3f4aa1a08eea15e7d8d3bc62ee8f70b1c0023f933865be3ef72\"" May 17 00:22:30.377624 kubelet[2288]: E0517 00:22:30.377594 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:30.385747 containerd[1595]: time="2025-05-17T00:22:30.385669534Z" level=info msg="CreateContainer within sandbox \"157bd6e4c8e8e3f4aa1a08eea15e7d8d3bc62ee8f70b1c0023f933865be3ef72\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 17 00:22:30.386549 containerd[1595]: time="2025-05-17T00:22:30.386495559Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.3-n-0e116cd3eb,Uid:379bfc971db3b4ae016ab6b991edb0cc,Namespace:kube-system,Attempt:0,} returns sandbox id \"2e597fb031df8d2b8eb988c7acb863708401aba10c3412946a69236ebba7dda9\"" May 17 00:22:30.388851 kubelet[2288]: E0517 00:22:30.388355 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:30.391252 containerd[1595]: time="2025-05-17T00:22:30.391200295Z" level=info msg="CreateContainer within sandbox \"2e597fb031df8d2b8eb988c7acb863708401aba10c3412946a69236ebba7dda9\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 17 00:22:30.393585 containerd[1595]: time="2025-05-17T00:22:30.393540859Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.3-n-0e116cd3eb,Uid:b237a751c3a466b82870764fcca590b0,Namespace:kube-system,Attempt:0,} returns sandbox id \"48678c2442a330d3b16dee63e6cd5419e14b6d4eaf30f2690c8590de3ab7de9e\"" May 17 00:22:30.395201 kubelet[2288]: E0517 00:22:30.394659 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:30.397749 containerd[1595]: time="2025-05-17T00:22:30.397696009Z" level=info msg="CreateContainer within sandbox \"48678c2442a330d3b16dee63e6cd5419e14b6d4eaf30f2690c8590de3ab7de9e\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 17 00:22:30.407047 containerd[1595]: time="2025-05-17T00:22:30.406978019Z" level=info msg="CreateContainer within sandbox \"157bd6e4c8e8e3f4aa1a08eea15e7d8d3bc62ee8f70b1c0023f933865be3ef72\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"42c360e1eafb52df4c0dd5ec7db944bf92fde54f754dcb7745b11d472f83f65e\"" May 17 00:22:30.408250 containerd[1595]: time="2025-05-17T00:22:30.408209690Z" level=info msg="StartContainer for \"42c360e1eafb52df4c0dd5ec7db944bf92fde54f754dcb7745b11d472f83f65e\"" May 17 00:22:30.409056 containerd[1595]: time="2025-05-17T00:22:30.409020924Z" level=info msg="CreateContainer within sandbox \"2e597fb031df8d2b8eb988c7acb863708401aba10c3412946a69236ebba7dda9\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"09505c04a0467f7f2a05b36414f01e823faab642cded8bcf5632a6785a0b0a30\"" May 17 00:22:30.410480 containerd[1595]: time="2025-05-17T00:22:30.410120573Z" level=info msg="CreateContainer within sandbox \"48678c2442a330d3b16dee63e6cd5419e14b6d4eaf30f2690c8590de3ab7de9e\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"8985ad185510c61b72d124867d88051297504f2502950892334ce299f90902a6\"" May 17 00:22:30.411059 containerd[1595]: time="2025-05-17T00:22:30.411035014Z" level=info msg="StartContainer for \"8985ad185510c61b72d124867d88051297504f2502950892334ce299f90902a6\"" May 17 00:22:30.413536 containerd[1595]: time="2025-05-17T00:22:30.412447711Z" level=info msg="StartContainer for \"09505c04a0467f7f2a05b36414f01e823faab642cded8bcf5632a6785a0b0a30\"" May 17 00:22:30.570427 containerd[1595]: time="2025-05-17T00:22:30.569490415Z" level=info msg="StartContainer for \"8985ad185510c61b72d124867d88051297504f2502950892334ce299f90902a6\" returns successfully" May 17 00:22:30.576107 kubelet[2288]: E0517 00:22:30.576037 2288 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://64.23.167.227:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.3-n-0e116cd3eb?timeout=10s\": dial tcp 64.23.167.227:6443: connect: connection refused" interval="1.6s" May 17 00:22:30.582653 containerd[1595]: time="2025-05-17T00:22:30.582582916Z" level=info msg="StartContainer for \"09505c04a0467f7f2a05b36414f01e823faab642cded8bcf5632a6785a0b0a30\" returns successfully" May 17 00:22:30.587777 containerd[1595]: time="2025-05-17T00:22:30.587726892Z" level=info msg="StartContainer for \"42c360e1eafb52df4c0dd5ec7db944bf92fde54f754dcb7745b11d472f83f65e\" returns successfully" May 17 00:22:30.678690 kubelet[2288]: W0517 00:22:30.678581 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://64.23.167.227:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:30.678905 kubelet[2288]: E0517 00:22:30.678874 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://64.23.167.227:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:30.692827 kubelet[2288]: W0517 00:22:30.692706 2288 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://64.23.167.227:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.167.227:6443: connect: connection refused May 17 00:22:30.692999 kubelet[2288]: E0517 00:22:30.692838 2288 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://64.23.167.227:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 64.23.167.227:6443: connect: connection refused" logger="UnhandledError" May 17 00:22:30.747348 kubelet[2288]: I0517 00:22:30.746436 2288 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:30.747348 kubelet[2288]: E0517 00:22:30.746942 2288 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://64.23.167.227:6443/api/v1/nodes\": dial tcp 64.23.167.227:6443: connect: connection refused" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:31.227420 kubelet[2288]: E0517 00:22:31.227178 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:31.236492 kubelet[2288]: E0517 00:22:31.236299 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:31.241610 kubelet[2288]: E0517 00:22:31.241505 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:32.247338 kubelet[2288]: E0517 00:22:32.245355 2288 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:32.348814 kubelet[2288]: I0517 00:22:32.348226 2288 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:32.484141 kubelet[2288]: E0517 00:22:32.484093 2288 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.3-n-0e116cd3eb\" not found" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:32.574125 kubelet[2288]: I0517 00:22:32.573989 2288 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:32.574125 kubelet[2288]: E0517 00:22:32.574036 2288 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4081.3.3-n-0e116cd3eb\": node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:32.589110 kubelet[2288]: E0517 00:22:32.589064 2288 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:32.689651 kubelet[2288]: E0517 00:22:32.689580 2288 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:32.790620 kubelet[2288]: E0517 00:22:32.790558 2288 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:32.891622 kubelet[2288]: E0517 00:22:32.891472 2288 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:33.152634 kubelet[2288]: I0517 00:22:33.152196 2288 apiserver.go:52] "Watching apiserver" May 17 00:22:33.171570 kubelet[2288]: I0517 00:22:33.171480 2288 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" May 17 00:22:34.536083 systemd[1]: Reloading requested from client PID 2558 ('systemctl') (unit session-7.scope)... May 17 00:22:34.536103 systemd[1]: Reloading... May 17 00:22:34.640370 zram_generator::config[2597]: No configuration found. May 17 00:22:34.831449 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:22:35.021628 systemd[1]: Reloading finished in 485 ms. May 17 00:22:35.066176 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 17 00:22:35.081173 systemd[1]: kubelet.service: Deactivated successfully. May 17 00:22:35.081608 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 17 00:22:35.088847 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 17 00:22:35.236591 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 17 00:22:35.249262 (kubelet)[2658]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 17 00:22:35.341702 kubelet[2658]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:22:35.341702 kubelet[2658]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 17 00:22:35.341702 kubelet[2658]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:22:35.341702 kubelet[2658]: I0517 00:22:35.339684 2658 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 17 00:22:35.347127 kubelet[2658]: I0517 00:22:35.347083 2658 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" May 17 00:22:35.347405 kubelet[2658]: I0517 00:22:35.347386 2658 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 17 00:22:35.347840 kubelet[2658]: I0517 00:22:35.347816 2658 server.go:934] "Client rotation is on, will bootstrap in background" May 17 00:22:35.349583 kubelet[2658]: I0517 00:22:35.349562 2658 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 17 00:22:35.356339 kubelet[2658]: I0517 00:22:35.356281 2658 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 17 00:22:35.361068 kubelet[2658]: E0517 00:22:35.361020 2658 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" May 17 00:22:35.361271 kubelet[2658]: I0517 00:22:35.361256 2658 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." May 17 00:22:35.366192 kubelet[2658]: I0517 00:22:35.366130 2658 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 17 00:22:35.367001 kubelet[2658]: I0517 00:22:35.366697 2658 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" May 17 00:22:35.367001 kubelet[2658]: I0517 00:22:35.366824 2658 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 17 00:22:35.367155 kubelet[2658]: I0517 00:22:35.366849 2658 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.3-n-0e116cd3eb","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} May 17 00:22:35.367261 kubelet[2658]: I0517 00:22:35.367176 2658 topology_manager.go:138] "Creating topology manager with none policy" May 17 00:22:35.367261 kubelet[2658]: I0517 00:22:35.367189 2658 container_manager_linux.go:300] "Creating device plugin manager" May 17 00:22:35.367261 kubelet[2658]: I0517 00:22:35.367226 2658 state_mem.go:36] "Initialized new in-memory state store" May 17 00:22:35.367451 kubelet[2658]: I0517 00:22:35.367436 2658 kubelet.go:408] "Attempting to sync node with API server" May 17 00:22:35.367741 kubelet[2658]: I0517 00:22:35.367719 2658 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" May 17 00:22:35.367841 kubelet[2658]: I0517 00:22:35.367833 2658 kubelet.go:314] "Adding apiserver pod source" May 17 00:22:35.367917 kubelet[2658]: I0517 00:22:35.367907 2658 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 17 00:22:35.374099 kubelet[2658]: I0517 00:22:35.374062 2658 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" May 17 00:22:35.374751 kubelet[2658]: I0517 00:22:35.374728 2658 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 17 00:22:35.376276 kubelet[2658]: I0517 00:22:35.375374 2658 server.go:1274] "Started kubelet" May 17 00:22:35.378896 kubelet[2658]: I0517 00:22:35.378874 2658 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 17 00:22:35.388208 kubelet[2658]: I0517 00:22:35.388143 2658 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 17 00:22:35.389509 kubelet[2658]: I0517 00:22:35.389481 2658 server.go:449] "Adding debug handlers to kubelet server" May 17 00:22:35.391065 kubelet[2658]: I0517 00:22:35.391027 2658 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 17 00:22:35.391437 kubelet[2658]: I0517 00:22:35.391414 2658 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 17 00:22:35.391855 kubelet[2658]: I0517 00:22:35.391835 2658 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 17 00:22:35.393633 kubelet[2658]: I0517 00:22:35.393609 2658 volume_manager.go:289] "Starting Kubelet Volume Manager" May 17 00:22:35.394042 kubelet[2658]: E0517 00:22:35.394018 2658 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4081.3.3-n-0e116cd3eb\" not found" May 17 00:22:35.396519 kubelet[2658]: I0517 00:22:35.396497 2658 desired_state_of_world_populator.go:147] "Desired state populator starts to run" May 17 00:22:35.396721 kubelet[2658]: I0517 00:22:35.396711 2658 reconciler.go:26] "Reconciler: start to sync state" May 17 00:22:35.399407 kubelet[2658]: I0517 00:22:35.399360 2658 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 17 00:22:35.401168 kubelet[2658]: I0517 00:22:35.401129 2658 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 17 00:22:35.401377 kubelet[2658]: I0517 00:22:35.401362 2658 status_manager.go:217] "Starting to sync pod status with apiserver" May 17 00:22:35.401466 kubelet[2658]: I0517 00:22:35.401458 2658 kubelet.go:2321] "Starting kubelet main sync loop" May 17 00:22:35.401628 kubelet[2658]: E0517 00:22:35.401610 2658 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 17 00:22:35.413068 kubelet[2658]: I0517 00:22:35.413042 2658 factory.go:221] Registration of the systemd container factory successfully May 17 00:22:35.413491 kubelet[2658]: I0517 00:22:35.413464 2658 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 17 00:22:35.418585 kubelet[2658]: E0517 00:22:35.418468 2658 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 17 00:22:35.419973 kubelet[2658]: I0517 00:22:35.419806 2658 factory.go:221] Registration of the containerd container factory successfully May 17 00:22:35.494413 kubelet[2658]: I0517 00:22:35.494384 2658 cpu_manager.go:214] "Starting CPU manager" policy="none" May 17 00:22:35.495249 kubelet[2658]: I0517 00:22:35.494737 2658 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 17 00:22:35.495249 kubelet[2658]: I0517 00:22:35.494769 2658 state_mem.go:36] "Initialized new in-memory state store" May 17 00:22:35.495249 kubelet[2658]: I0517 00:22:35.494947 2658 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 17 00:22:35.495249 kubelet[2658]: I0517 00:22:35.494959 2658 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 17 00:22:35.495249 kubelet[2658]: I0517 00:22:35.494981 2658 policy_none.go:49] "None policy: Start" May 17 00:22:35.496075 kubelet[2658]: I0517 00:22:35.496058 2658 memory_manager.go:170] "Starting memorymanager" policy="None" May 17 00:22:35.496185 kubelet[2658]: I0517 00:22:35.496178 2658 state_mem.go:35] "Initializing new in-memory state store" May 17 00:22:35.496568 kubelet[2658]: I0517 00:22:35.496546 2658 state_mem.go:75] "Updated machine memory state" May 17 00:22:35.500090 kubelet[2658]: I0517 00:22:35.498572 2658 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 17 00:22:35.500090 kubelet[2658]: I0517 00:22:35.498766 2658 eviction_manager.go:189] "Eviction manager: starting control loop" May 17 00:22:35.500090 kubelet[2658]: I0517 00:22:35.498777 2658 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 17 00:22:35.500090 kubelet[2658]: I0517 00:22:35.499704 2658 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 17 00:22:35.525009 kubelet[2658]: W0517 00:22:35.524858 2658 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 17 00:22:35.525715 kubelet[2658]: W0517 00:22:35.525531 2658 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 17 00:22:35.526000 kubelet[2658]: W0517 00:22:35.525940 2658 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 17 00:22:35.560116 sudo[2688]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 17 00:22:35.561677 sudo[2688]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) May 17 00:22:35.598382 kubelet[2658]: I0517 00:22:35.598058 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.600346 kubelet[2658]: I0517 00:22:35.598690 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.600346 kubelet[2658]: I0517 00:22:35.598779 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-ca-certs\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.600346 kubelet[2658]: I0517 00:22:35.598810 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.600346 kubelet[2658]: I0517 00:22:35.598872 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b237a751c3a466b82870764fcca590b0-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.3-n-0e116cd3eb\" (UID: \"b237a751c3a466b82870764fcca590b0\") " pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.606528 kubelet[2658]: I0517 00:22:35.605282 2658 kubelet_node_status.go:72] "Attempting to register node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.616894 kubelet[2658]: I0517 00:22:35.616848 2658 kubelet_node_status.go:111] "Node was previously registered" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.617068 kubelet[2658]: I0517 00:22:35.616962 2658 kubelet_node_status.go:75] "Successfully registered node" node="ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.701066 kubelet[2658]: I0517 00:22:35.699977 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/34fd926d0d45564074f8fc2aed24959b-kubeconfig\") pod \"kube-scheduler-ci-4081.3.3-n-0e116cd3eb\" (UID: \"34fd926d0d45564074f8fc2aed24959b\") " pod="kube-system/kube-scheduler-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.701066 kubelet[2658]: I0517 00:22:35.700071 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/379bfc971db3b4ae016ab6b991edb0cc-ca-certs\") pod \"kube-apiserver-ci-4081.3.3-n-0e116cd3eb\" (UID: \"379bfc971db3b4ae016ab6b991edb0cc\") " pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.701066 kubelet[2658]: I0517 00:22:35.700097 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/379bfc971db3b4ae016ab6b991edb0cc-k8s-certs\") pod \"kube-apiserver-ci-4081.3.3-n-0e116cd3eb\" (UID: \"379bfc971db3b4ae016ab6b991edb0cc\") " pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.701066 kubelet[2658]: I0517 00:22:35.700126 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/379bfc971db3b4ae016ab6b991edb0cc-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.3-n-0e116cd3eb\" (UID: \"379bfc971db3b4ae016ab6b991edb0cc\") " pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:35.826512 kubelet[2658]: E0517 00:22:35.826458 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:35.826832 kubelet[2658]: E0517 00:22:35.826689 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:35.829599 kubelet[2658]: E0517 00:22:35.829422 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:36.371729 kubelet[2658]: I0517 00:22:36.370868 2658 apiserver.go:52] "Watching apiserver" May 17 00:22:36.398607 kubelet[2658]: I0517 00:22:36.398500 2658 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" May 17 00:22:36.448610 sudo[2688]: pam_unix(sudo:session): session closed for user root May 17 00:22:36.457151 kubelet[2658]: E0517 00:22:36.456088 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:36.459437 kubelet[2658]: E0517 00:22:36.458824 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:36.469350 kubelet[2658]: W0517 00:22:36.466390 2658 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 17 00:22:36.469350 kubelet[2658]: E0517 00:22:36.466484 2658 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081.3.3-n-0e116cd3eb\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" May 17 00:22:36.469350 kubelet[2658]: E0517 00:22:36.466746 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:36.508727 kubelet[2658]: I0517 00:22:36.508648 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.3-n-0e116cd3eb" podStartSLOduration=1.5086086239999998 podStartE2EDuration="1.508608624s" podCreationTimestamp="2025-05-17 00:22:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:22:36.496433939 +0000 UTC m=+1.237649341" watchObservedRunningTime="2025-05-17 00:22:36.508608624 +0000 UTC m=+1.249824037" May 17 00:22:36.522374 kubelet[2658]: I0517 00:22:36.521040 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.3-n-0e116cd3eb" podStartSLOduration=1.521014442 podStartE2EDuration="1.521014442s" podCreationTimestamp="2025-05-17 00:22:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:22:36.509451977 +0000 UTC m=+1.250667384" watchObservedRunningTime="2025-05-17 00:22:36.521014442 +0000 UTC m=+1.262229854" May 17 00:22:37.458193 kubelet[2658]: E0517 00:22:37.457834 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:38.234265 sudo[1797]: pam_unix(sudo:session): session closed for user root May 17 00:22:38.240232 sshd[1791]: pam_unix(sshd:session): session closed for user core May 17 00:22:38.246238 systemd[1]: sshd@6-64.23.167.227:22-139.178.68.195:59616.service: Deactivated successfully. May 17 00:22:38.252930 systemd[1]: session-7.scope: Deactivated successfully. May 17 00:22:38.254380 systemd-logind[1571]: Session 7 logged out. Waiting for processes to exit. May 17 00:22:38.256271 systemd-logind[1571]: Removed session 7. May 17 00:22:40.296973 kubelet[2658]: I0517 00:22:40.296830 2658 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 17 00:22:40.297656 containerd[1595]: time="2025-05-17T00:22:40.297618043Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 17 00:22:40.298773 kubelet[2658]: I0517 00:22:40.298241 2658 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 17 00:22:40.467590 kubelet[2658]: E0517 00:22:40.467550 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:40.494810 kubelet[2658]: I0517 00:22:40.494656 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.3-n-0e116cd3eb" podStartSLOduration=5.494630239 podStartE2EDuration="5.494630239s" podCreationTimestamp="2025-05-17 00:22:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:22:36.521458299 +0000 UTC m=+1.262673706" watchObservedRunningTime="2025-05-17 00:22:40.494630239 +0000 UTC m=+5.235845649" May 17 00:22:41.322283 kubelet[2658]: W0517 00:22:41.322241 2658 reflector.go:561] object-"kube-system"/"cilium-config": failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:ci-4081.3.3-n-0e116cd3eb" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4081.3.3-n-0e116cd3eb' and this object May 17 00:22:41.324578 kubelet[2658]: E0517 00:22:41.324401 2658 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cilium-config\" is forbidden: User \"system:node:ci-4081.3.3-n-0e116cd3eb\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4081.3.3-n-0e116cd3eb' and this object" logger="UnhandledError" May 17 00:22:41.324578 kubelet[2658]: W0517 00:22:41.322840 2658 reflector.go:561] object-"kube-system"/"cilium-clustermesh": failed to list *v1.Secret: secrets "cilium-clustermesh" is forbidden: User "system:node:ci-4081.3.3-n-0e116cd3eb" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4081.3.3-n-0e116cd3eb' and this object May 17 00:22:41.324578 kubelet[2658]: E0517 00:22:41.324505 2658 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-clustermesh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-clustermesh\" is forbidden: User \"system:node:ci-4081.3.3-n-0e116cd3eb\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4081.3.3-n-0e116cd3eb' and this object" logger="UnhandledError" May 17 00:22:41.324578 kubelet[2658]: W0517 00:22:41.322889 2658 reflector.go:561] object-"kube-system"/"hubble-server-certs": failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ci-4081.3.3-n-0e116cd3eb" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4081.3.3-n-0e116cd3eb' and this object May 17 00:22:41.324927 kubelet[2658]: E0517 00:22:41.324532 2658 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"hubble-server-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"hubble-server-certs\" is forbidden: User \"system:node:ci-4081.3.3-n-0e116cd3eb\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4081.3.3-n-0e116cd3eb' and this object" logger="UnhandledError" May 17 00:22:41.350603 kubelet[2658]: I0517 00:22:41.350537 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-etc-cni-netd\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.350859 kubelet[2658]: I0517 00:22:41.350737 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-net\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.350859 kubelet[2658]: I0517 00:22:41.350819 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-cgroup\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.350859 kubelet[2658]: I0517 00:22:41.350842 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cni-path\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351073 kubelet[2658]: I0517 00:22:41.350901 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-xtables-lock\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351073 kubelet[2658]: I0517 00:22:41.350940 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/41055cf1-6ad8-42df-b307-0d050127e3d0-xtables-lock\") pod \"kube-proxy-m8ktq\" (UID: \"41055cf1-6ad8-42df-b307-0d050127e3d0\") " pod="kube-system/kube-proxy-m8ktq" May 17 00:22:41.351073 kubelet[2658]: I0517 00:22:41.350968 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-clustermesh-secrets\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351073 kubelet[2658]: I0517 00:22:41.350988 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-config-path\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351073 kubelet[2658]: I0517 00:22:41.351026 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hubble-tls\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351073 kubelet[2658]: I0517 00:22:41.351041 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/41055cf1-6ad8-42df-b307-0d050127e3d0-kube-proxy\") pod \"kube-proxy-m8ktq\" (UID: \"41055cf1-6ad8-42df-b307-0d050127e3d0\") " pod="kube-system/kube-proxy-m8ktq" May 17 00:22:41.351305 kubelet[2658]: I0517 00:22:41.351056 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-kernel\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351305 kubelet[2658]: I0517 00:22:41.351072 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/41055cf1-6ad8-42df-b307-0d050127e3d0-lib-modules\") pod \"kube-proxy-m8ktq\" (UID: \"41055cf1-6ad8-42df-b307-0d050127e3d0\") " pod="kube-system/kube-proxy-m8ktq" May 17 00:22:41.351305 kubelet[2658]: I0517 00:22:41.351093 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-lib-modules\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351305 kubelet[2658]: I0517 00:22:41.351110 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjn9t\" (UniqueName: \"kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-kube-api-access-rjn9t\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351305 kubelet[2658]: I0517 00:22:41.351127 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-bpf-maps\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351546 kubelet[2658]: I0517 00:22:41.351144 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mm22\" (UniqueName: \"kubernetes.io/projected/41055cf1-6ad8-42df-b307-0d050127e3d0-kube-api-access-7mm22\") pod \"kube-proxy-m8ktq\" (UID: \"41055cf1-6ad8-42df-b307-0d050127e3d0\") " pod="kube-system/kube-proxy-m8ktq" May 17 00:22:41.351546 kubelet[2658]: I0517 00:22:41.351159 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-run\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.351546 kubelet[2658]: I0517 00:22:41.351179 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hostproc\") pod \"cilium-nb77r\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " pod="kube-system/cilium-nb77r" May 17 00:22:41.472009 kubelet[2658]: E0517 00:22:41.471928 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:41.552535 kubelet[2658]: I0517 00:22:41.552475 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a77f0e64-ad05-4614-bce9-cfe0daa984dc-cilium-config-path\") pod \"cilium-operator-5d85765b45-nhgxg\" (UID: \"a77f0e64-ad05-4614-bce9-cfe0daa984dc\") " pod="kube-system/cilium-operator-5d85765b45-nhgxg" May 17 00:22:41.552535 kubelet[2658]: I0517 00:22:41.552540 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhj9t\" (UniqueName: \"kubernetes.io/projected/a77f0e64-ad05-4614-bce9-cfe0daa984dc-kube-api-access-mhj9t\") pod \"cilium-operator-5d85765b45-nhgxg\" (UID: \"a77f0e64-ad05-4614-bce9-cfe0daa984dc\") " pod="kube-system/cilium-operator-5d85765b45-nhgxg" May 17 00:22:41.614043 kubelet[2658]: E0517 00:22:41.613877 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:41.616823 containerd[1595]: time="2025-05-17T00:22:41.615966227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-m8ktq,Uid:41055cf1-6ad8-42df-b307-0d050127e3d0,Namespace:kube-system,Attempt:0,}" May 17 00:22:41.654005 containerd[1595]: time="2025-05-17T00:22:41.653467336Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:22:41.654005 containerd[1595]: time="2025-05-17T00:22:41.653558832Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:22:41.654005 containerd[1595]: time="2025-05-17T00:22:41.653571587Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:41.654005 containerd[1595]: time="2025-05-17T00:22:41.653681541Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:41.716958 containerd[1595]: time="2025-05-17T00:22:41.716905714Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-m8ktq,Uid:41055cf1-6ad8-42df-b307-0d050127e3d0,Namespace:kube-system,Attempt:0,} returns sandbox id \"7334a87d09225e70f77563737ad8a467890238f13f1ac120d61a860f009871fe\"" May 17 00:22:41.718238 kubelet[2658]: E0517 00:22:41.718203 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:41.724181 containerd[1595]: time="2025-05-17T00:22:41.723873205Z" level=info msg="CreateContainer within sandbox \"7334a87d09225e70f77563737ad8a467890238f13f1ac120d61a860f009871fe\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 17 00:22:41.741433 containerd[1595]: time="2025-05-17T00:22:41.741355712Z" level=info msg="CreateContainer within sandbox \"7334a87d09225e70f77563737ad8a467890238f13f1ac120d61a860f009871fe\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"40c23af4b545d637bfb2ab3e52413a26203003175cc939f99d9158637b5b58ed\"" May 17 00:22:41.743467 containerd[1595]: time="2025-05-17T00:22:41.742826773Z" level=info msg="StartContainer for \"40c23af4b545d637bfb2ab3e52413a26203003175cc939f99d9158637b5b58ed\"" May 17 00:22:41.831269 containerd[1595]: time="2025-05-17T00:22:41.831215867Z" level=info msg="StartContainer for \"40c23af4b545d637bfb2ab3e52413a26203003175cc939f99d9158637b5b58ed\" returns successfully" May 17 00:22:42.459252 kubelet[2658]: E0517 00:22:42.458751 2658 projected.go:263] Couldn't get secret kube-system/hubble-server-certs: failed to sync secret cache: timed out waiting for the condition May 17 00:22:42.459252 kubelet[2658]: E0517 00:22:42.458807 2658 projected.go:194] Error preparing data for projected volume hubble-tls for pod kube-system/cilium-nb77r: failed to sync secret cache: timed out waiting for the condition May 17 00:22:42.459252 kubelet[2658]: E0517 00:22:42.458911 2658 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hubble-tls podName:e02a6237-4e09-4ae1-82f5-f96d9fba9fd1 nodeName:}" failed. No retries permitted until 2025-05-17 00:22:42.958885382 +0000 UTC m=+7.700100781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "hubble-tls" (UniqueName: "kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hubble-tls") pod "cilium-nb77r" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1") : failed to sync secret cache: timed out waiting for the condition May 17 00:22:42.459252 kubelet[2658]: E0517 00:22:42.459017 2658 secret.go:189] Couldn't get secret kube-system/cilium-clustermesh: failed to sync secret cache: timed out waiting for the condition May 17 00:22:42.459252 kubelet[2658]: E0517 00:22:42.459152 2658 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-clustermesh-secrets podName:e02a6237-4e09-4ae1-82f5-f96d9fba9fd1 nodeName:}" failed. No retries permitted until 2025-05-17 00:22:42.959121914 +0000 UTC m=+7.700337324 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "clustermesh-secrets" (UniqueName: "kubernetes.io/secret/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-clustermesh-secrets") pod "cilium-nb77r" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1") : failed to sync secret cache: timed out waiting for the condition May 17 00:22:42.459252 kubelet[2658]: E0517 00:22:42.459179 2658 configmap.go:193] Couldn't get configMap kube-system/cilium-config: failed to sync configmap cache: timed out waiting for the condition May 17 00:22:42.460040 kubelet[2658]: E0517 00:22:42.459227 2658 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-config-path podName:e02a6237-4e09-4ae1-82f5-f96d9fba9fd1 nodeName:}" failed. No retries permitted until 2025-05-17 00:22:42.959216332 +0000 UTC m=+7.700431720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-config-path" (UniqueName: "kubernetes.io/configmap/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-config-path") pod "cilium-nb77r" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1") : failed to sync configmap cache: timed out waiting for the condition May 17 00:22:42.475828 kubelet[2658]: E0517 00:22:42.475094 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:42.654744 kubelet[2658]: E0517 00:22:42.654609 2658 configmap.go:193] Couldn't get configMap kube-system/cilium-config: failed to sync configmap cache: timed out waiting for the condition May 17 00:22:42.654744 kubelet[2658]: E0517 00:22:42.654722 2658 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a77f0e64-ad05-4614-bce9-cfe0daa984dc-cilium-config-path podName:a77f0e64-ad05-4614-bce9-cfe0daa984dc nodeName:}" failed. No retries permitted until 2025-05-17 00:22:43.154699091 +0000 UTC m=+7.895914480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-config-path" (UniqueName: "kubernetes.io/configmap/a77f0e64-ad05-4614-bce9-cfe0daa984dc-cilium-config-path") pod "cilium-operator-5d85765b45-nhgxg" (UID: "a77f0e64-ad05-4614-bce9-cfe0daa984dc") : failed to sync configmap cache: timed out waiting for the condition May 17 00:22:42.980909 kubelet[2658]: E0517 00:22:42.980439 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:43.002078 kubelet[2658]: I0517 00:22:43.001809 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-m8ktq" podStartSLOduration=2.001749273 podStartE2EDuration="2.001749273s" podCreationTimestamp="2025-05-17 00:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:22:42.489115145 +0000 UTC m=+7.230330554" watchObservedRunningTime="2025-05-17 00:22:43.001749273 +0000 UTC m=+7.742964682" May 17 00:22:43.122539 kubelet[2658]: E0517 00:22:43.122476 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:43.123598 containerd[1595]: time="2025-05-17T00:22:43.123359908Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nb77r,Uid:e02a6237-4e09-4ae1-82f5-f96d9fba9fd1,Namespace:kube-system,Attempt:0,}" May 17 00:22:43.156492 kubelet[2658]: E0517 00:22:43.154867 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:43.178465 containerd[1595]: time="2025-05-17T00:22:43.176816711Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:22:43.178465 containerd[1595]: time="2025-05-17T00:22:43.176972588Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:22:43.178465 containerd[1595]: time="2025-05-17T00:22:43.176989862Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:43.178465 containerd[1595]: time="2025-05-17T00:22:43.177157145Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:43.249637 containerd[1595]: time="2025-05-17T00:22:43.249483240Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nb77r,Uid:e02a6237-4e09-4ae1-82f5-f96d9fba9fd1,Namespace:kube-system,Attempt:0,} returns sandbox id \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\"" May 17 00:22:43.251410 kubelet[2658]: E0517 00:22:43.251173 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:43.255191 containerd[1595]: time="2025-05-17T00:22:43.255139745Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 17 00:22:43.278568 kubelet[2658]: E0517 00:22:43.278511 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:43.280009 containerd[1595]: time="2025-05-17T00:22:43.279526575Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-nhgxg,Uid:a77f0e64-ad05-4614-bce9-cfe0daa984dc,Namespace:kube-system,Attempt:0,}" May 17 00:22:43.308156 containerd[1595]: time="2025-05-17T00:22:43.308021618Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:22:43.308156 containerd[1595]: time="2025-05-17T00:22:43.308083199Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:22:43.308156 containerd[1595]: time="2025-05-17T00:22:43.308094197Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:43.308742 containerd[1595]: time="2025-05-17T00:22:43.308586488Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:22:43.392347 containerd[1595]: time="2025-05-17T00:22:43.392247670Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-nhgxg,Uid:a77f0e64-ad05-4614-bce9-cfe0daa984dc,Namespace:kube-system,Attempt:0,} returns sandbox id \"c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012\"" May 17 00:22:43.394044 kubelet[2658]: E0517 00:22:43.393994 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:43.480359 kubelet[2658]: E0517 00:22:43.479733 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:43.480359 kubelet[2658]: E0517 00:22:43.479755 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:48.370186 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2588220855.mount: Deactivated successfully. May 17 00:22:50.966230 containerd[1595]: time="2025-05-17T00:22:50.966151557Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:50.968182 containerd[1595]: time="2025-05-17T00:22:50.968088603Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" May 17 00:22:50.969225 containerd[1595]: time="2025-05-17T00:22:50.968913911Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:50.972881 containerd[1595]: time="2025-05-17T00:22:50.972807450Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 7.71727931s" May 17 00:22:50.972881 containerd[1595]: time="2025-05-17T00:22:50.972874862Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" May 17 00:22:50.976128 containerd[1595]: time="2025-05-17T00:22:50.975656986Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 17 00:22:50.993744 containerd[1595]: time="2025-05-17T00:22:50.993639538Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:22:51.096079 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount249754202.mount: Deactivated successfully. May 17 00:22:51.124383 containerd[1595]: time="2025-05-17T00:22:51.124135408Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e\"" May 17 00:22:51.126292 containerd[1595]: time="2025-05-17T00:22:51.125641538Z" level=info msg="StartContainer for \"bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e\"" May 17 00:22:51.257915 containerd[1595]: time="2025-05-17T00:22:51.257269589Z" level=info msg="StartContainer for \"bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e\" returns successfully" May 17 00:22:51.440699 containerd[1595]: time="2025-05-17T00:22:51.410164898Z" level=info msg="shim disconnected" id=bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e namespace=k8s.io May 17 00:22:51.440699 containerd[1595]: time="2025-05-17T00:22:51.440430907Z" level=warning msg="cleaning up after shim disconnected" id=bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e namespace=k8s.io May 17 00:22:51.440699 containerd[1595]: time="2025-05-17T00:22:51.440459067Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:22:51.545407 kubelet[2658]: E0517 00:22:51.544920 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:51.552500 containerd[1595]: time="2025-05-17T00:22:51.551845298Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 17 00:22:51.573625 containerd[1595]: time="2025-05-17T00:22:51.573427457Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0\"" May 17 00:22:51.574902 containerd[1595]: time="2025-05-17T00:22:51.574829193Z" level=info msg="StartContainer for \"977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0\"" May 17 00:22:51.655970 containerd[1595]: time="2025-05-17T00:22:51.654537928Z" level=info msg="StartContainer for \"977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0\" returns successfully" May 17 00:22:51.670718 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 17 00:22:51.671098 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 17 00:22:51.671202 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... May 17 00:22:51.684511 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 17 00:22:51.715968 containerd[1595]: time="2025-05-17T00:22:51.715839544Z" level=info msg="shim disconnected" id=977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0 namespace=k8s.io May 17 00:22:51.716707 containerd[1595]: time="2025-05-17T00:22:51.716056694Z" level=warning msg="cleaning up after shim disconnected" id=977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0 namespace=k8s.io May 17 00:22:51.716707 containerd[1595]: time="2025-05-17T00:22:51.716075983Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:22:51.730115 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 17 00:22:51.743416 containerd[1595]: time="2025-05-17T00:22:51.742644097Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:22:51Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io May 17 00:22:52.090174 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e-rootfs.mount: Deactivated successfully. May 17 00:22:52.574867 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4306257.mount: Deactivated successfully. May 17 00:22:52.580395 kubelet[2658]: E0517 00:22:52.578871 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:52.601164 containerd[1595]: time="2025-05-17T00:22:52.594598520Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 17 00:22:52.668461 containerd[1595]: time="2025-05-17T00:22:52.667553609Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491\"" May 17 00:22:52.671602 containerd[1595]: time="2025-05-17T00:22:52.671365741Z" level=info msg="StartContainer for \"aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491\"" May 17 00:22:52.791034 containerd[1595]: time="2025-05-17T00:22:52.790979663Z" level=info msg="StartContainer for \"aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491\" returns successfully" May 17 00:22:52.841451 containerd[1595]: time="2025-05-17T00:22:52.841113922Z" level=info msg="shim disconnected" id=aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491 namespace=k8s.io May 17 00:22:52.841451 containerd[1595]: time="2025-05-17T00:22:52.841192972Z" level=warning msg="cleaning up after shim disconnected" id=aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491 namespace=k8s.io May 17 00:22:52.841451 containerd[1595]: time="2025-05-17T00:22:52.841207613Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:22:53.304482 containerd[1595]: time="2025-05-17T00:22:53.304428774Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:53.305626 containerd[1595]: time="2025-05-17T00:22:53.305276431Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" May 17 00:22:53.305626 containerd[1595]: time="2025-05-17T00:22:53.305574305Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 17 00:22:53.310234 containerd[1595]: time="2025-05-17T00:22:53.310153114Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.334435068s" May 17 00:22:53.310234 containerd[1595]: time="2025-05-17T00:22:53.310206267Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" May 17 00:22:53.313861 containerd[1595]: time="2025-05-17T00:22:53.313812894Z" level=info msg="CreateContainer within sandbox \"c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 17 00:22:53.338345 containerd[1595]: time="2025-05-17T00:22:53.338172111Z" level=info msg="CreateContainer within sandbox \"c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\"" May 17 00:22:53.339281 containerd[1595]: time="2025-05-17T00:22:53.339091160Z" level=info msg="StartContainer for \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\"" May 17 00:22:53.413264 containerd[1595]: time="2025-05-17T00:22:53.413219684Z" level=info msg="StartContainer for \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\" returns successfully" May 17 00:22:53.588915 kubelet[2658]: E0517 00:22:53.587503 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:53.595711 kubelet[2658]: E0517 00:22:53.595682 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:53.597670 containerd[1595]: time="2025-05-17T00:22:53.596653884Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 17 00:22:53.618659 update_engine[1573]: I20250517 00:22:53.618508 1573 update_attempter.cc:509] Updating boot flags... May 17 00:22:53.628342 containerd[1595]: time="2025-05-17T00:22:53.627673400Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1\"" May 17 00:22:53.636595 containerd[1595]: time="2025-05-17T00:22:53.632545917Z" level=info msg="StartContainer for \"f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1\"" May 17 00:22:53.788355 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (3302) May 17 00:22:53.973357 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (3301) May 17 00:22:53.973520 containerd[1595]: time="2025-05-17T00:22:53.973443204Z" level=info msg="StartContainer for \"f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1\" returns successfully" May 17 00:22:54.088239 systemd[1]: run-containerd-runc-k8s.io-1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126-runc.pWV32Z.mount: Deactivated successfully. May 17 00:22:54.104766 containerd[1595]: time="2025-05-17T00:22:54.104635818Z" level=info msg="shim disconnected" id=f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1 namespace=k8s.io May 17 00:22:54.107353 containerd[1595]: time="2025-05-17T00:22:54.105042658Z" level=warning msg="cleaning up after shim disconnected" id=f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1 namespace=k8s.io May 17 00:22:54.107353 containerd[1595]: time="2025-05-17T00:22:54.105074055Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:22:54.701453 kubelet[2658]: E0517 00:22:54.699225 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:54.705351 kubelet[2658]: E0517 00:22:54.702751 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:54.714957 containerd[1595]: time="2025-05-17T00:22:54.713712613Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 17 00:22:54.747764 containerd[1595]: time="2025-05-17T00:22:54.747699603Z" level=info msg="CreateContainer within sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\"" May 17 00:22:54.750810 containerd[1595]: time="2025-05-17T00:22:54.750767065Z" level=info msg="StartContainer for \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\"" May 17 00:22:54.783339 kubelet[2658]: I0517 00:22:54.780930 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-nhgxg" podStartSLOduration=3.865276299 podStartE2EDuration="13.780896394s" podCreationTimestamp="2025-05-17 00:22:41 +0000 UTC" firstStartedPulling="2025-05-17 00:22:43.395384714 +0000 UTC m=+8.136600102" lastFinishedPulling="2025-05-17 00:22:53.311004796 +0000 UTC m=+18.052220197" observedRunningTime="2025-05-17 00:22:53.688887826 +0000 UTC m=+18.430103234" watchObservedRunningTime="2025-05-17 00:22:54.780896394 +0000 UTC m=+19.522111795" May 17 00:22:54.893272 containerd[1595]: time="2025-05-17T00:22:54.893223981Z" level=info msg="StartContainer for \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\" returns successfully" May 17 00:22:55.085040 kubelet[2658]: I0517 00:22:55.084986 2658 kubelet_node_status.go:488] "Fast updating node status as it just became ready" May 17 00:22:55.255400 kubelet[2658]: I0517 00:22:55.255102 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/91be9196-c9bb-4279-aa81-b029c58243b3-config-volume\") pod \"coredns-7c65d6cfc9-jsd6t\" (UID: \"91be9196-c9bb-4279-aa81-b029c58243b3\") " pod="kube-system/coredns-7c65d6cfc9-jsd6t" May 17 00:22:55.255400 kubelet[2658]: I0517 00:22:55.255180 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mncxw\" (UniqueName: \"kubernetes.io/projected/91be9196-c9bb-4279-aa81-b029c58243b3-kube-api-access-mncxw\") pod \"coredns-7c65d6cfc9-jsd6t\" (UID: \"91be9196-c9bb-4279-aa81-b029c58243b3\") " pod="kube-system/coredns-7c65d6cfc9-jsd6t" May 17 00:22:55.255400 kubelet[2658]: I0517 00:22:55.255254 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/428e1647-90ec-40ee-98d3-7d690a2cd390-config-volume\") pod \"coredns-7c65d6cfc9-vjgzr\" (UID: \"428e1647-90ec-40ee-98d3-7d690a2cd390\") " pod="kube-system/coredns-7c65d6cfc9-vjgzr" May 17 00:22:55.256435 kubelet[2658]: I0517 00:22:55.256396 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkhxc\" (UniqueName: \"kubernetes.io/projected/428e1647-90ec-40ee-98d3-7d690a2cd390-kube-api-access-nkhxc\") pod \"coredns-7c65d6cfc9-vjgzr\" (UID: \"428e1647-90ec-40ee-98d3-7d690a2cd390\") " pod="kube-system/coredns-7c65d6cfc9-vjgzr" May 17 00:22:55.436584 kubelet[2658]: E0517 00:22:55.433806 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:55.436752 containerd[1595]: time="2025-05-17T00:22:55.435011281Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-jsd6t,Uid:91be9196-c9bb-4279-aa81-b029c58243b3,Namespace:kube-system,Attempt:0,}" May 17 00:22:55.458206 kubelet[2658]: E0517 00:22:55.455950 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:55.458511 containerd[1595]: time="2025-05-17T00:22:55.456620084Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-vjgzr,Uid:428e1647-90ec-40ee-98d3-7d690a2cd390,Namespace:kube-system,Attempt:0,}" May 17 00:22:55.706727 kubelet[2658]: E0517 00:22:55.705650 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:55.729134 kubelet[2658]: I0517 00:22:55.729036 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-nb77r" podStartSLOduration=7.005667718 podStartE2EDuration="14.728992529s" podCreationTimestamp="2025-05-17 00:22:41 +0000 UTC" firstStartedPulling="2025-05-17 00:22:43.2519006 +0000 UTC m=+7.993115996" lastFinishedPulling="2025-05-17 00:22:50.975225405 +0000 UTC m=+15.716440807" observedRunningTime="2025-05-17 00:22:55.725780003 +0000 UTC m=+20.466995410" watchObservedRunningTime="2025-05-17 00:22:55.728992529 +0000 UTC m=+20.470207960" May 17 00:22:56.708414 kubelet[2658]: E0517 00:22:56.708076 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:57.209292 systemd-networkd[1219]: cilium_host: Link UP May 17 00:22:57.211662 systemd-networkd[1219]: cilium_net: Link UP May 17 00:22:57.211669 systemd-networkd[1219]: cilium_net: Gained carrier May 17 00:22:57.213008 systemd-networkd[1219]: cilium_host: Gained carrier May 17 00:22:57.366559 systemd-networkd[1219]: cilium_vxlan: Link UP May 17 00:22:57.366571 systemd-networkd[1219]: cilium_vxlan: Gained carrier May 17 00:22:57.503679 systemd-networkd[1219]: cilium_host: Gained IPv6LL May 17 00:22:57.599648 systemd-networkd[1219]: cilium_net: Gained IPv6LL May 17 00:22:57.711341 kubelet[2658]: E0517 00:22:57.709655 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:57.745458 kernel: NET: Registered PF_ALG protocol family May 17 00:22:58.520416 systemd-networkd[1219]: cilium_vxlan: Gained IPv6LL May 17 00:22:58.584474 systemd-networkd[1219]: lxc_health: Link UP May 17 00:22:58.595537 systemd-networkd[1219]: lxc_health: Gained carrier May 17 00:22:59.067251 systemd-networkd[1219]: lxcb3b364de493c: Link UP May 17 00:22:59.072369 kernel: eth0: renamed from tmp1edb6 May 17 00:22:59.076929 systemd-networkd[1219]: lxcb3b364de493c: Gained carrier May 17 00:22:59.126349 kubelet[2658]: E0517 00:22:59.125622 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:22:59.179739 kernel: eth0: renamed from tmpf2dfa May 17 00:22:59.183894 systemd-networkd[1219]: lxc1086cfd2b5d1: Link UP May 17 00:22:59.191900 systemd-networkd[1219]: lxc1086cfd2b5d1: Gained carrier May 17 00:22:59.719918 kubelet[2658]: E0517 00:22:59.719827 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:00.055600 systemd-networkd[1219]: lxc_health: Gained IPv6LL May 17 00:23:00.248425 systemd-networkd[1219]: lxc1086cfd2b5d1: Gained IPv6LL May 17 00:23:00.695570 systemd-networkd[1219]: lxcb3b364de493c: Gained IPv6LL May 17 00:23:00.720851 kubelet[2658]: E0517 00:23:00.720797 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:04.563345 containerd[1595]: time="2025-05-17T00:23:04.545284868Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:23:04.563345 containerd[1595]: time="2025-05-17T00:23:04.553735349Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:23:04.563345 containerd[1595]: time="2025-05-17T00:23:04.553751884Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:23:04.563345 containerd[1595]: time="2025-05-17T00:23:04.553883249Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:23:04.569872 containerd[1595]: time="2025-05-17T00:23:04.565542664Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:23:04.569872 containerd[1595]: time="2025-05-17T00:23:04.568537621Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:23:04.569872 containerd[1595]: time="2025-05-17T00:23:04.568673332Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:23:04.569872 containerd[1595]: time="2025-05-17T00:23:04.569183559Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:23:04.690401 containerd[1595]: time="2025-05-17T00:23:04.689704023Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-jsd6t,Uid:91be9196-c9bb-4279-aa81-b029c58243b3,Namespace:kube-system,Attempt:0,} returns sandbox id \"1edb63b66e9c973f02af02b00ebe295ab478e8d186b9062498532ddab4cbe8f0\"" May 17 00:23:04.695366 kubelet[2658]: E0517 00:23:04.694633 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:04.700525 containerd[1595]: time="2025-05-17T00:23:04.700193070Z" level=info msg="CreateContainer within sandbox \"1edb63b66e9c973f02af02b00ebe295ab478e8d186b9062498532ddab4cbe8f0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 17 00:23:04.718477 containerd[1595]: time="2025-05-17T00:23:04.718232889Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-vjgzr,Uid:428e1647-90ec-40ee-98d3-7d690a2cd390,Namespace:kube-system,Attempt:0,} returns sandbox id \"f2dfa82549384625052aff6a09c4b01d6e2c3ea8f22bfc4497ff10a3bcccd34a\"" May 17 00:23:04.719451 kubelet[2658]: E0517 00:23:04.719425 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:04.725237 containerd[1595]: time="2025-05-17T00:23:04.725194812Z" level=info msg="CreateContainer within sandbox \"f2dfa82549384625052aff6a09c4b01d6e2c3ea8f22bfc4497ff10a3bcccd34a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 17 00:23:04.755635 containerd[1595]: time="2025-05-17T00:23:04.755566387Z" level=info msg="CreateContainer within sandbox \"1edb63b66e9c973f02af02b00ebe295ab478e8d186b9062498532ddab4cbe8f0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"58bcb6d4abaf897efcb7627988412b33692d2e7518f948afec3caf5ee85693ca\"" May 17 00:23:04.756056 containerd[1595]: time="2025-05-17T00:23:04.755862782Z" level=info msg="CreateContainer within sandbox \"f2dfa82549384625052aff6a09c4b01d6e2c3ea8f22bfc4497ff10a3bcccd34a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"5111e0135958ca69c4d49779857d96b2a191e2aee21658fe0a7a9008a842eff0\"" May 17 00:23:04.757234 containerd[1595]: time="2025-05-17T00:23:04.757196473Z" level=info msg="StartContainer for \"58bcb6d4abaf897efcb7627988412b33692d2e7518f948afec3caf5ee85693ca\"" May 17 00:23:04.757791 containerd[1595]: time="2025-05-17T00:23:04.757691431Z" level=info msg="StartContainer for \"5111e0135958ca69c4d49779857d96b2a191e2aee21658fe0a7a9008a842eff0\"" May 17 00:23:04.845864 containerd[1595]: time="2025-05-17T00:23:04.845503991Z" level=info msg="StartContainer for \"5111e0135958ca69c4d49779857d96b2a191e2aee21658fe0a7a9008a842eff0\" returns successfully" May 17 00:23:04.847580 containerd[1595]: time="2025-05-17T00:23:04.846772888Z" level=info msg="StartContainer for \"58bcb6d4abaf897efcb7627988412b33692d2e7518f948afec3caf5ee85693ca\" returns successfully" May 17 00:23:05.564188 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4209503428.mount: Deactivated successfully. May 17 00:23:05.772401 kubelet[2658]: E0517 00:23:05.772360 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:05.782139 kubelet[2658]: E0517 00:23:05.780554 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:05.791184 kubelet[2658]: I0517 00:23:05.791031 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-vjgzr" podStartSLOduration=24.790957314 podStartE2EDuration="24.790957314s" podCreationTimestamp="2025-05-17 00:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:23:05.787654138 +0000 UTC m=+30.528869565" watchObservedRunningTime="2025-05-17 00:23:05.790957314 +0000 UTC m=+30.532172725" May 17 00:23:05.805785 kubelet[2658]: I0517 00:23:05.803689 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-jsd6t" podStartSLOduration=24.803663752 podStartE2EDuration="24.803663752s" podCreationTimestamp="2025-05-17 00:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:23:05.803340307 +0000 UTC m=+30.544555712" watchObservedRunningTime="2025-05-17 00:23:05.803663752 +0000 UTC m=+30.544879162" May 17 00:23:06.779712 kubelet[2658]: E0517 00:23:06.779418 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:06.782194 kubelet[2658]: E0517 00:23:06.781479 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:07.782017 kubelet[2658]: E0517 00:23:07.781464 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:07.782017 kubelet[2658]: E0517 00:23:07.781842 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:08.784735 kubelet[2658]: E0517 00:23:08.784624 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:14.657719 systemd[1]: Started sshd@7-64.23.167.227:22-139.178.68.195:50602.service - OpenSSH per-connection server daemon (139.178.68.195:50602). May 17 00:23:14.716268 sshd[4040]: Accepted publickey for core from 139.178.68.195 port 50602 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:14.719445 sshd[4040]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:14.725974 systemd-logind[1571]: New session 8 of user core. May 17 00:23:14.733822 systemd[1]: Started session-8.scope - Session 8 of User core. May 17 00:23:15.319653 sshd[4040]: pam_unix(sshd:session): session closed for user core May 17 00:23:15.324364 systemd[1]: sshd@7-64.23.167.227:22-139.178.68.195:50602.service: Deactivated successfully. May 17 00:23:15.327598 systemd-logind[1571]: Session 8 logged out. Waiting for processes to exit. May 17 00:23:15.327949 systemd[1]: session-8.scope: Deactivated successfully. May 17 00:23:15.330005 systemd-logind[1571]: Removed session 8. May 17 00:23:20.330735 systemd[1]: Started sshd@8-64.23.167.227:22-139.178.68.195:50616.service - OpenSSH per-connection server daemon (139.178.68.195:50616). May 17 00:23:20.376597 sshd[4055]: Accepted publickey for core from 139.178.68.195 port 50616 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:20.378765 sshd[4055]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:20.384796 systemd-logind[1571]: New session 9 of user core. May 17 00:23:20.391811 systemd[1]: Started session-9.scope - Session 9 of User core. May 17 00:23:20.547681 sshd[4055]: pam_unix(sshd:session): session closed for user core May 17 00:23:20.552106 systemd[1]: sshd@8-64.23.167.227:22-139.178.68.195:50616.service: Deactivated successfully. May 17 00:23:20.556302 systemd[1]: session-9.scope: Deactivated successfully. May 17 00:23:20.557406 systemd-logind[1571]: Session 9 logged out. Waiting for processes to exit. May 17 00:23:20.558897 systemd-logind[1571]: Removed session 9. May 17 00:23:25.560795 systemd[1]: Started sshd@9-64.23.167.227:22-139.178.68.195:46106.service - OpenSSH per-connection server daemon (139.178.68.195:46106). May 17 00:23:25.601159 sshd[4070]: Accepted publickey for core from 139.178.68.195 port 46106 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:25.603065 sshd[4070]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:25.608996 systemd-logind[1571]: New session 10 of user core. May 17 00:23:25.616026 systemd[1]: Started session-10.scope - Session 10 of User core. May 17 00:23:25.785784 sshd[4070]: pam_unix(sshd:session): session closed for user core May 17 00:23:25.791659 systemd[1]: sshd@9-64.23.167.227:22-139.178.68.195:46106.service: Deactivated successfully. May 17 00:23:25.795828 systemd[1]: session-10.scope: Deactivated successfully. May 17 00:23:25.796957 systemd-logind[1571]: Session 10 logged out. Waiting for processes to exit. May 17 00:23:25.798492 systemd-logind[1571]: Removed session 10. May 17 00:23:30.797178 systemd[1]: Started sshd@10-64.23.167.227:22-139.178.68.195:46116.service - OpenSSH per-connection server daemon (139.178.68.195:46116). May 17 00:23:30.860141 sshd[4085]: Accepted publickey for core from 139.178.68.195 port 46116 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:30.862170 sshd[4085]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:30.868278 systemd-logind[1571]: New session 11 of user core. May 17 00:23:30.872687 systemd[1]: Started session-11.scope - Session 11 of User core. May 17 00:23:31.024475 sshd[4085]: pam_unix(sshd:session): session closed for user core May 17 00:23:31.034634 systemd[1]: Started sshd@11-64.23.167.227:22-139.178.68.195:46118.service - OpenSSH per-connection server daemon (139.178.68.195:46118). May 17 00:23:31.035161 systemd[1]: sshd@10-64.23.167.227:22-139.178.68.195:46116.service: Deactivated successfully. May 17 00:23:31.042290 systemd[1]: session-11.scope: Deactivated successfully. May 17 00:23:31.045062 systemd-logind[1571]: Session 11 logged out. Waiting for processes to exit. May 17 00:23:31.046985 systemd-logind[1571]: Removed session 11. May 17 00:23:31.087093 sshd[4097]: Accepted publickey for core from 139.178.68.195 port 46118 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:31.089052 sshd[4097]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:31.095059 systemd-logind[1571]: New session 12 of user core. May 17 00:23:31.103922 systemd[1]: Started session-12.scope - Session 12 of User core. May 17 00:23:31.303606 sshd[4097]: pam_unix(sshd:session): session closed for user core May 17 00:23:31.314761 systemd[1]: Started sshd@12-64.23.167.227:22-139.178.68.195:46124.service - OpenSSH per-connection server daemon (139.178.68.195:46124). May 17 00:23:31.318019 systemd[1]: sshd@11-64.23.167.227:22-139.178.68.195:46118.service: Deactivated successfully. May 17 00:23:31.328670 systemd[1]: session-12.scope: Deactivated successfully. May 17 00:23:31.337534 systemd-logind[1571]: Session 12 logged out. Waiting for processes to exit. May 17 00:23:31.343396 systemd-logind[1571]: Removed session 12. May 17 00:23:31.392360 sshd[4109]: Accepted publickey for core from 139.178.68.195 port 46124 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:31.394171 sshd[4109]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:31.400144 systemd-logind[1571]: New session 13 of user core. May 17 00:23:31.405024 systemd[1]: Started session-13.scope - Session 13 of User core. May 17 00:23:31.558772 sshd[4109]: pam_unix(sshd:session): session closed for user core May 17 00:23:31.564341 systemd-logind[1571]: Session 13 logged out. Waiting for processes to exit. May 17 00:23:31.564838 systemd[1]: sshd@12-64.23.167.227:22-139.178.68.195:46124.service: Deactivated successfully. May 17 00:23:31.572127 systemd[1]: session-13.scope: Deactivated successfully. May 17 00:23:31.573453 systemd-logind[1571]: Removed session 13. May 17 00:23:36.569698 systemd[1]: Started sshd@13-64.23.167.227:22-139.178.68.195:33862.service - OpenSSH per-connection server daemon (139.178.68.195:33862). May 17 00:23:36.621510 sshd[4129]: Accepted publickey for core from 139.178.68.195 port 33862 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:36.623818 sshd[4129]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:36.630630 systemd-logind[1571]: New session 14 of user core. May 17 00:23:36.639918 systemd[1]: Started session-14.scope - Session 14 of User core. May 17 00:23:36.781619 sshd[4129]: pam_unix(sshd:session): session closed for user core May 17 00:23:36.784948 systemd[1]: sshd@13-64.23.167.227:22-139.178.68.195:33862.service: Deactivated successfully. May 17 00:23:36.790998 systemd[1]: session-14.scope: Deactivated successfully. May 17 00:23:36.792235 systemd-logind[1571]: Session 14 logged out. Waiting for processes to exit. May 17 00:23:36.793282 systemd-logind[1571]: Removed session 14. May 17 00:23:37.881668 systemd[1]: Started sshd@14-64.23.167.227:22-189.110.148.93:59568.service - OpenSSH per-connection server daemon (189.110.148.93:59568). May 17 00:23:38.898279 sshd[4142]: Invalid user es from 189.110.148.93 port 59568 May 17 00:23:39.092148 sshd[4142]: Received disconnect from 189.110.148.93 port 59568:11: Bye Bye [preauth] May 17 00:23:39.092148 sshd[4142]: Disconnected from invalid user es 189.110.148.93 port 59568 [preauth] May 17 00:23:39.094817 systemd[1]: sshd@14-64.23.167.227:22-189.110.148.93:59568.service: Deactivated successfully. May 17 00:23:41.791820 systemd[1]: Started sshd@15-64.23.167.227:22-139.178.68.195:33864.service - OpenSSH per-connection server daemon (139.178.68.195:33864). May 17 00:23:41.846951 sshd[4147]: Accepted publickey for core from 139.178.68.195 port 33864 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:41.849056 sshd[4147]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:41.857928 systemd-logind[1571]: New session 15 of user core. May 17 00:23:41.865820 systemd[1]: Started session-15.scope - Session 15 of User core. May 17 00:23:42.009682 sshd[4147]: pam_unix(sshd:session): session closed for user core May 17 00:23:42.016434 systemd[1]: sshd@15-64.23.167.227:22-139.178.68.195:33864.service: Deactivated successfully. May 17 00:23:42.020377 systemd[1]: session-15.scope: Deactivated successfully. May 17 00:23:42.020427 systemd-logind[1571]: Session 15 logged out. Waiting for processes to exit. May 17 00:23:42.022162 systemd-logind[1571]: Removed session 15. May 17 00:23:47.021730 systemd[1]: Started sshd@16-64.23.167.227:22-139.178.68.195:34676.service - OpenSSH per-connection server daemon (139.178.68.195:34676). May 17 00:23:47.064831 sshd[4163]: Accepted publickey for core from 139.178.68.195 port 34676 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:47.066894 sshd[4163]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:47.074758 systemd-logind[1571]: New session 16 of user core. May 17 00:23:47.079347 systemd[1]: Started session-16.scope - Session 16 of User core. May 17 00:23:47.227398 sshd[4163]: pam_unix(sshd:session): session closed for user core May 17 00:23:47.235791 systemd[1]: Started sshd@17-64.23.167.227:22-139.178.68.195:34680.service - OpenSSH per-connection server daemon (139.178.68.195:34680). May 17 00:23:47.237520 systemd[1]: sshd@16-64.23.167.227:22-139.178.68.195:34676.service: Deactivated successfully. May 17 00:23:47.244610 systemd[1]: session-16.scope: Deactivated successfully. May 17 00:23:47.247742 systemd-logind[1571]: Session 16 logged out. Waiting for processes to exit. May 17 00:23:47.249543 systemd-logind[1571]: Removed session 16. May 17 00:23:47.293382 sshd[4174]: Accepted publickey for core from 139.178.68.195 port 34680 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:47.294793 sshd[4174]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:47.301594 systemd-logind[1571]: New session 17 of user core. May 17 00:23:47.312001 systemd[1]: Started session-17.scope - Session 17 of User core. May 17 00:23:47.618724 sshd[4174]: pam_unix(sshd:session): session closed for user core May 17 00:23:47.635866 systemd[1]: Started sshd@18-64.23.167.227:22-139.178.68.195:34688.service - OpenSSH per-connection server daemon (139.178.68.195:34688). May 17 00:23:47.636513 systemd[1]: sshd@17-64.23.167.227:22-139.178.68.195:34680.service: Deactivated successfully. May 17 00:23:47.640669 systemd[1]: session-17.scope: Deactivated successfully. May 17 00:23:47.642991 systemd-logind[1571]: Session 17 logged out. Waiting for processes to exit. May 17 00:23:47.647332 systemd-logind[1571]: Removed session 17. May 17 00:23:47.697661 sshd[4187]: Accepted publickey for core from 139.178.68.195 port 34688 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:47.699878 sshd[4187]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:47.706809 systemd-logind[1571]: New session 18 of user core. May 17 00:23:47.713783 systemd[1]: Started session-18.scope - Session 18 of User core. May 17 00:23:49.612633 sshd[4187]: pam_unix(sshd:session): session closed for user core May 17 00:23:49.636636 systemd[1]: Started sshd@19-64.23.167.227:22-139.178.68.195:34698.service - OpenSSH per-connection server daemon (139.178.68.195:34698). May 17 00:23:49.637187 systemd[1]: sshd@18-64.23.167.227:22-139.178.68.195:34688.service: Deactivated successfully. May 17 00:23:49.657767 systemd[1]: session-18.scope: Deactivated successfully. May 17 00:23:49.660839 systemd-logind[1571]: Session 18 logged out. Waiting for processes to exit. May 17 00:23:49.670064 systemd-logind[1571]: Removed session 18. May 17 00:23:49.732557 sshd[4205]: Accepted publickey for core from 139.178.68.195 port 34698 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:49.734552 sshd[4205]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:49.741448 systemd-logind[1571]: New session 19 of user core. May 17 00:23:49.746812 systemd[1]: Started session-19.scope - Session 19 of User core. May 17 00:23:50.085815 sshd[4205]: pam_unix(sshd:session): session closed for user core May 17 00:23:50.098778 systemd[1]: Started sshd@20-64.23.167.227:22-139.178.68.195:34704.service - OpenSSH per-connection server daemon (139.178.68.195:34704). May 17 00:23:50.100246 systemd[1]: sshd@19-64.23.167.227:22-139.178.68.195:34698.service: Deactivated successfully. May 17 00:23:50.108218 systemd[1]: session-19.scope: Deactivated successfully. May 17 00:23:50.108713 systemd-logind[1571]: Session 19 logged out. Waiting for processes to exit. May 17 00:23:50.111931 systemd-logind[1571]: Removed session 19. May 17 00:23:50.159079 sshd[4217]: Accepted publickey for core from 139.178.68.195 port 34704 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:50.161221 sshd[4217]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:50.168803 systemd-logind[1571]: New session 20 of user core. May 17 00:23:50.174840 systemd[1]: Started session-20.scope - Session 20 of User core. May 17 00:23:50.317033 sshd[4217]: pam_unix(sshd:session): session closed for user core May 17 00:23:50.321666 systemd[1]: sshd@20-64.23.167.227:22-139.178.68.195:34704.service: Deactivated successfully. May 17 00:23:50.327786 systemd[1]: session-20.scope: Deactivated successfully. May 17 00:23:50.330846 systemd-logind[1571]: Session 20 logged out. Waiting for processes to exit. May 17 00:23:50.331994 systemd-logind[1571]: Removed session 20. May 17 00:23:55.334508 systemd[1]: Started sshd@21-64.23.167.227:22-139.178.68.195:52868.service - OpenSSH per-connection server daemon (139.178.68.195:52868). May 17 00:23:55.381966 sshd[4234]: Accepted publickey for core from 139.178.68.195 port 52868 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:23:55.384580 sshd[4234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:23:55.391676 systemd-logind[1571]: New session 21 of user core. May 17 00:23:55.396763 systemd[1]: Started session-21.scope - Session 21 of User core. May 17 00:23:55.557066 sshd[4234]: pam_unix(sshd:session): session closed for user core May 17 00:23:55.565449 systemd[1]: sshd@21-64.23.167.227:22-139.178.68.195:52868.service: Deactivated successfully. May 17 00:23:55.570023 systemd[1]: session-21.scope: Deactivated successfully. May 17 00:23:55.571593 systemd-logind[1571]: Session 21 logged out. Waiting for processes to exit. May 17 00:23:55.573678 systemd-logind[1571]: Removed session 21. May 17 00:23:56.403101 kubelet[2658]: E0517 00:23:56.403030 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:23:58.402292 kubelet[2658]: E0517 00:23:58.402246 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:00.402858 kubelet[2658]: E0517 00:24:00.402764 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:00.567960 systemd[1]: Started sshd@22-64.23.167.227:22-139.178.68.195:52880.service - OpenSSH per-connection server daemon (139.178.68.195:52880). May 17 00:24:00.624194 sshd[4250]: Accepted publickey for core from 139.178.68.195 port 52880 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:24:00.626247 sshd[4250]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:24:00.633419 systemd-logind[1571]: New session 22 of user core. May 17 00:24:00.642217 systemd[1]: Started session-22.scope - Session 22 of User core. May 17 00:24:00.785892 sshd[4250]: pam_unix(sshd:session): session closed for user core May 17 00:24:00.791685 systemd[1]: sshd@22-64.23.167.227:22-139.178.68.195:52880.service: Deactivated successfully. May 17 00:24:00.796258 systemd-logind[1571]: Session 22 logged out. Waiting for processes to exit. May 17 00:24:00.796641 systemd[1]: session-22.scope: Deactivated successfully. May 17 00:24:00.798901 systemd-logind[1571]: Removed session 22. May 17 00:24:04.402238 kubelet[2658]: E0517 00:24:04.402192 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:05.404722 kubelet[2658]: E0517 00:24:05.403014 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:05.796797 systemd[1]: Started sshd@23-64.23.167.227:22-139.178.68.195:48992.service - OpenSSH per-connection server daemon (139.178.68.195:48992). May 17 00:24:05.837567 sshd[4264]: Accepted publickey for core from 139.178.68.195 port 48992 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:24:05.839813 sshd[4264]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:24:05.847678 systemd-logind[1571]: New session 23 of user core. May 17 00:24:05.855861 systemd[1]: Started session-23.scope - Session 23 of User core. May 17 00:24:05.992760 sshd[4264]: pam_unix(sshd:session): session closed for user core May 17 00:24:05.996742 systemd[1]: sshd@23-64.23.167.227:22-139.178.68.195:48992.service: Deactivated successfully. May 17 00:24:05.997145 systemd-logind[1571]: Session 23 logged out. Waiting for processes to exit. May 17 00:24:06.002595 systemd[1]: session-23.scope: Deactivated successfully. May 17 00:24:06.006099 systemd-logind[1571]: Removed session 23. May 17 00:24:06.402779 kubelet[2658]: E0517 00:24:06.402693 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:11.004119 systemd[1]: Started sshd@24-64.23.167.227:22-139.178.68.195:49006.service - OpenSSH per-connection server daemon (139.178.68.195:49006). May 17 00:24:11.046900 sshd[4278]: Accepted publickey for core from 139.178.68.195 port 49006 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:24:11.049203 sshd[4278]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:24:11.055185 systemd-logind[1571]: New session 24 of user core. May 17 00:24:11.060704 systemd[1]: Started session-24.scope - Session 24 of User core. May 17 00:24:11.205598 sshd[4278]: pam_unix(sshd:session): session closed for user core May 17 00:24:11.215893 systemd[1]: Started sshd@25-64.23.167.227:22-139.178.68.195:49020.service - OpenSSH per-connection server daemon (139.178.68.195:49020). May 17 00:24:11.216651 systemd[1]: sshd@24-64.23.167.227:22-139.178.68.195:49006.service: Deactivated successfully. May 17 00:24:11.227544 systemd-logind[1571]: Session 24 logged out. Waiting for processes to exit. May 17 00:24:11.229237 systemd[1]: session-24.scope: Deactivated successfully. May 17 00:24:11.232503 systemd-logind[1571]: Removed session 24. May 17 00:24:11.259775 sshd[4288]: Accepted publickey for core from 139.178.68.195 port 49020 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:24:11.262344 sshd[4288]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:24:11.270608 systemd-logind[1571]: New session 25 of user core. May 17 00:24:11.278895 systemd[1]: Started session-25.scope - Session 25 of User core. May 17 00:24:12.663939 containerd[1595]: time="2025-05-17T00:24:12.663793397Z" level=info msg="StopContainer for \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\" with timeout 30 (s)" May 17 00:24:12.665476 containerd[1595]: time="2025-05-17T00:24:12.665052624Z" level=info msg="Stop container \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\" with signal terminated" May 17 00:24:12.689586 systemd[1]: run-containerd-runc-k8s.io-a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c-runc.BfTbQa.mount: Deactivated successfully. May 17 00:24:12.718024 containerd[1595]: time="2025-05-17T00:24:12.717930371Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 17 00:24:12.727578 containerd[1595]: time="2025-05-17T00:24:12.727521146Z" level=info msg="StopContainer for \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\" with timeout 2 (s)" May 17 00:24:12.728262 containerd[1595]: time="2025-05-17T00:24:12.728203795Z" level=info msg="Stop container \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\" with signal terminated" May 17 00:24:12.742536 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126-rootfs.mount: Deactivated successfully. May 17 00:24:12.743425 systemd-networkd[1219]: lxc_health: Link DOWN May 17 00:24:12.743429 systemd-networkd[1219]: lxc_health: Lost carrier May 17 00:24:12.765879 containerd[1595]: time="2025-05-17T00:24:12.764604641Z" level=info msg="shim disconnected" id=1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126 namespace=k8s.io May 17 00:24:12.765879 containerd[1595]: time="2025-05-17T00:24:12.764695928Z" level=warning msg="cleaning up after shim disconnected" id=1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126 namespace=k8s.io May 17 00:24:12.765879 containerd[1595]: time="2025-05-17T00:24:12.764710633Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:12.806063 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c-rootfs.mount: Deactivated successfully. May 17 00:24:12.811160 containerd[1595]: time="2025-05-17T00:24:12.811098824Z" level=info msg="StopContainer for \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\" returns successfully" May 17 00:24:12.811981 containerd[1595]: time="2025-05-17T00:24:12.811560706Z" level=info msg="shim disconnected" id=a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c namespace=k8s.io May 17 00:24:12.811981 containerd[1595]: time="2025-05-17T00:24:12.811637704Z" level=warning msg="cleaning up after shim disconnected" id=a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c namespace=k8s.io May 17 00:24:12.811981 containerd[1595]: time="2025-05-17T00:24:12.811647018Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:12.813119 containerd[1595]: time="2025-05-17T00:24:12.812711082Z" level=info msg="StopPodSandbox for \"c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012\"" May 17 00:24:12.813119 containerd[1595]: time="2025-05-17T00:24:12.812750794Z" level=info msg="Container to stop \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:24:12.817924 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012-shm.mount: Deactivated successfully. May 17 00:24:12.836743 containerd[1595]: time="2025-05-17T00:24:12.836511824Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:24:12Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io May 17 00:24:12.839559 containerd[1595]: time="2025-05-17T00:24:12.839426017Z" level=info msg="StopContainer for \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\" returns successfully" May 17 00:24:12.840104 containerd[1595]: time="2025-05-17T00:24:12.839991683Z" level=info msg="StopPodSandbox for \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\"" May 17 00:24:12.840104 containerd[1595]: time="2025-05-17T00:24:12.840053806Z" level=info msg="Container to stop \"977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:24:12.840104 containerd[1595]: time="2025-05-17T00:24:12.840073278Z" level=info msg="Container to stop \"aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:24:12.840104 containerd[1595]: time="2025-05-17T00:24:12.840086953Z" level=info msg="Container to stop \"f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:24:12.840104 containerd[1595]: time="2025-05-17T00:24:12.840100269Z" level=info msg="Container to stop \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:24:12.840777 containerd[1595]: time="2025-05-17T00:24:12.840113776Z" level=info msg="Container to stop \"bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:24:12.887286 containerd[1595]: time="2025-05-17T00:24:12.884697466Z" level=info msg="shim disconnected" id=c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012 namespace=k8s.io May 17 00:24:12.887286 containerd[1595]: time="2025-05-17T00:24:12.887231857Z" level=warning msg="cleaning up after shim disconnected" id=c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012 namespace=k8s.io May 17 00:24:12.887286 containerd[1595]: time="2025-05-17T00:24:12.887244136Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:12.941469 containerd[1595]: time="2025-05-17T00:24:12.939497230Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:24:12Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io May 17 00:24:12.941469 containerd[1595]: time="2025-05-17T00:24:12.940601295Z" level=info msg="TearDown network for sandbox \"c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012\" successfully" May 17 00:24:12.941469 containerd[1595]: time="2025-05-17T00:24:12.940624866Z" level=info msg="StopPodSandbox for \"c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012\" returns successfully" May 17 00:24:12.948239 containerd[1595]: time="2025-05-17T00:24:12.948116156Z" level=info msg="shim disconnected" id=e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e namespace=k8s.io May 17 00:24:12.948239 containerd[1595]: time="2025-05-17T00:24:12.948182847Z" level=warning msg="cleaning up after shim disconnected" id=e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e namespace=k8s.io May 17 00:24:12.948239 containerd[1595]: time="2025-05-17T00:24:12.948191863Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:12.979271 kubelet[2658]: I0517 00:24:12.976150 2658 scope.go:117] "RemoveContainer" containerID="1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126" May 17 00:24:12.989340 containerd[1595]: time="2025-05-17T00:24:12.986616909Z" level=info msg="RemoveContainer for \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\"" May 17 00:24:13.007622 containerd[1595]: time="2025-05-17T00:24:13.004591345Z" level=info msg="RemoveContainer for \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\" returns successfully" May 17 00:24:13.013115 kubelet[2658]: I0517 00:24:13.013013 2658 scope.go:117] "RemoveContainer" containerID="1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126" May 17 00:24:13.043109 containerd[1595]: time="2025-05-17T00:24:13.014686196Z" level=error msg="ContainerStatus for \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\": not found" May 17 00:24:13.043362 kubelet[2658]: E0517 00:24:13.043306 2658 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\": not found" containerID="1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126" May 17 00:24:13.043626 kubelet[2658]: I0517 00:24:13.043383 2658 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126"} err="failed to get container status \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\": rpc error: code = NotFound desc = an error occurred when try to find container \"1635cd59ba76f4c7c7e05375efc99064fa219189d2312a71fbbba5a07752a126\": not found" May 17 00:24:13.052491 containerd[1595]: time="2025-05-17T00:24:13.052448661Z" level=info msg="TearDown network for sandbox \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" successfully" May 17 00:24:13.053175 containerd[1595]: time="2025-05-17T00:24:13.052635502Z" level=info msg="StopPodSandbox for \"e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e\" returns successfully" May 17 00:24:13.107813 kubelet[2658]: I0517 00:24:13.107745 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhj9t\" (UniqueName: \"kubernetes.io/projected/a77f0e64-ad05-4614-bce9-cfe0daa984dc-kube-api-access-mhj9t\") pod \"a77f0e64-ad05-4614-bce9-cfe0daa984dc\" (UID: \"a77f0e64-ad05-4614-bce9-cfe0daa984dc\") " May 17 00:24:13.107813 kubelet[2658]: I0517 00:24:13.107810 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a77f0e64-ad05-4614-bce9-cfe0daa984dc-cilium-config-path\") pod \"a77f0e64-ad05-4614-bce9-cfe0daa984dc\" (UID: \"a77f0e64-ad05-4614-bce9-cfe0daa984dc\") " May 17 00:24:13.110435 kubelet[2658]: I0517 00:24:13.110303 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a77f0e64-ad05-4614-bce9-cfe0daa984dc-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "a77f0e64-ad05-4614-bce9-cfe0daa984dc" (UID: "a77f0e64-ad05-4614-bce9-cfe0daa984dc"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 17 00:24:13.112648 kubelet[2658]: I0517 00:24:13.112589 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a77f0e64-ad05-4614-bce9-cfe0daa984dc-kube-api-access-mhj9t" (OuterVolumeSpecName: "kube-api-access-mhj9t") pod "a77f0e64-ad05-4614-bce9-cfe0daa984dc" (UID: "a77f0e64-ad05-4614-bce9-cfe0daa984dc"). InnerVolumeSpecName "kube-api-access-mhj9t". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:24:13.208198 kubelet[2658]: I0517 00:24:13.208127 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-config-path\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.208198 kubelet[2658]: I0517 00:24:13.208215 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-bpf-maps\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.208750 kubelet[2658]: I0517 00:24:13.208244 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-xtables-lock\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.208750 kubelet[2658]: I0517 00:24:13.208272 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hubble-tls\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.208750 kubelet[2658]: I0517 00:24:13.208297 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hostproc\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.208750 kubelet[2658]: I0517 00:24:13.208350 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-lib-modules\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.208750 kubelet[2658]: I0517 00:24:13.208380 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-clustermesh-secrets\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.208750 kubelet[2658]: I0517 00:24:13.208407 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjn9t\" (UniqueName: \"kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-kube-api-access-rjn9t\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.209133 kubelet[2658]: I0517 00:24:13.208430 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-run\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.209133 kubelet[2658]: I0517 00:24:13.208456 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-cgroup\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.209133 kubelet[2658]: I0517 00:24:13.208481 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-kernel\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.209133 kubelet[2658]: I0517 00:24:13.208506 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-etc-cni-netd\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.209133 kubelet[2658]: I0517 00:24:13.208526 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-net\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.209133 kubelet[2658]: I0517 00:24:13.208547 2658 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cni-path\") pod \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\" (UID: \"e02a6237-4e09-4ae1-82f5-f96d9fba9fd1\") " May 17 00:24:13.209337 kubelet[2658]: I0517 00:24:13.208604 2658 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhj9t\" (UniqueName: \"kubernetes.io/projected/a77f0e64-ad05-4614-bce9-cfe0daa984dc-kube-api-access-mhj9t\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.209337 kubelet[2658]: I0517 00:24:13.208620 2658 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a77f0e64-ad05-4614-bce9-cfe0daa984dc-cilium-config-path\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.209337 kubelet[2658]: I0517 00:24:13.208688 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cni-path" (OuterVolumeSpecName: "cni-path") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.211255 kubelet[2658]: I0517 00:24:13.211196 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 17 00:24:13.211805 kubelet[2658]: I0517 00:24:13.211612 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.211805 kubelet[2658]: I0517 00:24:13.211669 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.212644 kubelet[2658]: I0517 00:24:13.212519 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.212644 kubelet[2658]: I0517 00:24:13.212568 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.212644 kubelet[2658]: I0517 00:24:13.212586 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.212644 kubelet[2658]: I0517 00:24:13.212600 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.212644 kubelet[2658]: I0517 00:24:13.212617 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.213080 kubelet[2658]: I0517 00:24:13.212883 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hostproc" (OuterVolumeSpecName: "hostproc") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.213080 kubelet[2658]: I0517 00:24:13.212955 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:24:13.214467 kubelet[2658]: I0517 00:24:13.214423 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-kube-api-access-rjn9t" (OuterVolumeSpecName: "kube-api-access-rjn9t") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "kube-api-access-rjn9t". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:24:13.216394 kubelet[2658]: I0517 00:24:13.216030 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 17 00:24:13.217029 kubelet[2658]: I0517 00:24:13.216972 2658 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" (UID: "e02a6237-4e09-4ae1-82f5-f96d9fba9fd1"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:24:13.309554 kubelet[2658]: I0517 00:24:13.309476 2658 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hostproc\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309554 kubelet[2658]: I0517 00:24:13.309522 2658 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-lib-modules\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309554 kubelet[2658]: I0517 00:24:13.309536 2658 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-clustermesh-secrets\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309554 kubelet[2658]: I0517 00:24:13.309555 2658 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjn9t\" (UniqueName: \"kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-kube-api-access-rjn9t\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309554 kubelet[2658]: I0517 00:24:13.309568 2658 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-run\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309554 kubelet[2658]: I0517 00:24:13.309577 2658 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-cgroup\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309585 2658 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-kernel\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309595 2658 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-etc-cni-netd\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309607 2658 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-host-proc-sys-net\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309615 2658 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cni-path\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309624 2658 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-xtables-lock\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309634 2658 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-cilium-config-path\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309644 2658 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-bpf-maps\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.309884 kubelet[2658]: I0517 00:24:13.309652 2658 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1-hubble-tls\") on node \"ci-4081.3.3-n-0e116cd3eb\" DevicePath \"\"" May 17 00:24:13.678475 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c1889552a35a8fea469a58089094927d2b31233cea9128f34c19058d5260d012-rootfs.mount: Deactivated successfully. May 17 00:24:13.678710 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e-rootfs.mount: Deactivated successfully. May 17 00:24:13.678880 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e53eb6df139bc8ce369b7929aeeddc8f676928735b7c97283d12c72771e9c71e-shm.mount: Deactivated successfully. May 17 00:24:13.679045 systemd[1]: var-lib-kubelet-pods-e02a6237\x2d4e09\x2d4ae1\x2d82f5\x2df96d9fba9fd1-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 17 00:24:13.679178 systemd[1]: var-lib-kubelet-pods-e02a6237\x2d4e09\x2d4ae1\x2d82f5\x2df96d9fba9fd1-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 17 00:24:13.679273 systemd[1]: var-lib-kubelet-pods-a77f0e64\x2dad05\x2d4614\x2dbce9\x2dcfe0daa984dc-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dmhj9t.mount: Deactivated successfully. May 17 00:24:13.680026 systemd[1]: var-lib-kubelet-pods-e02a6237\x2d4e09\x2d4ae1\x2d82f5\x2df96d9fba9fd1-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2drjn9t.mount: Deactivated successfully. May 17 00:24:14.023155 kubelet[2658]: I0517 00:24:14.023103 2658 scope.go:117] "RemoveContainer" containerID="a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c" May 17 00:24:14.028260 containerd[1595]: time="2025-05-17T00:24:14.027518882Z" level=info msg="RemoveContainer for \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\"" May 17 00:24:14.030974 containerd[1595]: time="2025-05-17T00:24:14.030868792Z" level=info msg="RemoveContainer for \"a155c15d4e155dff9921041ab2016b6fba732c97f2aa986c72037b61d041055c\" returns successfully" May 17 00:24:14.032209 kubelet[2658]: I0517 00:24:14.032173 2658 scope.go:117] "RemoveContainer" containerID="f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1" May 17 00:24:14.034659 containerd[1595]: time="2025-05-17T00:24:14.034630273Z" level=info msg="RemoveContainer for \"f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1\"" May 17 00:24:14.037207 containerd[1595]: time="2025-05-17T00:24:14.037108538Z" level=info msg="RemoveContainer for \"f366f1acdde0bfc01f25c251968598d718374f2d962783a70978a9e50d1535e1\" returns successfully" May 17 00:24:14.037378 kubelet[2658]: I0517 00:24:14.037351 2658 scope.go:117] "RemoveContainer" containerID="aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491" May 17 00:24:14.040136 containerd[1595]: time="2025-05-17T00:24:14.040106499Z" level=info msg="RemoveContainer for \"aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491\"" May 17 00:24:14.048985 containerd[1595]: time="2025-05-17T00:24:14.048750067Z" level=info msg="RemoveContainer for \"aca53803d051459b7991969f402d67bced20f3bdef922b629a4213033a21f491\" returns successfully" May 17 00:24:14.049338 kubelet[2658]: I0517 00:24:14.049207 2658 scope.go:117] "RemoveContainer" containerID="977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0" May 17 00:24:14.053185 containerd[1595]: time="2025-05-17T00:24:14.053075468Z" level=info msg="RemoveContainer for \"977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0\"" May 17 00:24:14.059171 containerd[1595]: time="2025-05-17T00:24:14.058341439Z" level=info msg="RemoveContainer for \"977c0154b96b69a7911a4ed220399a421e11de08da8f331c6cc2e6ced55180b0\" returns successfully" May 17 00:24:14.060015 kubelet[2658]: I0517 00:24:14.059850 2658 scope.go:117] "RemoveContainer" containerID="bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e" May 17 00:24:14.063026 containerd[1595]: time="2025-05-17T00:24:14.062755932Z" level=info msg="RemoveContainer for \"bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e\"" May 17 00:24:14.069278 containerd[1595]: time="2025-05-17T00:24:14.069203944Z" level=info msg="RemoveContainer for \"bd499152e2d7c47d898c054a0b73cd05f67a5f96cc9f7cece2cbf2302780832e\" returns successfully" May 17 00:24:14.605119 sshd[4288]: pam_unix(sshd:session): session closed for user core May 17 00:24:14.613819 systemd[1]: Started sshd@26-64.23.167.227:22-139.178.68.195:48086.service - OpenSSH per-connection server daemon (139.178.68.195:48086). May 17 00:24:14.616241 systemd[1]: sshd@25-64.23.167.227:22-139.178.68.195:49020.service: Deactivated successfully. May 17 00:24:14.621894 systemd[1]: session-25.scope: Deactivated successfully. May 17 00:24:14.628079 systemd-logind[1571]: Session 25 logged out. Waiting for processes to exit. May 17 00:24:14.630082 systemd-logind[1571]: Removed session 25. May 17 00:24:14.685359 sshd[4460]: Accepted publickey for core from 139.178.68.195 port 48086 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:24:14.687579 sshd[4460]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:24:14.693368 systemd-logind[1571]: New session 26 of user core. May 17 00:24:14.696668 systemd[1]: Started session-26.scope - Session 26 of User core. May 17 00:24:15.406409 kubelet[2658]: E0517 00:24:15.404962 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:15.411721 kubelet[2658]: I0517 00:24:15.411673 2658 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a77f0e64-ad05-4614-bce9-cfe0daa984dc" path="/var/lib/kubelet/pods/a77f0e64-ad05-4614-bce9-cfe0daa984dc/volumes" May 17 00:24:15.412174 kubelet[2658]: I0517 00:24:15.412151 2658 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" path="/var/lib/kubelet/pods/e02a6237-4e09-4ae1-82f5-f96d9fba9fd1/volumes" May 17 00:24:15.466353 sshd[4460]: pam_unix(sshd:session): session closed for user core May 17 00:24:15.479543 systemd[1]: Started sshd@27-64.23.167.227:22-139.178.68.195:48094.service - OpenSSH per-connection server daemon (139.178.68.195:48094). May 17 00:24:15.480558 systemd[1]: sshd@26-64.23.167.227:22-139.178.68.195:48086.service: Deactivated successfully. May 17 00:24:15.491087 systemd[1]: session-26.scope: Deactivated successfully. May 17 00:24:15.491353 systemd-logind[1571]: Session 26 logged out. Waiting for processes to exit. May 17 00:24:15.496916 systemd-logind[1571]: Removed session 26. May 17 00:24:15.512200 kubelet[2658]: E0517 00:24:15.510965 2658 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a77f0e64-ad05-4614-bce9-cfe0daa984dc" containerName="cilium-operator" May 17 00:24:15.515356 kubelet[2658]: E0517 00:24:15.513956 2658 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" containerName="cilium-agent" May 17 00:24:15.515356 kubelet[2658]: E0517 00:24:15.514010 2658 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" containerName="mount-cgroup" May 17 00:24:15.515356 kubelet[2658]: E0517 00:24:15.514018 2658 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" containerName="apply-sysctl-overwrites" May 17 00:24:15.515356 kubelet[2658]: E0517 00:24:15.514026 2658 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" containerName="mount-bpf-fs" May 17 00:24:15.515356 kubelet[2658]: E0517 00:24:15.514032 2658 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" containerName="clean-cilium-state" May 17 00:24:15.515356 kubelet[2658]: I0517 00:24:15.514105 2658 memory_manager.go:354] "RemoveStaleState removing state" podUID="a77f0e64-ad05-4614-bce9-cfe0daa984dc" containerName="cilium-operator" May 17 00:24:15.515356 kubelet[2658]: I0517 00:24:15.514116 2658 memory_manager.go:354] "RemoveStaleState removing state" podUID="e02a6237-4e09-4ae1-82f5-f96d9fba9fd1" containerName="cilium-agent" May 17 00:24:15.543344 kubelet[2658]: E0517 00:24:15.539083 2658 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 17 00:24:15.576375 sshd[4474]: Accepted publickey for core from 139.178.68.195 port 48094 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:24:15.578013 sshd[4474]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:24:15.585360 systemd-logind[1571]: New session 27 of user core. May 17 00:24:15.596789 systemd[1]: Started session-27.scope - Session 27 of User core. May 17 00:24:15.622292 kubelet[2658]: I0517 00:24:15.622133 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-hostproc\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.622292 kubelet[2658]: I0517 00:24:15.622216 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-xtables-lock\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.622292 kubelet[2658]: I0517 00:24:15.622249 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6a064c2c-b535-4ece-90d9-357c03bc2a85-hubble-tls\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.622680 kubelet[2658]: I0517 00:24:15.622431 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-host-proc-sys-kernel\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.622680 kubelet[2658]: I0517 00:24:15.622457 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6a064c2c-b535-4ece-90d9-357c03bc2a85-cilium-ipsec-secrets\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623088 kubelet[2658]: I0517 00:24:15.622777 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-lib-modules\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623088 kubelet[2658]: I0517 00:24:15.622805 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9gjg\" (UniqueName: \"kubernetes.io/projected/6a064c2c-b535-4ece-90d9-357c03bc2a85-kube-api-access-g9gjg\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623088 kubelet[2658]: I0517 00:24:15.622848 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-cilium-cgroup\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623088 kubelet[2658]: I0517 00:24:15.622862 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6a064c2c-b535-4ece-90d9-357c03bc2a85-cilium-config-path\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623088 kubelet[2658]: I0517 00:24:15.622877 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-etc-cni-netd\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623248 kubelet[2658]: I0517 00:24:15.622940 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-host-proc-sys-net\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623248 kubelet[2658]: I0517 00:24:15.622962 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-cilium-run\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623248 kubelet[2658]: I0517 00:24:15.622987 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-bpf-maps\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623248 kubelet[2658]: I0517 00:24:15.623006 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6a064c2c-b535-4ece-90d9-357c03bc2a85-cni-path\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.623248 kubelet[2658]: I0517 00:24:15.623024 2658 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6a064c2c-b535-4ece-90d9-357c03bc2a85-clustermesh-secrets\") pod \"cilium-bpmf4\" (UID: \"6a064c2c-b535-4ece-90d9-357c03bc2a85\") " pod="kube-system/cilium-bpmf4" May 17 00:24:15.659507 sshd[4474]: pam_unix(sshd:session): session closed for user core May 17 00:24:15.666053 systemd[1]: sshd@27-64.23.167.227:22-139.178.68.195:48094.service: Deactivated successfully. May 17 00:24:15.669078 systemd-logind[1571]: Session 27 logged out. Waiting for processes to exit. May 17 00:24:15.675699 systemd[1]: Started sshd@28-64.23.167.227:22-139.178.68.195:48096.service - OpenSSH per-connection server daemon (139.178.68.195:48096). May 17 00:24:15.676169 systemd[1]: session-27.scope: Deactivated successfully. May 17 00:24:15.677496 systemd-logind[1571]: Removed session 27. May 17 00:24:15.734507 sshd[4486]: Accepted publickey for core from 139.178.68.195 port 48096 ssh2: RSA SHA256:TM7Vm5JNsRT9OkRUxlGPKsAsv9oxy8GzboZ61mm4KqQ May 17 00:24:15.747763 sshd[4486]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 17 00:24:15.772564 systemd-logind[1571]: New session 28 of user core. May 17 00:24:15.780830 systemd[1]: Started session-28.scope - Session 28 of User core. May 17 00:24:15.840179 kubelet[2658]: E0517 00:24:15.840137 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:15.841112 containerd[1595]: time="2025-05-17T00:24:15.841056555Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-bpmf4,Uid:6a064c2c-b535-4ece-90d9-357c03bc2a85,Namespace:kube-system,Attempt:0,}" May 17 00:24:15.876741 containerd[1595]: time="2025-05-17T00:24:15.876533987Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:24:15.876741 containerd[1595]: time="2025-05-17T00:24:15.876651546Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:24:15.876741 containerd[1595]: time="2025-05-17T00:24:15.876668743Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:24:15.877468 containerd[1595]: time="2025-05-17T00:24:15.876848767Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:24:15.943812 containerd[1595]: time="2025-05-17T00:24:15.938146531Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-bpmf4,Uid:6a064c2c-b535-4ece-90d9-357c03bc2a85,Namespace:kube-system,Attempt:0,} returns sandbox id \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\"" May 17 00:24:15.950785 kubelet[2658]: E0517 00:24:15.948608 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:15.951949 containerd[1595]: time="2025-05-17T00:24:15.951909332Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:24:15.974818 containerd[1595]: time="2025-05-17T00:24:15.974764225Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"f03adf5e1c55314b3c186677ca3138d7e3bcb809317987e42ef4464685ee2158\"" May 17 00:24:15.977960 containerd[1595]: time="2025-05-17T00:24:15.976643145Z" level=info msg="StartContainer for \"f03adf5e1c55314b3c186677ca3138d7e3bcb809317987e42ef4464685ee2158\"" May 17 00:24:16.054557 containerd[1595]: time="2025-05-17T00:24:16.054187862Z" level=info msg="StartContainer for \"f03adf5e1c55314b3c186677ca3138d7e3bcb809317987e42ef4464685ee2158\" returns successfully" May 17 00:24:16.093805 containerd[1595]: time="2025-05-17T00:24:16.093650658Z" level=info msg="shim disconnected" id=f03adf5e1c55314b3c186677ca3138d7e3bcb809317987e42ef4464685ee2158 namespace=k8s.io May 17 00:24:16.093805 containerd[1595]: time="2025-05-17T00:24:16.093798285Z" level=warning msg="cleaning up after shim disconnected" id=f03adf5e1c55314b3c186677ca3138d7e3bcb809317987e42ef4464685ee2158 namespace=k8s.io May 17 00:24:16.093805 containerd[1595]: time="2025-05-17T00:24:16.093807738Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:17.042731 kubelet[2658]: E0517 00:24:17.042207 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:17.045614 containerd[1595]: time="2025-05-17T00:24:17.045292172Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 17 00:24:17.060749 containerd[1595]: time="2025-05-17T00:24:17.060704552Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"0187102883277b2eea85f081938bd664e00ea033f47c0a8f74e6e12fc2b459e2\"" May 17 00:24:17.067128 containerd[1595]: time="2025-05-17T00:24:17.065159718Z" level=info msg="StartContainer for \"0187102883277b2eea85f081938bd664e00ea033f47c0a8f74e6e12fc2b459e2\"" May 17 00:24:17.130373 containerd[1595]: time="2025-05-17T00:24:17.129408448Z" level=info msg="StartContainer for \"0187102883277b2eea85f081938bd664e00ea033f47c0a8f74e6e12fc2b459e2\" returns successfully" May 17 00:24:17.161757 containerd[1595]: time="2025-05-17T00:24:17.161687611Z" level=info msg="shim disconnected" id=0187102883277b2eea85f081938bd664e00ea033f47c0a8f74e6e12fc2b459e2 namespace=k8s.io May 17 00:24:17.161757 containerd[1595]: time="2025-05-17T00:24:17.161748540Z" level=warning msg="cleaning up after shim disconnected" id=0187102883277b2eea85f081938bd664e00ea033f47c0a8f74e6e12fc2b459e2 namespace=k8s.io May 17 00:24:17.161757 containerd[1595]: time="2025-05-17T00:24:17.161759910Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:17.736733 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0187102883277b2eea85f081938bd664e00ea033f47c0a8f74e6e12fc2b459e2-rootfs.mount: Deactivated successfully. May 17 00:24:17.881167 kubelet[2658]: I0517 00:24:17.880517 2658 setters.go:600] "Node became not ready" node="ci-4081.3.3-n-0e116cd3eb" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-17T00:24:17Z","lastTransitionTime":"2025-05-17T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 17 00:24:18.047908 kubelet[2658]: E0517 00:24:18.047685 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:18.055168 containerd[1595]: time="2025-05-17T00:24:18.053786079Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 17 00:24:18.084341 containerd[1595]: time="2025-05-17T00:24:18.082890606Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"912dcaa7e1000d988a62cb53e266fac6d79debec318f0b90a403bbafb1f6da6e\"" May 17 00:24:18.084607 containerd[1595]: time="2025-05-17T00:24:18.084582984Z" level=info msg="StartContainer for \"912dcaa7e1000d988a62cb53e266fac6d79debec318f0b90a403bbafb1f6da6e\"" May 17 00:24:18.161883 containerd[1595]: time="2025-05-17T00:24:18.161825324Z" level=info msg="StartContainer for \"912dcaa7e1000d988a62cb53e266fac6d79debec318f0b90a403bbafb1f6da6e\" returns successfully" May 17 00:24:18.199611 containerd[1595]: time="2025-05-17T00:24:18.199487855Z" level=info msg="shim disconnected" id=912dcaa7e1000d988a62cb53e266fac6d79debec318f0b90a403bbafb1f6da6e namespace=k8s.io May 17 00:24:18.199611 containerd[1595]: time="2025-05-17T00:24:18.199561285Z" level=warning msg="cleaning up after shim disconnected" id=912dcaa7e1000d988a62cb53e266fac6d79debec318f0b90a403bbafb1f6da6e namespace=k8s.io May 17 00:24:18.199611 containerd[1595]: time="2025-05-17T00:24:18.199574541Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:18.735798 systemd[1]: run-containerd-runc-k8s.io-912dcaa7e1000d988a62cb53e266fac6d79debec318f0b90a403bbafb1f6da6e-runc.qliDzi.mount: Deactivated successfully. May 17 00:24:18.736091 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-912dcaa7e1000d988a62cb53e266fac6d79debec318f0b90a403bbafb1f6da6e-rootfs.mount: Deactivated successfully. May 17 00:24:19.053809 kubelet[2658]: E0517 00:24:19.053667 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:19.063490 containerd[1595]: time="2025-05-17T00:24:19.063206597Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 17 00:24:19.086742 containerd[1595]: time="2025-05-17T00:24:19.086680425Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"be3c7983eef6fbb5227234dcac9e503a7685d4ecdd4fee3fe2e08972c2242b24\"" May 17 00:24:19.087592 containerd[1595]: time="2025-05-17T00:24:19.087554799Z" level=info msg="StartContainer for \"be3c7983eef6fbb5227234dcac9e503a7685d4ecdd4fee3fe2e08972c2242b24\"" May 17 00:24:19.130799 systemd[1]: run-containerd-runc-k8s.io-be3c7983eef6fbb5227234dcac9e503a7685d4ecdd4fee3fe2e08972c2242b24-runc.s9Vqom.mount: Deactivated successfully. May 17 00:24:19.173186 containerd[1595]: time="2025-05-17T00:24:19.173121250Z" level=info msg="StartContainer for \"be3c7983eef6fbb5227234dcac9e503a7685d4ecdd4fee3fe2e08972c2242b24\" returns successfully" May 17 00:24:19.204211 containerd[1595]: time="2025-05-17T00:24:19.204128900Z" level=info msg="shim disconnected" id=be3c7983eef6fbb5227234dcac9e503a7685d4ecdd4fee3fe2e08972c2242b24 namespace=k8s.io May 17 00:24:19.204211 containerd[1595]: time="2025-05-17T00:24:19.204206960Z" level=warning msg="cleaning up after shim disconnected" id=be3c7983eef6fbb5227234dcac9e503a7685d4ecdd4fee3fe2e08972c2242b24 namespace=k8s.io May 17 00:24:19.204211 containerd[1595]: time="2025-05-17T00:24:19.204220155Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 17 00:24:19.736157 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-be3c7983eef6fbb5227234dcac9e503a7685d4ecdd4fee3fe2e08972c2242b24-rootfs.mount: Deactivated successfully. May 17 00:24:20.058763 kubelet[2658]: E0517 00:24:20.058139 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:20.065874 containerd[1595]: time="2025-05-17T00:24:20.065087679Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 17 00:24:20.090368 containerd[1595]: time="2025-05-17T00:24:20.090125271Z" level=info msg="CreateContainer within sandbox \"7185d0bf8143bfb352a9891a7b15a167d010f5fa986a69b3462001c66a66b823\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"edc0b730ddf186224e5b534e8dac6a7050cd5f47f73ca619145d364b595ccbb3\"" May 17 00:24:20.093423 containerd[1595]: time="2025-05-17T00:24:20.092518005Z" level=info msg="StartContainer for \"edc0b730ddf186224e5b534e8dac6a7050cd5f47f73ca619145d364b595ccbb3\"" May 17 00:24:20.176760 containerd[1595]: time="2025-05-17T00:24:20.176617990Z" level=info msg="StartContainer for \"edc0b730ddf186224e5b534e8dac6a7050cd5f47f73ca619145d364b595ccbb3\" returns successfully" May 17 00:24:20.705359 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) May 17 00:24:20.738191 systemd[1]: run-containerd-runc-k8s.io-edc0b730ddf186224e5b534e8dac6a7050cd5f47f73ca619145d364b595ccbb3-runc.dsAsFV.mount: Deactivated successfully. May 17 00:24:21.065386 kubelet[2658]: E0517 00:24:21.065334 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:22.068825 kubelet[2658]: E0517 00:24:22.068780 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:22.595559 systemd[1]: run-containerd-runc-k8s.io-edc0b730ddf186224e5b534e8dac6a7050cd5f47f73ca619145d364b595ccbb3-runc.vP4tUd.mount: Deactivated successfully. May 17 00:24:23.069484 kubelet[2658]: E0517 00:24:23.069451 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:23.975003 systemd-networkd[1219]: lxc_health: Link UP May 17 00:24:23.987629 systemd-networkd[1219]: lxc_health: Gained carrier May 17 00:24:25.843763 kubelet[2658]: E0517 00:24:25.843711 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:25.876199 kubelet[2658]: I0517 00:24:25.875566 2658 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-bpmf4" podStartSLOduration=10.87553679 podStartE2EDuration="10.87553679s" podCreationTimestamp="2025-05-17 00:24:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:24:21.085122683 +0000 UTC m=+105.826338089" watchObservedRunningTime="2025-05-17 00:24:25.87553679 +0000 UTC m=+110.616752198" May 17 00:24:25.943540 systemd-networkd[1219]: lxc_health: Gained IPv6LL May 17 00:24:26.079887 kubelet[2658]: E0517 00:24:26.079842 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:27.088433 kubelet[2658]: E0517 00:24:27.087241 2658 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.2 67.207.67.3 67.207.67.2" May 17 00:24:29.344786 sshd[4486]: pam_unix(sshd:session): session closed for user core May 17 00:24:29.349799 systemd[1]: sshd@28-64.23.167.227:22-139.178.68.195:48096.service: Deactivated successfully. May 17 00:24:29.360864 systemd[1]: session-28.scope: Deactivated successfully. May 17 00:24:29.362164 systemd-logind[1571]: Session 28 logged out. Waiting for processes to exit. May 17 00:24:29.365549 systemd-logind[1571]: Removed session 28.