Mar 25 01:28:18.947604 kernel: Linux version 6.6.83-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Mon Mar 24 23:38:35 -00 2025 Mar 25 01:28:18.948718 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=e7a00b7ee8d97e8d255663e9d3fa92277da8316702fb7f6d664fd7b137c307e9 Mar 25 01:28:18.948740 kernel: BIOS-provided physical RAM map: Mar 25 01:28:18.948748 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Mar 25 01:28:18.948754 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Mar 25 01:28:18.948761 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Mar 25 01:28:18.948778 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Mar 25 01:28:18.948786 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Mar 25 01:28:18.948792 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 25 01:28:18.948799 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Mar 25 01:28:18.948809 kernel: NX (Execute Disable) protection: active Mar 25 01:28:18.948822 kernel: APIC: Static calls initialized Mar 25 01:28:18.948829 kernel: SMBIOS 2.8 present. Mar 25 01:28:18.948837 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Mar 25 01:28:18.948845 kernel: Hypervisor detected: KVM Mar 25 01:28:18.948853 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 25 01:28:18.948869 kernel: kvm-clock: using sched offset of 3790502699 cycles Mar 25 01:28:18.948885 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 25 01:28:18.948896 kernel: tsc: Detected 2494.134 MHz processor Mar 25 01:28:18.948908 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 25 01:28:18.948920 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 25 01:28:18.948928 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Mar 25 01:28:18.948936 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Mar 25 01:28:18.948944 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 25 01:28:18.948953 kernel: ACPI: Early table checksum verification disabled Mar 25 01:28:18.948964 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Mar 25 01:28:18.948973 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 25 01:28:18.948981 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 25 01:28:18.948989 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 25 01:28:18.948997 kernel: ACPI: FACS 0x000000007FFE0000 000040 Mar 25 01:28:18.949005 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 25 01:28:18.949013 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 25 01:28:18.949021 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 25 01:28:18.949031 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 25 01:28:18.949039 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Mar 25 01:28:18.949047 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Mar 25 01:28:18.949055 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Mar 25 01:28:18.949063 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Mar 25 01:28:18.949071 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Mar 25 01:28:18.949079 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Mar 25 01:28:18.949091 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Mar 25 01:28:18.949102 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Mar 25 01:28:18.949113 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Mar 25 01:28:18.949122 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Mar 25 01:28:18.949145 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Mar 25 01:28:18.949154 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Mar 25 01:28:18.949162 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Mar 25 01:28:18.949171 kernel: Zone ranges: Mar 25 01:28:18.949182 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 25 01:28:18.949191 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Mar 25 01:28:18.949202 kernel: Normal empty Mar 25 01:28:18.949215 kernel: Movable zone start for each node Mar 25 01:28:18.949229 kernel: Early memory node ranges Mar 25 01:28:18.949241 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Mar 25 01:28:18.949254 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Mar 25 01:28:18.949267 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Mar 25 01:28:18.949275 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 25 01:28:18.949288 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Mar 25 01:28:18.949299 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Mar 25 01:28:18.949308 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 25 01:28:18.949316 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 25 01:28:18.949325 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 25 01:28:18.949339 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 25 01:28:18.949351 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 25 01:28:18.949363 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 25 01:28:18.949375 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 25 01:28:18.949393 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 25 01:28:18.949405 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 25 01:28:18.949416 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 25 01:28:18.949428 kernel: TSC deadline timer available Mar 25 01:28:18.949439 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Mar 25 01:28:18.949453 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 25 01:28:18.949466 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Mar 25 01:28:18.949479 kernel: Booting paravirtualized kernel on KVM Mar 25 01:28:18.949490 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 25 01:28:18.949510 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Mar 25 01:28:18.949525 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Mar 25 01:28:18.949534 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Mar 25 01:28:18.949543 kernel: pcpu-alloc: [0] 0 1 Mar 25 01:28:18.949551 kernel: kvm-guest: PV spinlocks disabled, no host support Mar 25 01:28:18.949562 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=e7a00b7ee8d97e8d255663e9d3fa92277da8316702fb7f6d664fd7b137c307e9 Mar 25 01:28:18.949570 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Mar 25 01:28:18.949579 kernel: random: crng init done Mar 25 01:28:18.949590 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 25 01:28:18.949599 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Mar 25 01:28:18.949608 kernel: Fallback order for Node 0: 0 Mar 25 01:28:18.949616 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Mar 25 01:28:18.949624 kernel: Policy zone: DMA32 Mar 25 01:28:18.950693 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 25 01:28:18.950705 kernel: Memory: 1967108K/2096612K available (14336K kernel code, 2304K rwdata, 25060K rodata, 43592K init, 1472K bss, 129244K reserved, 0K cma-reserved) Mar 25 01:28:18.950713 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 25 01:28:18.950722 kernel: Kernel/User page tables isolation: enabled Mar 25 01:28:18.950738 kernel: ftrace: allocating 37985 entries in 149 pages Mar 25 01:28:18.950750 kernel: ftrace: allocated 149 pages with 4 groups Mar 25 01:28:18.950764 kernel: Dynamic Preempt: voluntary Mar 25 01:28:18.950777 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 25 01:28:18.950790 kernel: rcu: RCU event tracing is enabled. Mar 25 01:28:18.950802 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 25 01:28:18.950815 kernel: Trampoline variant of Tasks RCU enabled. Mar 25 01:28:18.950828 kernel: Rude variant of Tasks RCU enabled. Mar 25 01:28:18.950841 kernel: Tracing variant of Tasks RCU enabled. Mar 25 01:28:18.950855 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 25 01:28:18.950863 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 25 01:28:18.950872 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Mar 25 01:28:18.950887 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 25 01:28:18.950895 kernel: Console: colour VGA+ 80x25 Mar 25 01:28:18.950904 kernel: printk: console [tty0] enabled Mar 25 01:28:18.950912 kernel: printk: console [ttyS0] enabled Mar 25 01:28:18.950921 kernel: ACPI: Core revision 20230628 Mar 25 01:28:18.950930 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 25 01:28:18.950941 kernel: APIC: Switch to symmetric I/O mode setup Mar 25 01:28:18.950951 kernel: x2apic enabled Mar 25 01:28:18.950963 kernel: APIC: Switched APIC routing to: physical x2apic Mar 25 01:28:18.950973 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 25 01:28:18.950982 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f3946f721, max_idle_ns: 440795294991 ns Mar 25 01:28:18.950990 kernel: Calibrating delay loop (skipped) preset value.. 4988.26 BogoMIPS (lpj=2494134) Mar 25 01:28:18.950999 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Mar 25 01:28:18.951008 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Mar 25 01:28:18.951030 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 25 01:28:18.951041 kernel: Spectre V2 : Mitigation: Retpolines Mar 25 01:28:18.951056 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Mar 25 01:28:18.951070 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Mar 25 01:28:18.951086 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Mar 25 01:28:18.951101 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Mar 25 01:28:18.951113 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Mar 25 01:28:18.951122 kernel: MDS: Mitigation: Clear CPU buffers Mar 25 01:28:18.951131 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Mar 25 01:28:18.951147 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 25 01:28:18.951156 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 25 01:28:18.951165 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 25 01:28:18.951174 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 25 01:28:18.951183 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Mar 25 01:28:18.951192 kernel: Freeing SMP alternatives memory: 32K Mar 25 01:28:18.951201 kernel: pid_max: default: 32768 minimum: 301 Mar 25 01:28:18.951209 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 25 01:28:18.951222 kernel: landlock: Up and running. Mar 25 01:28:18.951230 kernel: SELinux: Initializing. Mar 25 01:28:18.951239 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Mar 25 01:28:18.951248 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Mar 25 01:28:18.951259 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Mar 25 01:28:18.951272 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 25 01:28:18.951280 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 25 01:28:18.951289 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 25 01:28:18.951299 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Mar 25 01:28:18.951310 kernel: signal: max sigframe size: 1776 Mar 25 01:28:18.951319 kernel: rcu: Hierarchical SRCU implementation. Mar 25 01:28:18.951328 kernel: rcu: Max phase no-delay instances is 400. Mar 25 01:28:18.951337 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 25 01:28:18.951346 kernel: smp: Bringing up secondary CPUs ... Mar 25 01:28:18.951355 kernel: smpboot: x86: Booting SMP configuration: Mar 25 01:28:18.951428 kernel: .... node #0, CPUs: #1 Mar 25 01:28:18.951447 kernel: smp: Brought up 1 node, 2 CPUs Mar 25 01:28:18.951460 kernel: smpboot: Max logical packages: 1 Mar 25 01:28:18.951478 kernel: smpboot: Total of 2 processors activated (9976.53 BogoMIPS) Mar 25 01:28:18.951491 kernel: devtmpfs: initialized Mar 25 01:28:18.951504 kernel: x86/mm: Memory block size: 128MB Mar 25 01:28:18.951517 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 25 01:28:18.951531 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 25 01:28:18.951544 kernel: pinctrl core: initialized pinctrl subsystem Mar 25 01:28:18.951553 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 25 01:28:18.951562 kernel: audit: initializing netlink subsys (disabled) Mar 25 01:28:18.951572 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 25 01:28:18.951593 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 25 01:28:18.951623 kernel: audit: type=2000 audit(1742866097.399:1): state=initialized audit_enabled=0 res=1 Mar 25 01:28:18.954732 kernel: cpuidle: using governor menu Mar 25 01:28:18.954745 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 25 01:28:18.954754 kernel: dca service started, version 1.12.1 Mar 25 01:28:18.954764 kernel: PCI: Using configuration type 1 for base access Mar 25 01:28:18.954773 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 25 01:28:18.954782 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 25 01:28:18.954791 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 25 01:28:18.954811 kernel: ACPI: Added _OSI(Module Device) Mar 25 01:28:18.954820 kernel: ACPI: Added _OSI(Processor Device) Mar 25 01:28:18.954830 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Mar 25 01:28:18.954839 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 25 01:28:18.954847 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 25 01:28:18.954861 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Mar 25 01:28:18.954874 kernel: ACPI: Interpreter enabled Mar 25 01:28:18.954887 kernel: ACPI: PM: (supports S0 S5) Mar 25 01:28:18.954900 kernel: ACPI: Using IOAPIC for interrupt routing Mar 25 01:28:18.954914 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 25 01:28:18.954927 kernel: PCI: Using E820 reservations for host bridge windows Mar 25 01:28:18.954936 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Mar 25 01:28:18.954945 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 25 01:28:18.955206 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Mar 25 01:28:18.955323 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Mar 25 01:28:18.955488 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Mar 25 01:28:18.955515 kernel: acpiphp: Slot [3] registered Mar 25 01:28:18.955529 kernel: acpiphp: Slot [4] registered Mar 25 01:28:18.955543 kernel: acpiphp: Slot [5] registered Mar 25 01:28:18.955557 kernel: acpiphp: Slot [6] registered Mar 25 01:28:18.955571 kernel: acpiphp: Slot [7] registered Mar 25 01:28:18.955585 kernel: acpiphp: Slot [8] registered Mar 25 01:28:18.955600 kernel: acpiphp: Slot [9] registered Mar 25 01:28:18.955615 kernel: acpiphp: Slot [10] registered Mar 25 01:28:18.955647 kernel: acpiphp: Slot [11] registered Mar 25 01:28:18.955661 kernel: acpiphp: Slot [12] registered Mar 25 01:28:18.955681 kernel: acpiphp: Slot [13] registered Mar 25 01:28:18.955695 kernel: acpiphp: Slot [14] registered Mar 25 01:28:18.955710 kernel: acpiphp: Slot [15] registered Mar 25 01:28:18.955725 kernel: acpiphp: Slot [16] registered Mar 25 01:28:18.955740 kernel: acpiphp: Slot [17] registered Mar 25 01:28:18.955752 kernel: acpiphp: Slot [18] registered Mar 25 01:28:18.955764 kernel: acpiphp: Slot [19] registered Mar 25 01:28:18.955776 kernel: acpiphp: Slot [20] registered Mar 25 01:28:18.955788 kernel: acpiphp: Slot [21] registered Mar 25 01:28:18.955803 kernel: acpiphp: Slot [22] registered Mar 25 01:28:18.955815 kernel: acpiphp: Slot [23] registered Mar 25 01:28:18.955828 kernel: acpiphp: Slot [24] registered Mar 25 01:28:18.955839 kernel: acpiphp: Slot [25] registered Mar 25 01:28:18.955851 kernel: acpiphp: Slot [26] registered Mar 25 01:28:18.955863 kernel: acpiphp: Slot [27] registered Mar 25 01:28:18.955874 kernel: acpiphp: Slot [28] registered Mar 25 01:28:18.955886 kernel: acpiphp: Slot [29] registered Mar 25 01:28:18.955898 kernel: acpiphp: Slot [30] registered Mar 25 01:28:18.955912 kernel: acpiphp: Slot [31] registered Mar 25 01:28:18.955929 kernel: PCI host bridge to bus 0000:00 Mar 25 01:28:18.956101 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 25 01:28:18.956195 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 25 01:28:18.956305 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 25 01:28:18.956423 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Mar 25 01:28:18.956524 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Mar 25 01:28:18.956610 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 25 01:28:18.956818 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Mar 25 01:28:18.956936 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Mar 25 01:28:18.957059 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Mar 25 01:28:18.957155 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Mar 25 01:28:18.957250 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Mar 25 01:28:18.957352 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Mar 25 01:28:18.957474 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Mar 25 01:28:18.957595 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Mar 25 01:28:18.957734 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Mar 25 01:28:18.957855 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Mar 25 01:28:18.957967 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Mar 25 01:28:18.958065 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Mar 25 01:28:18.958161 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Mar 25 01:28:18.958283 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Mar 25 01:28:18.958401 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Mar 25 01:28:18.958517 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Mar 25 01:28:18.961042 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Mar 25 01:28:18.961245 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Mar 25 01:28:18.961362 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 25 01:28:18.961501 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Mar 25 01:28:18.961602 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Mar 25 01:28:18.961797 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Mar 25 01:28:18.961910 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Mar 25 01:28:18.962029 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Mar 25 01:28:18.962153 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Mar 25 01:28:18.962251 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Mar 25 01:28:18.962363 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Mar 25 01:28:18.962475 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Mar 25 01:28:18.962612 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Mar 25 01:28:18.964479 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Mar 25 01:28:18.964609 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Mar 25 01:28:18.964802 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Mar 25 01:28:18.964937 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Mar 25 01:28:18.965097 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Mar 25 01:28:18.965217 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Mar 25 01:28:18.965441 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Mar 25 01:28:18.965554 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Mar 25 01:28:18.967722 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Mar 25 01:28:18.967874 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Mar 25 01:28:18.967999 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Mar 25 01:28:18.968111 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Mar 25 01:28:18.968208 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Mar 25 01:28:18.968221 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 25 01:28:18.968230 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 25 01:28:18.968239 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 25 01:28:18.968248 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 25 01:28:18.968257 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Mar 25 01:28:18.968270 kernel: iommu: Default domain type: Translated Mar 25 01:28:18.968279 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 25 01:28:18.968288 kernel: PCI: Using ACPI for IRQ routing Mar 25 01:28:18.968297 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 25 01:28:18.968305 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Mar 25 01:28:18.968314 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Mar 25 01:28:18.968411 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Mar 25 01:28:18.968509 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Mar 25 01:28:18.969717 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 25 01:28:18.969767 kernel: vgaarb: loaded Mar 25 01:28:18.969781 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 25 01:28:18.969790 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 25 01:28:18.969799 kernel: clocksource: Switched to clocksource kvm-clock Mar 25 01:28:18.969808 kernel: VFS: Disk quotas dquot_6.6.0 Mar 25 01:28:18.969818 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 25 01:28:18.969827 kernel: pnp: PnP ACPI init Mar 25 01:28:18.969836 kernel: pnp: PnP ACPI: found 4 devices Mar 25 01:28:18.969845 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 25 01:28:18.969858 kernel: NET: Registered PF_INET protocol family Mar 25 01:28:18.969867 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 25 01:28:18.969876 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Mar 25 01:28:18.969885 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 25 01:28:18.969894 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Mar 25 01:28:18.969903 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Mar 25 01:28:18.969911 kernel: TCP: Hash tables configured (established 16384 bind 16384) Mar 25 01:28:18.969921 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Mar 25 01:28:18.969932 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Mar 25 01:28:18.969941 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 25 01:28:18.969950 kernel: NET: Registered PF_XDP protocol family Mar 25 01:28:18.970088 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 25 01:28:18.970211 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 25 01:28:18.970299 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 25 01:28:18.970385 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Mar 25 01:28:18.970495 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Mar 25 01:28:18.971720 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Mar 25 01:28:18.971898 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Mar 25 01:28:18.971915 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Mar 25 01:28:18.972022 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x7a0 took 32254 usecs Mar 25 01:28:18.972035 kernel: PCI: CLS 0 bytes, default 64 Mar 25 01:28:18.972045 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Mar 25 01:28:18.972054 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f3946f721, max_idle_ns: 440795294991 ns Mar 25 01:28:18.972063 kernel: Initialise system trusted keyrings Mar 25 01:28:18.972073 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Mar 25 01:28:18.972087 kernel: Key type asymmetric registered Mar 25 01:28:18.972096 kernel: Asymmetric key parser 'x509' registered Mar 25 01:28:18.972106 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Mar 25 01:28:18.972115 kernel: io scheduler mq-deadline registered Mar 25 01:28:18.972124 kernel: io scheduler kyber registered Mar 25 01:28:18.972133 kernel: io scheduler bfq registered Mar 25 01:28:18.972142 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 25 01:28:18.972151 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Mar 25 01:28:18.972160 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Mar 25 01:28:18.972172 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Mar 25 01:28:18.972181 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 25 01:28:18.972190 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 25 01:28:18.972199 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 25 01:28:18.972208 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 25 01:28:18.972216 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 25 01:28:18.972350 kernel: rtc_cmos 00:03: RTC can wake from S4 Mar 25 01:28:18.972386 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 25 01:28:18.972484 kernel: rtc_cmos 00:03: registered as rtc0 Mar 25 01:28:18.972575 kernel: rtc_cmos 00:03: setting system clock to 2025-03-25T01:28:18 UTC (1742866098) Mar 25 01:28:18.973750 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Mar 25 01:28:18.973781 kernel: intel_pstate: CPU model not supported Mar 25 01:28:18.973794 kernel: NET: Registered PF_INET6 protocol family Mar 25 01:28:18.973806 kernel: Segment Routing with IPv6 Mar 25 01:28:18.973818 kernel: In-situ OAM (IOAM) with IPv6 Mar 25 01:28:18.973832 kernel: NET: Registered PF_PACKET protocol family Mar 25 01:28:18.973846 kernel: Key type dns_resolver registered Mar 25 01:28:18.973867 kernel: IPI shorthand broadcast: enabled Mar 25 01:28:18.973881 kernel: sched_clock: Marking stable (927005866, 118318953)->(1170091462, -124766643) Mar 25 01:28:18.973901 kernel: registered taskstats version 1 Mar 25 01:28:18.973911 kernel: Loading compiled-in X.509 certificates Mar 25 01:28:18.973920 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.83-flatcar: eff01054e94a599f8e404b9a9482f4e2220f5386' Mar 25 01:28:18.973929 kernel: Key type .fscrypt registered Mar 25 01:28:18.973938 kernel: Key type fscrypt-provisioning registered Mar 25 01:28:18.973947 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 25 01:28:18.973959 kernel: ima: Allocated hash algorithm: sha1 Mar 25 01:28:18.973968 kernel: ima: No architecture policies found Mar 25 01:28:18.973977 kernel: clk: Disabling unused clocks Mar 25 01:28:18.973986 kernel: Freeing unused kernel image (initmem) memory: 43592K Mar 25 01:28:18.973996 kernel: Write protecting the kernel read-only data: 40960k Mar 25 01:28:18.974027 kernel: Freeing unused kernel image (rodata/data gap) memory: 1564K Mar 25 01:28:18.974047 kernel: Run /init as init process Mar 25 01:28:18.974060 kernel: with arguments: Mar 25 01:28:18.974075 kernel: /init Mar 25 01:28:18.974090 kernel: with environment: Mar 25 01:28:18.974109 kernel: HOME=/ Mar 25 01:28:18.974123 kernel: TERM=linux Mar 25 01:28:18.974134 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Mar 25 01:28:18.974145 systemd[1]: Successfully made /usr/ read-only. Mar 25 01:28:18.974159 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 25 01:28:18.974169 systemd[1]: Detected virtualization kvm. Mar 25 01:28:18.974179 systemd[1]: Detected architecture x86-64. Mar 25 01:28:18.974191 systemd[1]: Running in initrd. Mar 25 01:28:18.974200 systemd[1]: No hostname configured, using default hostname. Mar 25 01:28:18.974210 systemd[1]: Hostname set to . Mar 25 01:28:18.974219 systemd[1]: Initializing machine ID from VM UUID. Mar 25 01:28:18.974229 systemd[1]: Queued start job for default target initrd.target. Mar 25 01:28:18.974239 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 25 01:28:18.974248 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 25 01:28:18.974260 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 25 01:28:18.974272 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 25 01:28:18.974283 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 25 01:28:18.974298 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 25 01:28:18.974315 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 25 01:28:18.974331 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 25 01:28:18.974344 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 25 01:28:18.974353 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 25 01:28:18.974366 systemd[1]: Reached target paths.target - Path Units. Mar 25 01:28:18.974376 systemd[1]: Reached target slices.target - Slice Units. Mar 25 01:28:18.974389 systemd[1]: Reached target swap.target - Swaps. Mar 25 01:28:18.974401 systemd[1]: Reached target timers.target - Timer Units. Mar 25 01:28:18.974411 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 25 01:28:18.974423 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 25 01:28:18.974435 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 25 01:28:18.974451 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 25 01:28:18.974464 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 25 01:28:18.974480 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 25 01:28:18.974495 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 25 01:28:18.974511 systemd[1]: Reached target sockets.target - Socket Units. Mar 25 01:28:18.974527 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 25 01:28:18.974541 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 25 01:28:18.974562 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 25 01:28:18.974577 systemd[1]: Starting systemd-fsck-usr.service... Mar 25 01:28:18.974591 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 25 01:28:18.974607 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 25 01:28:18.974617 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 25 01:28:18.975684 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 25 01:28:18.975715 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 25 01:28:18.975734 systemd[1]: Finished systemd-fsck-usr.service. Mar 25 01:28:18.975803 systemd-journald[181]: Collecting audit messages is disabled. Mar 25 01:28:18.975832 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 25 01:28:18.975844 systemd-journald[181]: Journal started Mar 25 01:28:18.975866 systemd-journald[181]: Runtime Journal (/run/log/journal/98ef30a558654dd4b4bca9ef930af956) is 4.9M, max 39.3M, 34.3M free. Mar 25 01:28:18.976725 systemd[1]: Started systemd-journald.service - Journal Service. Mar 25 01:28:18.966089 systemd-modules-load[183]: Inserted module 'overlay' Mar 25 01:28:18.991337 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 25 01:28:18.997841 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 25 01:28:19.009756 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 25 01:28:19.008921 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 25 01:28:19.010834 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 25 01:28:19.013660 kernel: Bridge firewalling registered Mar 25 01:28:19.014811 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 25 01:28:19.015722 systemd-modules-load[183]: Inserted module 'br_netfilter' Mar 25 01:28:19.026700 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 25 01:28:19.035834 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 25 01:28:19.043391 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 25 01:28:19.045599 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 25 01:28:19.053196 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 25 01:28:19.055762 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 25 01:28:19.071538 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 25 01:28:19.076995 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 25 01:28:19.088319 dracut-cmdline[216]: dracut-dracut-053 Mar 25 01:28:19.094681 dracut-cmdline[216]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=e7a00b7ee8d97e8d255663e9d3fa92277da8316702fb7f6d664fd7b137c307e9 Mar 25 01:28:19.125193 systemd-resolved[219]: Positive Trust Anchors: Mar 25 01:28:19.125216 systemd-resolved[219]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 25 01:28:19.125255 systemd-resolved[219]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 25 01:28:19.128738 systemd-resolved[219]: Defaulting to hostname 'linux'. Mar 25 01:28:19.130291 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 25 01:28:19.130787 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 25 01:28:19.223691 kernel: SCSI subsystem initialized Mar 25 01:28:19.235762 kernel: Loading iSCSI transport class v2.0-870. Mar 25 01:28:19.247699 kernel: iscsi: registered transport (tcp) Mar 25 01:28:19.272807 kernel: iscsi: registered transport (qla4xxx) Mar 25 01:28:19.272928 kernel: QLogic iSCSI HBA Driver Mar 25 01:28:19.338317 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 25 01:28:19.342156 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 25 01:28:19.383729 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 25 01:28:19.383837 kernel: device-mapper: uevent: version 1.0.3 Mar 25 01:28:19.384747 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 25 01:28:19.439731 kernel: raid6: avx2x4 gen() 18073 MB/s Mar 25 01:28:19.454707 kernel: raid6: avx2x2 gen() 20109 MB/s Mar 25 01:28:19.471963 kernel: raid6: avx2x1 gen() 19747 MB/s Mar 25 01:28:19.472072 kernel: raid6: using algorithm avx2x2 gen() 20109 MB/s Mar 25 01:28:19.489853 kernel: raid6: .... xor() 17151 MB/s, rmw enabled Mar 25 01:28:19.489972 kernel: raid6: using avx2x2 recovery algorithm Mar 25 01:28:19.513695 kernel: xor: automatically using best checksumming function avx Mar 25 01:28:19.692696 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 25 01:28:19.708278 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 25 01:28:19.710717 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 25 01:28:19.742681 systemd-udevd[402]: Using default interface naming scheme 'v255'. Mar 25 01:28:19.750038 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 25 01:28:19.755547 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 25 01:28:19.790136 dracut-pre-trigger[409]: rd.md=0: removing MD RAID activation Mar 25 01:28:19.836836 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 25 01:28:19.840125 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 25 01:28:19.925709 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 25 01:28:19.929801 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 25 01:28:19.966486 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 25 01:28:19.969295 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 25 01:28:19.970743 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 25 01:28:19.972232 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 25 01:28:19.975870 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 25 01:28:20.011504 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 25 01:28:20.032684 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Mar 25 01:28:20.068606 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Mar 25 01:28:20.072291 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 25 01:28:20.072326 kernel: GPT:9289727 != 125829119 Mar 25 01:28:20.072838 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 25 01:28:20.072882 kernel: GPT:9289727 != 125829119 Mar 25 01:28:20.072902 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 25 01:28:20.072919 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 25 01:28:20.072949 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Mar 25 01:28:20.073229 kernel: virtio_blk virtio5: [vdb] 976 512-byte logical blocks (500 kB/488 KiB) Mar 25 01:28:20.078541 kernel: scsi host0: Virtio SCSI HBA Mar 25 01:28:20.090739 kernel: cryptd: max_cpu_qlen set to 1000 Mar 25 01:28:20.153127 kernel: AVX2 version of gcm_enc/dec engaged. Mar 25 01:28:20.153208 kernel: AES CTR mode by8 optimization enabled Mar 25 01:28:20.155665 kernel: libata version 3.00 loaded. Mar 25 01:28:20.165738 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 25 01:28:20.184920 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 25 01:28:20.186883 kernel: ata_piix 0000:00:01.1: version 2.13 Mar 25 01:28:20.200289 kernel: scsi host1: ata_piix Mar 25 01:28:20.201505 kernel: scsi host2: ata_piix Mar 25 01:28:20.205467 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Mar 25 01:28:20.205498 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Mar 25 01:28:20.205516 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by (udev-worker) (448) Mar 25 01:28:20.205535 kernel: BTRFS: device fsid 6d9424cd-1432-492b-b006-b311869817e2 devid 1 transid 39 /dev/vda3 scanned by (udev-worker) (463) Mar 25 01:28:20.185122 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 25 01:28:20.188004 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 25 01:28:20.188709 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 25 01:28:20.188978 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 25 01:28:20.190528 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 25 01:28:20.216186 kernel: ACPI: bus type USB registered Mar 25 01:28:20.216214 kernel: usbcore: registered new interface driver usbfs Mar 25 01:28:20.195064 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 25 01:28:20.195993 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 25 01:28:20.232779 kernel: usbcore: registered new interface driver hub Mar 25 01:28:20.237678 kernel: usbcore: registered new device driver usb Mar 25 01:28:20.239049 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 25 01:28:20.265677 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 25 01:28:20.275238 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 25 01:28:20.275944 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 25 01:28:20.286277 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 25 01:28:20.287987 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 25 01:28:20.289799 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 25 01:28:20.308244 disk-uuid[540]: Primary Header is updated. Mar 25 01:28:20.308244 disk-uuid[540]: Secondary Entries is updated. Mar 25 01:28:20.308244 disk-uuid[540]: Secondary Header is updated. Mar 25 01:28:20.321677 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 25 01:28:20.327528 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 25 01:28:20.472984 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Mar 25 01:28:20.483915 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Mar 25 01:28:20.484177 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Mar 25 01:28:20.484394 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Mar 25 01:28:20.484602 kernel: hub 1-0:1.0: USB hub found Mar 25 01:28:20.484878 kernel: hub 1-0:1.0: 2 ports detected Mar 25 01:28:21.333690 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 25 01:28:21.334515 disk-uuid[541]: The operation has completed successfully. Mar 25 01:28:21.386514 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 25 01:28:21.386701 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 25 01:28:21.443667 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 25 01:28:21.458942 sh[561]: Success Mar 25 01:28:21.473699 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Mar 25 01:28:21.543034 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 25 01:28:21.547777 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 25 01:28:21.564932 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 25 01:28:21.584249 kernel: BTRFS info (device dm-0): first mount of filesystem 6d9424cd-1432-492b-b006-b311869817e2 Mar 25 01:28:21.584357 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 25 01:28:21.584379 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 25 01:28:21.586014 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 25 01:28:21.586088 kernel: BTRFS info (device dm-0): using free space tree Mar 25 01:28:21.596061 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 25 01:28:21.597161 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 25 01:28:21.599876 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 25 01:28:21.602873 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 25 01:28:21.640813 kernel: BTRFS info (device vda6): first mount of filesystem a72930ba-1354-475c-94df-b83a66efea67 Mar 25 01:28:21.640888 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 25 01:28:21.641998 kernel: BTRFS info (device vda6): using free space tree Mar 25 01:28:21.651749 kernel: BTRFS info (device vda6): auto enabling async discard Mar 25 01:28:21.658682 kernel: BTRFS info (device vda6): last unmount of filesystem a72930ba-1354-475c-94df-b83a66efea67 Mar 25 01:28:21.663145 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 25 01:28:21.668612 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 25 01:28:21.760772 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 25 01:28:21.764826 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 25 01:28:21.813596 systemd-networkd[742]: lo: Link UP Mar 25 01:28:21.813611 systemd-networkd[742]: lo: Gained carrier Mar 25 01:28:21.816539 systemd-networkd[742]: Enumeration completed Mar 25 01:28:21.817019 systemd-networkd[742]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Mar 25 01:28:21.817026 systemd-networkd[742]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Mar 25 01:28:21.817884 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 25 01:28:21.818450 systemd[1]: Reached target network.target - Network. Mar 25 01:28:21.820327 systemd-networkd[742]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 25 01:28:21.820332 systemd-networkd[742]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 25 01:28:21.821023 systemd-networkd[742]: eth0: Link UP Mar 25 01:28:21.821028 systemd-networkd[742]: eth0: Gained carrier Mar 25 01:28:21.821039 systemd-networkd[742]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Mar 25 01:28:21.826094 systemd-networkd[742]: eth1: Link UP Mar 25 01:28:21.826101 systemd-networkd[742]: eth1: Gained carrier Mar 25 01:28:21.826120 systemd-networkd[742]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 25 01:28:21.836623 ignition[660]: Ignition 2.20.0 Mar 25 01:28:21.836789 ignition[660]: Stage: fetch-offline Mar 25 01:28:21.836835 ignition[660]: no configs at "/usr/lib/ignition/base.d" Mar 25 01:28:21.838307 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 25 01:28:21.836848 ignition[660]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Mar 25 01:28:21.836999 ignition[660]: parsed url from cmdline: "" Mar 25 01:28:21.837005 ignition[660]: no config URL provided Mar 25 01:28:21.837011 ignition[660]: reading system config file "/usr/lib/ignition/user.ign" Mar 25 01:28:21.837020 ignition[660]: no config at "/usr/lib/ignition/user.ign" Mar 25 01:28:21.840740 systemd-networkd[742]: eth0: DHCPv4 address 143.198.155.16/20, gateway 143.198.144.1 acquired from 169.254.169.253 Mar 25 01:28:21.837027 ignition[660]: failed to fetch config: resource requires networking Mar 25 01:28:21.841861 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 25 01:28:21.837239 ignition[660]: Ignition finished successfully Mar 25 01:28:21.844808 systemd-networkd[742]: eth1: DHCPv4 address 10.124.0.18/20 acquired from 169.254.169.253 Mar 25 01:28:21.867741 ignition[750]: Ignition 2.20.0 Mar 25 01:28:21.867756 ignition[750]: Stage: fetch Mar 25 01:28:21.868012 ignition[750]: no configs at "/usr/lib/ignition/base.d" Mar 25 01:28:21.868030 ignition[750]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Mar 25 01:28:21.868165 ignition[750]: parsed url from cmdline: "" Mar 25 01:28:21.868170 ignition[750]: no config URL provided Mar 25 01:28:21.868179 ignition[750]: reading system config file "/usr/lib/ignition/user.ign" Mar 25 01:28:21.868191 ignition[750]: no config at "/usr/lib/ignition/user.ign" Mar 25 01:28:21.868223 ignition[750]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Mar 25 01:28:21.884349 ignition[750]: GET result: OK Mar 25 01:28:21.884465 ignition[750]: parsing config with SHA512: 5cf1c09860c6f989533a763342096b148e8f9154405562f5c41827e98be97b5a73945f9c1f1835d359a2afc9d81bd85e44fa00e25566693d57efd31924368c87 Mar 25 01:28:21.891983 unknown[750]: fetched base config from "system" Mar 25 01:28:21.892456 ignition[750]: fetch: fetch complete Mar 25 01:28:21.892000 unknown[750]: fetched base config from "system" Mar 25 01:28:21.892462 ignition[750]: fetch: fetch passed Mar 25 01:28:21.892009 unknown[750]: fetched user config from "digitalocean" Mar 25 01:28:21.892533 ignition[750]: Ignition finished successfully Mar 25 01:28:21.896246 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 25 01:28:21.897866 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 25 01:28:21.936266 ignition[758]: Ignition 2.20.0 Mar 25 01:28:21.936278 ignition[758]: Stage: kargs Mar 25 01:28:21.936517 ignition[758]: no configs at "/usr/lib/ignition/base.d" Mar 25 01:28:21.936530 ignition[758]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Mar 25 01:28:21.937617 ignition[758]: kargs: kargs passed Mar 25 01:28:21.938794 ignition[758]: Ignition finished successfully Mar 25 01:28:21.940341 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 25 01:28:21.943216 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 25 01:28:21.978856 ignition[765]: Ignition 2.20.0 Mar 25 01:28:21.978866 ignition[765]: Stage: disks Mar 25 01:28:21.979131 ignition[765]: no configs at "/usr/lib/ignition/base.d" Mar 25 01:28:21.979143 ignition[765]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Mar 25 01:28:21.983060 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 25 01:28:21.980277 ignition[765]: disks: disks passed Mar 25 01:28:21.980337 ignition[765]: Ignition finished successfully Mar 25 01:28:21.986035 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 25 01:28:21.986916 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 25 01:28:21.987865 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 25 01:28:21.988805 systemd[1]: Reached target sysinit.target - System Initialization. Mar 25 01:28:21.989482 systemd[1]: Reached target basic.target - Basic System. Mar 25 01:28:21.991538 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 25 01:28:22.023731 systemd-fsck[774]: ROOT: clean, 14/553520 files, 52654/553472 blocks Mar 25 01:28:22.027994 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 25 01:28:22.030924 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 25 01:28:22.148702 kernel: EXT4-fs (vda9): mounted filesystem 4e6dca82-2e50-453c-be25-61f944b72008 r/w with ordered data mode. Quota mode: none. Mar 25 01:28:22.149110 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 25 01:28:22.150066 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 25 01:28:22.152433 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 25 01:28:22.156795 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 25 01:28:22.167914 systemd[1]: Starting flatcar-afterburn-network.service - Flatcar Afterburn network service... Mar 25 01:28:22.173377 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 25 01:28:22.175069 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 25 01:28:22.175996 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 25 01:28:22.180124 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 25 01:28:22.186936 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 25 01:28:22.196460 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (782) Mar 25 01:28:22.199662 kernel: BTRFS info (device vda6): first mount of filesystem a72930ba-1354-475c-94df-b83a66efea67 Mar 25 01:28:22.203024 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 25 01:28:22.203122 kernel: BTRFS info (device vda6): using free space tree Mar 25 01:28:22.209077 kernel: BTRFS info (device vda6): auto enabling async discard Mar 25 01:28:22.211886 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 25 01:28:22.275571 coreos-metadata[785]: Mar 25 01:28:22.275 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Mar 25 01:28:22.282677 initrd-setup-root[813]: cut: /sysroot/etc/passwd: No such file or directory Mar 25 01:28:22.290388 coreos-metadata[785]: Mar 25 01:28:22.288 INFO Fetch successful Mar 25 01:28:22.292917 coreos-metadata[784]: Mar 25 01:28:22.292 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Mar 25 01:28:22.295783 coreos-metadata[785]: Mar 25 01:28:22.295 INFO wrote hostname ci-4284.0.0-c-20cdc03f82 to /sysroot/etc/hostname Mar 25 01:28:22.297555 initrd-setup-root[820]: cut: /sysroot/etc/group: No such file or directory Mar 25 01:28:22.299706 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 25 01:28:22.304436 initrd-setup-root[828]: cut: /sysroot/etc/shadow: No such file or directory Mar 25 01:28:22.305708 coreos-metadata[784]: Mar 25 01:28:22.305 INFO Fetch successful Mar 25 01:28:22.313438 initrd-setup-root[835]: cut: /sysroot/etc/gshadow: No such file or directory Mar 25 01:28:22.315019 systemd[1]: flatcar-afterburn-network.service: Deactivated successfully. Mar 25 01:28:22.315856 systemd[1]: Finished flatcar-afterburn-network.service - Flatcar Afterburn network service. Mar 25 01:28:22.445466 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 25 01:28:22.448487 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 25 01:28:22.451850 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 25 01:28:22.475680 kernel: BTRFS info (device vda6): last unmount of filesystem a72930ba-1354-475c-94df-b83a66efea67 Mar 25 01:28:22.493079 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 25 01:28:22.514721 ignition[904]: INFO : Ignition 2.20.0 Mar 25 01:28:22.514721 ignition[904]: INFO : Stage: mount Mar 25 01:28:22.515982 ignition[904]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 25 01:28:22.515982 ignition[904]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Mar 25 01:28:22.516934 ignition[904]: INFO : mount: mount passed Mar 25 01:28:22.516934 ignition[904]: INFO : Ignition finished successfully Mar 25 01:28:22.518028 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 25 01:28:22.520775 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 25 01:28:22.583864 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 25 01:28:22.586361 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 25 01:28:22.615085 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/vda6 scanned by mount (915) Mar 25 01:28:22.615174 kernel: BTRFS info (device vda6): first mount of filesystem a72930ba-1354-475c-94df-b83a66efea67 Mar 25 01:28:22.616882 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 25 01:28:22.616979 kernel: BTRFS info (device vda6): using free space tree Mar 25 01:28:22.620661 kernel: BTRFS info (device vda6): auto enabling async discard Mar 25 01:28:22.624102 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 25 01:28:22.673980 ignition[931]: INFO : Ignition 2.20.0 Mar 25 01:28:22.673980 ignition[931]: INFO : Stage: files Mar 25 01:28:22.675074 ignition[931]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 25 01:28:22.675074 ignition[931]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Mar 25 01:28:22.676274 ignition[931]: DEBUG : files: compiled without relabeling support, skipping Mar 25 01:28:22.677534 ignition[931]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 25 01:28:22.677534 ignition[931]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 25 01:28:22.683079 ignition[931]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 25 01:28:22.684112 ignition[931]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 25 01:28:22.685010 unknown[931]: wrote ssh authorized keys file for user: core Mar 25 01:28:22.685742 ignition[931]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 25 01:28:22.687233 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Mar 25 01:28:22.688268 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Mar 25 01:28:22.728706 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 25 01:28:22.916566 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Mar 25 01:28:22.916566 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 25 01:28:22.916566 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Mar 25 01:28:23.401771 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 25 01:28:23.491846 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 25 01:28:23.492663 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 25 01:28:23.501084 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 25 01:28:23.501084 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Mar 25 01:28:23.501084 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Mar 25 01:28:23.501084 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Mar 25 01:28:23.501084 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 Mar 25 01:28:23.745816 systemd-networkd[742]: eth1: Gained IPv6LL Mar 25 01:28:23.769835 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 25 01:28:23.874080 systemd-networkd[742]: eth0: Gained IPv6LL Mar 25 01:28:24.078256 ignition[931]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Mar 25 01:28:24.079268 ignition[931]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Mar 25 01:28:24.081198 ignition[931]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 25 01:28:24.082917 ignition[931]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 25 01:28:24.082917 ignition[931]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Mar 25 01:28:24.082917 ignition[931]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" Mar 25 01:28:24.082917 ignition[931]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" Mar 25 01:28:24.082917 ignition[931]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 25 01:28:24.082917 ignition[931]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 25 01:28:24.082917 ignition[931]: INFO : files: files passed Mar 25 01:28:24.082917 ignition[931]: INFO : Ignition finished successfully Mar 25 01:28:24.083069 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 25 01:28:24.088825 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 25 01:28:24.092884 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 25 01:28:24.102823 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 25 01:28:24.102965 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 25 01:28:24.113963 initrd-setup-root-after-ignition[962]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 25 01:28:24.113963 initrd-setup-root-after-ignition[962]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 25 01:28:24.116911 initrd-setup-root-after-ignition[965]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 25 01:28:24.117621 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 25 01:28:24.119000 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 25 01:28:24.120603 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 25 01:28:24.182267 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 25 01:28:24.182421 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 25 01:28:24.184379 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 25 01:28:24.185009 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 25 01:28:24.185816 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 25 01:28:24.187246 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 25 01:28:24.228231 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 25 01:28:24.232109 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 25 01:28:24.261585 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 25 01:28:24.263463 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 25 01:28:24.264314 systemd[1]: Stopped target timers.target - Timer Units. Mar 25 01:28:24.265277 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 25 01:28:24.265534 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 25 01:28:24.267653 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 25 01:28:24.268451 systemd[1]: Stopped target basic.target - Basic System. Mar 25 01:28:24.269203 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 25 01:28:24.269940 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 25 01:28:24.270893 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 25 01:28:24.271923 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 25 01:28:24.272868 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 25 01:28:24.273658 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 25 01:28:24.274503 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 25 01:28:24.275522 systemd[1]: Stopped target swap.target - Swaps. Mar 25 01:28:24.276151 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 25 01:28:24.276392 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 25 01:28:24.277864 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 25 01:28:24.279169 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 25 01:28:24.279970 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 25 01:28:24.280212 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 25 01:28:24.281022 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 25 01:28:24.281330 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 25 01:28:24.283075 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 25 01:28:24.283368 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 25 01:28:24.284553 systemd[1]: ignition-files.service: Deactivated successfully. Mar 25 01:28:24.284832 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 25 01:28:24.285362 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 25 01:28:24.285559 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 25 01:28:24.289018 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 25 01:28:24.292145 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 25 01:28:24.292525 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 25 01:28:24.294942 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 25 01:28:24.295932 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 25 01:28:24.297860 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 25 01:28:24.298528 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 25 01:28:24.298816 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 25 01:28:24.309924 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 25 01:28:24.310084 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 25 01:28:24.339963 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 25 01:28:24.346928 ignition[986]: INFO : Ignition 2.20.0 Mar 25 01:28:24.346928 ignition[986]: INFO : Stage: umount Mar 25 01:28:24.346928 ignition[986]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 25 01:28:24.346928 ignition[986]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Mar 25 01:28:24.346928 ignition[986]: INFO : umount: umount passed Mar 25 01:28:24.346928 ignition[986]: INFO : Ignition finished successfully Mar 25 01:28:24.346676 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 25 01:28:24.346845 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 25 01:28:24.347751 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 25 01:28:24.347913 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 25 01:28:24.349819 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 25 01:28:24.349971 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 25 01:28:24.351169 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 25 01:28:24.351257 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 25 01:28:24.352028 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 25 01:28:24.352106 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 25 01:28:24.352766 systemd[1]: Stopped target network.target - Network. Mar 25 01:28:24.353415 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 25 01:28:24.353498 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 25 01:28:24.354107 systemd[1]: Stopped target paths.target - Path Units. Mar 25 01:28:24.354715 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 25 01:28:24.358774 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 25 01:28:24.359331 systemd[1]: Stopped target slices.target - Slice Units. Mar 25 01:28:24.360262 systemd[1]: Stopped target sockets.target - Socket Units. Mar 25 01:28:24.361102 systemd[1]: iscsid.socket: Deactivated successfully. Mar 25 01:28:24.361194 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 25 01:28:24.361910 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 25 01:28:24.361966 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 25 01:28:24.362737 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 25 01:28:24.362843 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 25 01:28:24.363574 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 25 01:28:24.363732 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 25 01:28:24.364267 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 25 01:28:24.364322 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 25 01:28:24.365299 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 25 01:28:24.365979 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 25 01:28:24.375612 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 25 01:28:24.375913 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 25 01:28:24.382254 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Mar 25 01:28:24.382775 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 25 01:28:24.383030 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 25 01:28:24.385337 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Mar 25 01:28:24.386652 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 25 01:28:24.386789 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 25 01:28:24.390539 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 25 01:28:24.391070 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 25 01:28:24.391176 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 25 01:28:24.392360 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 25 01:28:24.392438 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 25 01:28:24.393028 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 25 01:28:24.393100 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 25 01:28:24.393741 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 25 01:28:24.393808 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 25 01:28:24.395004 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 25 01:28:24.401155 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 25 01:28:24.401296 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Mar 25 01:28:24.424145 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 25 01:28:24.424380 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 25 01:28:24.425429 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 25 01:28:24.425505 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 25 01:28:24.426164 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 25 01:28:24.426224 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 25 01:28:24.428735 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 25 01:28:24.428867 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 25 01:28:24.430482 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 25 01:28:24.430601 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 25 01:28:24.432126 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 25 01:28:24.432248 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 25 01:28:24.436903 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 25 01:28:24.437401 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 25 01:28:24.437525 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 25 01:28:24.439876 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 25 01:28:24.439984 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 25 01:28:24.447717 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 25 01:28:24.447871 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 25 01:28:24.468707 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 25 01:28:24.470085 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 25 01:28:24.471621 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 25 01:28:24.471896 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 25 01:28:24.473503 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 25 01:28:24.476820 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 25 01:28:24.502841 systemd[1]: Switching root. Mar 25 01:28:24.544421 systemd-journald[181]: Journal stopped Mar 25 01:28:26.041721 systemd-journald[181]: Received SIGTERM from PID 1 (systemd). Mar 25 01:28:26.041845 kernel: SELinux: policy capability network_peer_controls=1 Mar 25 01:28:26.041880 kernel: SELinux: policy capability open_perms=1 Mar 25 01:28:26.041906 kernel: SELinux: policy capability extended_socket_class=1 Mar 25 01:28:26.041935 kernel: SELinux: policy capability always_check_network=0 Mar 25 01:28:26.041956 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 25 01:28:26.041991 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 25 01:28:26.042010 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 25 01:28:26.042029 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 25 01:28:26.042047 kernel: audit: type=1403 audit(1742866104.736:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 25 01:28:26.042070 systemd[1]: Successfully loaded SELinux policy in 53.286ms. Mar 25 01:28:26.042101 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 17.571ms. Mar 25 01:28:26.042126 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 25 01:28:26.042149 systemd[1]: Detected virtualization kvm. Mar 25 01:28:26.042172 systemd[1]: Detected architecture x86-64. Mar 25 01:28:26.042198 systemd[1]: Detected first boot. Mar 25 01:28:26.042220 systemd[1]: Hostname set to . Mar 25 01:28:26.042241 systemd[1]: Initializing machine ID from VM UUID. Mar 25 01:28:26.042265 zram_generator::config[1035]: No configuration found. Mar 25 01:28:26.042288 kernel: Guest personality initialized and is inactive Mar 25 01:28:26.042308 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Mar 25 01:28:26.042330 kernel: Initialized host personality Mar 25 01:28:26.042356 kernel: NET: Registered PF_VSOCK protocol family Mar 25 01:28:26.042382 systemd[1]: Populated /etc with preset unit settings. Mar 25 01:28:26.042406 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Mar 25 01:28:26.042428 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 25 01:28:26.042449 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 25 01:28:26.042469 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 25 01:28:26.042491 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 25 01:28:26.042512 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 25 01:28:26.042535 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 25 01:28:26.042557 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 25 01:28:26.042585 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 25 01:28:26.042613 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 25 01:28:26.043705 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 25 01:28:26.043751 systemd[1]: Created slice user.slice - User and Session Slice. Mar 25 01:28:26.043774 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 25 01:28:26.043797 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 25 01:28:26.043822 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 25 01:28:26.043846 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 25 01:28:26.043879 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 25 01:28:26.043902 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 25 01:28:26.043925 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 25 01:28:26.043945 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 25 01:28:26.043964 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 25 01:28:26.043989 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 25 01:28:26.044016 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 25 01:28:26.044034 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 25 01:28:26.044062 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 25 01:28:26.044081 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 25 01:28:26.044104 systemd[1]: Reached target slices.target - Slice Units. Mar 25 01:28:26.044126 systemd[1]: Reached target swap.target - Swaps. Mar 25 01:28:26.044149 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 25 01:28:26.044172 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 25 01:28:26.044197 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 25 01:28:26.044227 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 25 01:28:26.044249 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 25 01:28:26.044272 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 25 01:28:26.044296 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 25 01:28:26.044318 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 25 01:28:26.044340 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 25 01:28:26.044362 systemd[1]: Mounting media.mount - External Media Directory... Mar 25 01:28:26.044386 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:26.044410 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 25 01:28:26.044439 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 25 01:28:26.044462 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 25 01:28:26.044486 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 25 01:28:26.044508 systemd[1]: Reached target machines.target - Containers. Mar 25 01:28:26.044533 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 25 01:28:26.044555 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 25 01:28:26.044578 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 25 01:28:26.044602 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 25 01:28:26.044624 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 25 01:28:26.044679 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 25 01:28:26.044702 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 25 01:28:26.044725 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 25 01:28:26.044747 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 25 01:28:26.044773 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 25 01:28:26.044797 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 25 01:28:26.044820 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 25 01:28:26.044844 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 25 01:28:26.044869 systemd[1]: Stopped systemd-fsck-usr.service. Mar 25 01:28:26.044897 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 25 01:28:26.044918 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 25 01:28:26.044942 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 25 01:28:26.044967 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 25 01:28:26.044992 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 25 01:28:26.045015 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 25 01:28:26.045038 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 25 01:28:26.045060 systemd[1]: verity-setup.service: Deactivated successfully. Mar 25 01:28:26.045099 systemd[1]: Stopped verity-setup.service. Mar 25 01:28:26.045131 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:26.045228 systemd-journald[1106]: Collecting audit messages is disabled. Mar 25 01:28:26.045278 systemd-journald[1106]: Journal started Mar 25 01:28:26.045324 systemd-journald[1106]: Runtime Journal (/run/log/journal/98ef30a558654dd4b4bca9ef930af956) is 4.9M, max 39.3M, 34.3M free. Mar 25 01:28:26.052189 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 25 01:28:25.723833 systemd[1]: Queued start job for default target multi-user.target. Mar 25 01:28:25.732795 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 25 01:28:25.733525 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 25 01:28:26.061765 systemd[1]: Started systemd-journald.service - Journal Service. Mar 25 01:28:26.058381 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 25 01:28:26.063977 systemd[1]: Mounted media.mount - External Media Directory. Mar 25 01:28:26.066268 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 25 01:28:26.067177 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 25 01:28:26.068535 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 25 01:28:26.070987 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 25 01:28:26.073312 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 25 01:28:26.073605 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 25 01:28:26.091688 kernel: fuse: init (API version 7.39) Mar 25 01:28:26.076515 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 25 01:28:26.076864 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 25 01:28:26.078542 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 25 01:28:26.079793 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 25 01:28:26.093499 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 25 01:28:26.097220 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 25 01:28:26.100870 kernel: loop: module loaded Mar 25 01:28:26.099313 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 25 01:28:26.103212 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 25 01:28:26.105247 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 25 01:28:26.117812 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 25 01:28:26.139811 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 25 01:28:26.162912 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 25 01:28:26.179902 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 25 01:28:26.181687 kernel: ACPI: bus type drm_connector registered Mar 25 01:28:26.190882 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 25 01:28:26.192250 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 25 01:28:26.192764 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 25 01:28:26.199550 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 25 01:28:26.216996 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 25 01:28:26.221450 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 25 01:28:26.224091 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 25 01:28:26.227687 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 25 01:28:26.238049 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 25 01:28:26.238820 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 25 01:28:26.246188 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 25 01:28:26.247055 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 25 01:28:26.252457 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 25 01:28:26.260183 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 25 01:28:26.268865 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 25 01:28:26.269268 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 25 01:28:26.271837 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 25 01:28:26.273357 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 25 01:28:26.275361 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 25 01:28:26.277770 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 25 01:28:26.306618 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 25 01:28:26.321683 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 25 01:28:26.355882 kernel: loop0: detected capacity change from 0 to 109808 Mar 25 01:28:26.384071 systemd-journald[1106]: Time spent on flushing to /var/log/journal/98ef30a558654dd4b4bca9ef930af956 is 106.936ms for 1004 entries. Mar 25 01:28:26.384071 systemd-journald[1106]: System Journal (/var/log/journal/98ef30a558654dd4b4bca9ef930af956) is 8M, max 195.6M, 187.6M free. Mar 25 01:28:26.530424 systemd-journald[1106]: Received client request to flush runtime journal. Mar 25 01:28:26.530542 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 25 01:28:26.530584 kernel: loop1: detected capacity change from 0 to 151640 Mar 25 01:28:26.382772 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 25 01:28:26.384815 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 25 01:28:26.396012 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 25 01:28:26.430791 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 25 01:28:26.509097 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 25 01:28:26.537141 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 25 01:28:26.554678 kernel: loop2: detected capacity change from 0 to 218376 Mar 25 01:28:26.561319 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 25 01:28:26.568166 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 25 01:28:26.609294 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 25 01:28:26.612535 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 25 01:28:26.632093 kernel: loop3: detected capacity change from 0 to 8 Mar 25 01:28:26.666139 systemd-tmpfiles[1176]: ACLs are not supported, ignoring. Mar 25 01:28:26.666167 systemd-tmpfiles[1176]: ACLs are not supported, ignoring. Mar 25 01:28:26.684557 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 25 01:28:26.690151 udevadm[1180]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 25 01:28:26.694537 kernel: loop4: detected capacity change from 0 to 109808 Mar 25 01:28:26.738574 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 25 01:28:26.745756 kernel: loop5: detected capacity change from 0 to 151640 Mar 25 01:28:26.773685 kernel: loop6: detected capacity change from 0 to 218376 Mar 25 01:28:26.797674 kernel: loop7: detected capacity change from 0 to 8 Mar 25 01:28:26.801489 (sd-merge)[1183]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. Mar 25 01:28:26.805049 (sd-merge)[1183]: Merged extensions into '/usr'. Mar 25 01:28:26.823982 systemd[1]: Reload requested from client PID 1155 ('systemd-sysext') (unit systemd-sysext.service)... Mar 25 01:28:26.824225 systemd[1]: Reloading... Mar 25 01:28:27.107096 zram_generator::config[1224]: No configuration found. Mar 25 01:28:27.258518 ldconfig[1147]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 25 01:28:27.360402 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 25 01:28:27.468963 systemd[1]: Reloading finished in 643 ms. Mar 25 01:28:27.490200 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 25 01:28:27.493660 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 25 01:28:27.509446 systemd[1]: Starting ensure-sysext.service... Mar 25 01:28:27.516030 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 25 01:28:27.567896 systemd[1]: Reload requested from client PID 1254 ('systemctl') (unit ensure-sysext.service)... Mar 25 01:28:27.567922 systemd[1]: Reloading... Mar 25 01:28:27.610972 systemd-tmpfiles[1255]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 25 01:28:27.611622 systemd-tmpfiles[1255]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 25 01:28:27.613732 systemd-tmpfiles[1255]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 25 01:28:27.615490 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. Mar 25 01:28:27.615617 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. Mar 25 01:28:27.628786 systemd-tmpfiles[1255]: Detected autofs mount point /boot during canonicalization of boot. Mar 25 01:28:27.628809 systemd-tmpfiles[1255]: Skipping /boot Mar 25 01:28:27.673207 systemd-tmpfiles[1255]: Detected autofs mount point /boot during canonicalization of boot. Mar 25 01:28:27.673228 systemd-tmpfiles[1255]: Skipping /boot Mar 25 01:28:27.747940 zram_generator::config[1287]: No configuration found. Mar 25 01:28:27.935680 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 25 01:28:28.046478 systemd[1]: Reloading finished in 477 ms. Mar 25 01:28:28.062870 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 25 01:28:28.082346 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 25 01:28:28.097897 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 25 01:28:28.104190 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 25 01:28:28.109131 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 25 01:28:28.120339 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 25 01:28:28.125269 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 25 01:28:28.137336 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 25 01:28:28.145080 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.145464 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 25 01:28:28.155309 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 25 01:28:28.173799 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 25 01:28:28.187089 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 25 01:28:28.188164 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 25 01:28:28.188429 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 25 01:28:28.188624 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.210665 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 25 01:28:28.223228 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.225953 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 25 01:28:28.226468 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 25 01:28:28.226828 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 25 01:28:28.227094 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.229363 systemd-udevd[1333]: Using default interface naming scheme 'v255'. Mar 25 01:28:28.232297 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 25 01:28:28.248508 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.248923 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 25 01:28:28.254266 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 25 01:28:28.255174 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 25 01:28:28.255514 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 25 01:28:28.256498 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.266194 systemd[1]: Finished ensure-sysext.service. Mar 25 01:28:28.272735 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 25 01:28:28.299845 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 25 01:28:28.302400 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 25 01:28:28.316044 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 25 01:28:28.316360 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 25 01:28:28.329600 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 25 01:28:28.329969 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 25 01:28:28.330948 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 25 01:28:28.336476 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 25 01:28:28.338737 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 25 01:28:28.345530 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 25 01:28:28.353391 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 25 01:28:28.354289 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 25 01:28:28.354591 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 25 01:28:28.362187 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 25 01:28:28.362784 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 25 01:28:28.365296 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 25 01:28:28.381750 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 25 01:28:28.396433 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 25 01:28:28.412324 augenrules[1383]: No rules Mar 25 01:28:28.419090 systemd[1]: audit-rules.service: Deactivated successfully. Mar 25 01:28:28.419411 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 25 01:28:28.595001 systemd-resolved[1332]: Positive Trust Anchors: Mar 25 01:28:28.595020 systemd-resolved[1332]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 25 01:28:28.595059 systemd-resolved[1332]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 25 01:28:28.601666 systemd-resolved[1332]: Using system hostname 'ci-4284.0.0-c-20cdc03f82'. Mar 25 01:28:28.604293 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 25 01:28:28.605038 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 25 01:28:28.620999 systemd-networkd[1366]: lo: Link UP Mar 25 01:28:28.621524 systemd-networkd[1366]: lo: Gained carrier Mar 25 01:28:28.625370 systemd-networkd[1366]: Enumeration completed Mar 25 01:28:28.628098 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 25 01:28:28.629413 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 25 01:28:28.629924 systemd[1]: Reached target network.target - Network. Mar 25 01:28:28.630952 systemd[1]: Reached target time-set.target - System Time Set. Mar 25 01:28:28.634116 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 25 01:28:28.637327 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 25 01:28:28.683371 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 25 01:28:28.689853 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1365) Mar 25 01:28:28.707548 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Mar 25 01:28:28.729534 systemd[1]: Condition check resulted in dev-disk-by\x2dlabel-config\x2d2.device - /dev/disk/by-label/config-2 being skipped. Mar 25 01:28:28.732189 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Mar 25 01:28:28.733217 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.733828 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 25 01:28:28.737506 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 25 01:28:28.741936 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 25 01:28:28.746959 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 25 01:28:28.748205 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 25 01:28:28.748279 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 25 01:28:28.748330 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 25 01:28:28.748363 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 25 01:28:28.776104 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 25 01:28:28.776366 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 25 01:28:28.800672 kernel: ISO 9660 Extensions: RRIP_1991A Mar 25 01:28:28.804197 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Mar 25 01:28:28.813336 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 25 01:28:28.814825 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 25 01:28:28.816096 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 25 01:28:28.819680 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 25 01:28:28.820730 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 25 01:28:28.832582 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 25 01:28:28.888588 systemd-networkd[1366]: eth0: Configuring with /run/systemd/network/10-a6:b0:6e:12:c4:80.network. Mar 25 01:28:28.890933 systemd-networkd[1366]: eth0: Link UP Mar 25 01:28:28.890945 systemd-networkd[1366]: eth0: Gained carrier Mar 25 01:28:28.896043 systemd-networkd[1366]: eth1: Configuring with /run/systemd/network/10-fa:46:ea:1e:7d:d7.network. Mar 25 01:28:28.898863 systemd-networkd[1366]: eth1: Link UP Mar 25 01:28:28.899018 systemd-networkd[1366]: eth1: Gained carrier Mar 25 01:28:28.903841 systemd-timesyncd[1352]: Network configuration changed, trying to establish connection. Mar 25 01:28:28.917687 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Mar 25 01:28:28.931665 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Mar 25 01:28:28.935803 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 25 01:28:28.938660 kernel: ACPI: button: Power Button [PWRF] Mar 25 01:28:28.939310 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 25 01:28:28.961711 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Mar 25 01:28:28.992487 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 25 01:28:28.997744 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Mar 25 01:28:29.003701 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Mar 25 01:28:29.015746 kernel: Console: switching to colour dummy device 80x25 Mar 25 01:28:29.016939 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Mar 25 01:28:29.017039 kernel: [drm] features: -context_init Mar 25 01:28:29.033829 kernel: [drm] number of scanouts: 1 Mar 25 01:28:29.033955 kernel: [drm] number of cap sets: 0 Mar 25 01:28:29.047679 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 Mar 25 01:28:29.068746 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 25 01:28:29.090944 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Mar 25 01:28:29.091110 kernel: Console: switching to colour frame buffer device 128x48 Mar 25 01:28:29.098690 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Mar 25 01:28:29.101713 kernel: mousedev: PS/2 mouse device common for all mice Mar 25 01:28:29.136279 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 25 01:28:29.136626 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 25 01:28:29.144218 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 25 01:28:29.149776 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 25 01:28:29.169949 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 25 01:28:29.170267 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 25 01:28:29.173928 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 25 01:28:29.282856 kernel: EDAC MC: Ver: 3.0.0 Mar 25 01:28:29.311157 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 25 01:28:29.321539 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 25 01:28:29.327307 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 25 01:28:29.360129 lvm[1446]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 25 01:28:29.398988 systemd-timesyncd[1352]: Contacted time server 108.61.73.243:123 (0.flatcar.pool.ntp.org). Mar 25 01:28:29.399132 systemd-timesyncd[1352]: Initial clock synchronization to Tue 2025-03-25 01:28:29.335176 UTC. Mar 25 01:28:29.401762 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 25 01:28:29.403781 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 25 01:28:29.404050 systemd[1]: Reached target sysinit.target - System Initialization. Mar 25 01:28:29.404510 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 25 01:28:29.405228 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 25 01:28:29.405957 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 25 01:28:29.406554 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 25 01:28:29.407021 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 25 01:28:29.407271 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 25 01:28:29.407458 systemd[1]: Reached target paths.target - Path Units. Mar 25 01:28:29.407722 systemd[1]: Reached target timers.target - Timer Units. Mar 25 01:28:29.412822 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 25 01:28:29.416410 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 25 01:28:29.423532 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 25 01:28:29.424921 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 25 01:28:29.426896 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 25 01:28:29.441479 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 25 01:28:29.447363 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 25 01:28:29.452467 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 25 01:28:29.461546 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 25 01:28:29.464482 systemd[1]: Reached target sockets.target - Socket Units. Mar 25 01:28:29.466773 systemd[1]: Reached target basic.target - Basic System. Mar 25 01:28:29.469860 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 25 01:28:29.469907 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 25 01:28:29.473886 systemd[1]: Starting containerd.service - containerd container runtime... Mar 25 01:28:29.480867 lvm[1450]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 25 01:28:29.481610 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 25 01:28:29.490035 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 25 01:28:29.498791 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 25 01:28:29.514953 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 25 01:28:29.515907 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 25 01:28:29.529265 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 25 01:28:29.539043 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 25 01:28:29.547012 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 25 01:28:29.551624 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 25 01:28:29.565714 jq[1454]: false Mar 25 01:28:29.569904 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 25 01:28:29.574247 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 25 01:28:29.579198 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 25 01:28:29.583316 systemd[1]: Starting update-engine.service - Update Engine... Mar 25 01:28:29.589138 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 25 01:28:29.594525 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 25 01:28:29.598743 extend-filesystems[1455]: Found loop4 Mar 25 01:28:29.608460 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 25 01:28:29.616881 extend-filesystems[1455]: Found loop5 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found loop6 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found loop7 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda1 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda2 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda3 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found usr Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda4 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda6 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda7 Mar 25 01:28:29.616881 extend-filesystems[1455]: Found vda9 Mar 25 01:28:29.616881 extend-filesystems[1455]: Checking size of /dev/vda9 Mar 25 01:28:29.610765 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 25 01:28:29.661156 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 25 01:28:29.712731 extend-filesystems[1455]: Resized partition /dev/vda9 Mar 25 01:28:29.661528 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 25 01:28:29.705226 (ntainerd)[1487]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 25 01:28:29.727476 update_engine[1464]: I20250325 01:28:29.704066 1464 main.cc:92] Flatcar Update Engine starting Mar 25 01:28:29.738309 extend-filesystems[1488]: resize2fs 1.47.2 (1-Jan-2025) Mar 25 01:28:29.734175 dbus-daemon[1453]: [system] SELinux support is enabled Mar 25 01:28:29.767809 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Mar 25 01:28:29.767872 jq[1465]: true Mar 25 01:28:29.750209 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 25 01:28:29.759192 systemd[1]: motdgen.service: Deactivated successfully. Mar 25 01:28:29.761747 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 25 01:28:29.777123 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 25 01:28:29.777213 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 25 01:28:29.781667 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 25 01:28:29.781856 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Mar 25 01:28:29.781899 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 25 01:28:29.797290 coreos-metadata[1452]: Mar 25 01:28:29.796 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Mar 25 01:28:29.801096 tar[1470]: linux-amd64/LICENSE Mar 25 01:28:29.801096 tar[1470]: linux-amd64/helm Mar 25 01:28:29.801700 update_engine[1464]: I20250325 01:28:29.800806 1464 update_check_scheduler.cc:74] Next update check in 11m40s Mar 25 01:28:29.800712 systemd[1]: Started update-engine.service - Update Engine. Mar 25 01:28:29.808969 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 25 01:28:29.822490 systemd-logind[1463]: New seat seat0. Mar 25 01:28:29.844052 systemd-logind[1463]: Watching system buttons on /dev/input/event1 (Power Button) Mar 25 01:28:29.844322 systemd-logind[1463]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 25 01:28:29.845073 systemd[1]: Started systemd-logind.service - User Login Management. Mar 25 01:28:29.850862 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1374) Mar 25 01:28:29.853702 coreos-metadata[1452]: Mar 25 01:28:29.851 INFO Fetch successful Mar 25 01:28:29.946971 jq[1491]: true Mar 25 01:28:30.083466 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Mar 25 01:28:30.088982 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 25 01:28:30.113678 extend-filesystems[1488]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 25 01:28:30.113678 extend-filesystems[1488]: old_desc_blocks = 1, new_desc_blocks = 8 Mar 25 01:28:30.113678 extend-filesystems[1488]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Mar 25 01:28:30.128413 extend-filesystems[1455]: Resized filesystem in /dev/vda9 Mar 25 01:28:30.128413 extend-filesystems[1455]: Found vdb Mar 25 01:28:30.118523 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 25 01:28:30.121805 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 25 01:28:30.140442 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 25 01:28:30.231213 bash[1525]: Updated "/home/core/.ssh/authorized_keys" Mar 25 01:28:30.233349 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 25 01:28:30.251964 systemd[1]: Starting sshkeys.service... Mar 25 01:28:30.326328 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Mar 25 01:28:30.330582 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Mar 25 01:28:30.340859 systemd-networkd[1366]: eth0: Gained IPv6LL Mar 25 01:28:30.357232 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 25 01:28:30.363417 systemd[1]: Reached target network-online.target - Network is Online. Mar 25 01:28:30.372276 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:28:30.383120 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 25 01:28:30.434014 coreos-metadata[1530]: Mar 25 01:28:30.433 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Mar 25 01:28:30.453523 coreos-metadata[1530]: Mar 25 01:28:30.445 INFO Fetch successful Mar 25 01:28:30.466658 unknown[1530]: wrote ssh authorized keys file for user: core Mar 25 01:28:30.467320 locksmithd[1495]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 25 01:28:30.523087 update-ssh-keys[1541]: Updated "/home/core/.ssh/authorized_keys" Mar 25 01:28:30.525419 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Mar 25 01:28:30.530996 systemd-networkd[1366]: eth1: Gained IPv6LL Mar 25 01:28:30.532741 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 25 01:28:30.535293 systemd[1]: Finished sshkeys.service. Mar 25 01:28:30.596906 containerd[1487]: time="2025-03-25T01:28:30Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Mar 25 01:28:30.605671 containerd[1487]: time="2025-03-25T01:28:30.603882201Z" level=info msg="starting containerd" revision=88aa2f531d6c2922003cc7929e51daf1c14caa0a version=v2.0.1 Mar 25 01:28:30.636361 sshd_keygen[1486]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 25 01:28:30.644249 containerd[1487]: time="2025-03-25T01:28:30.644160835Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.592µs" Mar 25 01:28:30.644249 containerd[1487]: time="2025-03-25T01:28:30.644220871Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Mar 25 01:28:30.644249 containerd[1487]: time="2025-03-25T01:28:30.644254590Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Mar 25 01:28:30.644545 containerd[1487]: time="2025-03-25T01:28:30.644508712Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Mar 25 01:28:30.644545 containerd[1487]: time="2025-03-25T01:28:30.644539294Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Mar 25 01:28:30.644694 containerd[1487]: time="2025-03-25T01:28:30.644581927Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 25 01:28:30.646745 containerd[1487]: time="2025-03-25T01:28:30.645870086Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 25 01:28:30.646745 containerd[1487]: time="2025-03-25T01:28:30.645935543Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 25 01:28:30.651764 containerd[1487]: time="2025-03-25T01:28:30.651640167Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 25 01:28:30.651764 containerd[1487]: time="2025-03-25T01:28:30.651726744Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 25 01:28:30.651764 containerd[1487]: time="2025-03-25T01:28:30.651754457Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 25 01:28:30.651764 containerd[1487]: time="2025-03-25T01:28:30.651771735Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Mar 25 01:28:30.652125 containerd[1487]: time="2025-03-25T01:28:30.652015408Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Mar 25 01:28:30.653425 containerd[1487]: time="2025-03-25T01:28:30.652375712Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 25 01:28:30.653425 containerd[1487]: time="2025-03-25T01:28:30.652438321Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 25 01:28:30.653425 containerd[1487]: time="2025-03-25T01:28:30.652479692Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Mar 25 01:28:30.654427 containerd[1487]: time="2025-03-25T01:28:30.652876971Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Mar 25 01:28:30.655298 containerd[1487]: time="2025-03-25T01:28:30.655246389Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Mar 25 01:28:30.655468 containerd[1487]: time="2025-03-25T01:28:30.655436463Z" level=info msg="metadata content store policy set" policy=shared Mar 25 01:28:30.662491 containerd[1487]: time="2025-03-25T01:28:30.662389476Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Mar 25 01:28:30.662491 containerd[1487]: time="2025-03-25T01:28:30.662506772Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662532376Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662555735Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662577655Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662597280Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662698319Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662727713Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662745965Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662776888Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Mar 25 01:28:30.662789 containerd[1487]: time="2025-03-25T01:28:30.662797139Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Mar 25 01:28:30.663156 containerd[1487]: time="2025-03-25T01:28:30.662828543Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Mar 25 01:28:30.663156 containerd[1487]: time="2025-03-25T01:28:30.663056146Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Mar 25 01:28:30.663156 containerd[1487]: time="2025-03-25T01:28:30.663091586Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Mar 25 01:28:30.663156 containerd[1487]: time="2025-03-25T01:28:30.663114594Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Mar 25 01:28:30.663156 containerd[1487]: time="2025-03-25T01:28:30.663136897Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Mar 25 01:28:30.663156 containerd[1487]: time="2025-03-25T01:28:30.663155448Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Mar 25 01:28:30.663387 containerd[1487]: time="2025-03-25T01:28:30.663174522Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Mar 25 01:28:30.663387 containerd[1487]: time="2025-03-25T01:28:30.663194518Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Mar 25 01:28:30.663387 containerd[1487]: time="2025-03-25T01:28:30.663212628Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Mar 25 01:28:30.663387 containerd[1487]: time="2025-03-25T01:28:30.663232249Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Mar 25 01:28:30.663387 containerd[1487]: time="2025-03-25T01:28:30.663251040Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Mar 25 01:28:30.663387 containerd[1487]: time="2025-03-25T01:28:30.663268213Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Mar 25 01:28:30.663387 containerd[1487]: time="2025-03-25T01:28:30.663373542Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Mar 25 01:28:30.663699 containerd[1487]: time="2025-03-25T01:28:30.663395265Z" level=info msg="Start snapshots syncer" Mar 25 01:28:30.663699 containerd[1487]: time="2025-03-25T01:28:30.663453375Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Mar 25 01:28:30.668420 containerd[1487]: time="2025-03-25T01:28:30.667024618Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Mar 25 01:28:30.668420 containerd[1487]: time="2025-03-25T01:28:30.667134960Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Mar 25 01:28:30.668890 containerd[1487]: time="2025-03-25T01:28:30.667283024Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Mar 25 01:28:30.668890 containerd[1487]: time="2025-03-25T01:28:30.667508490Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Mar 25 01:28:30.668890 containerd[1487]: time="2025-03-25T01:28:30.667548300Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Mar 25 01:28:30.668890 containerd[1487]: time="2025-03-25T01:28:30.667566682Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Mar 25 01:28:30.668890 containerd[1487]: time="2025-03-25T01:28:30.667583316Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Mar 25 01:28:30.668890 containerd[1487]: time="2025-03-25T01:28:30.667607998Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670093001Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670158008Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670208304Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670232081Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670247026Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670324350Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670360373Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670379155Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670396512Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670409496Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670425835Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670443324Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670469829Z" level=info msg="runtime interface created" Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670480841Z" level=info msg="created NRI interface" Mar 25 01:28:30.671673 containerd[1487]: time="2025-03-25T01:28:30.670495923Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Mar 25 01:28:30.672380 containerd[1487]: time="2025-03-25T01:28:30.670524100Z" level=info msg="Connect containerd service" Mar 25 01:28:30.672380 containerd[1487]: time="2025-03-25T01:28:30.670575239Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 25 01:28:30.680962 containerd[1487]: time="2025-03-25T01:28:30.680881313Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 25 01:28:30.700489 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 25 01:28:30.709277 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 25 01:28:30.759128 systemd[1]: issuegen.service: Deactivated successfully. Mar 25 01:28:30.760029 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 25 01:28:30.769806 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 25 01:28:30.823869 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 25 01:28:30.830145 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 25 01:28:30.835156 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 25 01:28:30.838718 systemd[1]: Reached target getty.target - Login Prompts. Mar 25 01:28:30.966618 containerd[1487]: time="2025-03-25T01:28:30.966467007Z" level=info msg="Start subscribing containerd event" Mar 25 01:28:30.966618 containerd[1487]: time="2025-03-25T01:28:30.966563497Z" level=info msg="Start recovering state" Mar 25 01:28:30.968234 containerd[1487]: time="2025-03-25T01:28:30.967987532Z" level=info msg="Start event monitor" Mar 25 01:28:30.968234 containerd[1487]: time="2025-03-25T01:28:30.968127700Z" level=info msg="Start cni network conf syncer for default" Mar 25 01:28:30.968234 containerd[1487]: time="2025-03-25T01:28:30.968143830Z" level=info msg="Start streaming server" Mar 25 01:28:30.968234 containerd[1487]: time="2025-03-25T01:28:30.968166196Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Mar 25 01:28:30.968234 containerd[1487]: time="2025-03-25T01:28:30.968175921Z" level=info msg="runtime interface starting up..." Mar 25 01:28:30.968533 containerd[1487]: time="2025-03-25T01:28:30.968183368Z" level=info msg="starting plugins..." Mar 25 01:28:30.968533 containerd[1487]: time="2025-03-25T01:28:30.968295140Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Mar 25 01:28:30.971404 containerd[1487]: time="2025-03-25T01:28:30.970952169Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 25 01:28:30.971404 containerd[1487]: time="2025-03-25T01:28:30.971040599Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 25 01:28:30.971404 containerd[1487]: time="2025-03-25T01:28:30.971224742Z" level=info msg="containerd successfully booted in 0.376778s" Mar 25 01:28:30.971448 systemd[1]: Started containerd.service - containerd container runtime. Mar 25 01:28:31.229258 tar[1470]: linux-amd64/README.md Mar 25 01:28:31.250477 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 25 01:28:31.920696 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:28:31.922407 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 25 01:28:31.929317 systemd[1]: Startup finished in 1.067s (kernel) + 6.021s (initrd) + 7.243s (userspace) = 14.332s. Mar 25 01:28:31.933575 (kubelet)[1590]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 25 01:28:32.407924 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 25 01:28:32.411565 systemd[1]: Started sshd@0-143.198.155.16:22-139.178.89.65:45556.service - OpenSSH per-connection server daemon (139.178.89.65:45556). Mar 25 01:28:32.512275 sshd[1600]: Accepted publickey for core from 139.178.89.65 port 45556 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:28:32.514284 sshd-session[1600]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:28:32.531052 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 25 01:28:32.534806 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 25 01:28:32.542016 systemd-logind[1463]: New session 1 of user core. Mar 25 01:28:32.571356 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 25 01:28:32.578670 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 25 01:28:32.596211 (systemd)[1604]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 25 01:28:32.604504 systemd-logind[1463]: New session c1 of user core. Mar 25 01:28:32.694825 kubelet[1590]: E0325 01:28:32.694680 1590 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 25 01:28:32.697916 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 25 01:28:32.698147 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 25 01:28:32.699018 systemd[1]: kubelet.service: Consumed 1.364s CPU time, 253.6M memory peak. Mar 25 01:28:32.793572 systemd[1604]: Queued start job for default target default.target. Mar 25 01:28:32.801715 systemd[1604]: Created slice app.slice - User Application Slice. Mar 25 01:28:32.801772 systemd[1604]: Reached target paths.target - Paths. Mar 25 01:28:32.801840 systemd[1604]: Reached target timers.target - Timers. Mar 25 01:28:32.804473 systemd[1604]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 25 01:28:32.839092 systemd[1604]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 25 01:28:32.839449 systemd[1604]: Reached target sockets.target - Sockets. Mar 25 01:28:32.839549 systemd[1604]: Reached target basic.target - Basic System. Mar 25 01:28:32.839613 systemd[1604]: Reached target default.target - Main User Target. Mar 25 01:28:32.839922 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 25 01:28:32.840267 systemd[1604]: Startup finished in 221ms. Mar 25 01:28:32.850347 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 25 01:28:32.926967 systemd[1]: Started sshd@1-143.198.155.16:22-139.178.89.65:45570.service - OpenSSH per-connection server daemon (139.178.89.65:45570). Mar 25 01:28:33.010515 sshd[1617]: Accepted publickey for core from 139.178.89.65 port 45570 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:28:33.012694 sshd-session[1617]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:28:33.020189 systemd-logind[1463]: New session 2 of user core. Mar 25 01:28:33.029923 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 25 01:28:33.093308 sshd[1619]: Connection closed by 139.178.89.65 port 45570 Mar 25 01:28:33.094118 sshd-session[1617]: pam_unix(sshd:session): session closed for user core Mar 25 01:28:33.107277 systemd[1]: sshd@1-143.198.155.16:22-139.178.89.65:45570.service: Deactivated successfully. Mar 25 01:28:33.109559 systemd[1]: session-2.scope: Deactivated successfully. Mar 25 01:28:33.112142 systemd-logind[1463]: Session 2 logged out. Waiting for processes to exit. Mar 25 01:28:33.114069 systemd[1]: Started sshd@2-143.198.155.16:22-139.178.89.65:45576.service - OpenSSH per-connection server daemon (139.178.89.65:45576). Mar 25 01:28:33.116712 systemd-logind[1463]: Removed session 2. Mar 25 01:28:33.174497 sshd[1624]: Accepted publickey for core from 139.178.89.65 port 45576 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:28:33.176259 sshd-session[1624]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:28:33.184002 systemd-logind[1463]: New session 3 of user core. Mar 25 01:28:33.196021 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 25 01:28:33.256288 sshd[1627]: Connection closed by 139.178.89.65 port 45576 Mar 25 01:28:33.256081 sshd-session[1624]: pam_unix(sshd:session): session closed for user core Mar 25 01:28:33.273527 systemd[1]: sshd@2-143.198.155.16:22-139.178.89.65:45576.service: Deactivated successfully. Mar 25 01:28:33.276719 systemd[1]: session-3.scope: Deactivated successfully. Mar 25 01:28:33.277927 systemd-logind[1463]: Session 3 logged out. Waiting for processes to exit. Mar 25 01:28:33.282905 systemd[1]: Started sshd@3-143.198.155.16:22-139.178.89.65:45590.service - OpenSSH per-connection server daemon (139.178.89.65:45590). Mar 25 01:28:33.283851 systemd-logind[1463]: Removed session 3. Mar 25 01:28:33.346678 sshd[1632]: Accepted publickey for core from 139.178.89.65 port 45590 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:28:33.349998 sshd-session[1632]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:28:33.357296 systemd-logind[1463]: New session 4 of user core. Mar 25 01:28:33.368113 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 25 01:28:33.432217 sshd[1635]: Connection closed by 139.178.89.65 port 45590 Mar 25 01:28:33.433247 sshd-session[1632]: pam_unix(sshd:session): session closed for user core Mar 25 01:28:33.451323 systemd[1]: sshd@3-143.198.155.16:22-139.178.89.65:45590.service: Deactivated successfully. Mar 25 01:28:33.454019 systemd[1]: session-4.scope: Deactivated successfully. Mar 25 01:28:33.455995 systemd-logind[1463]: Session 4 logged out. Waiting for processes to exit. Mar 25 01:28:33.458186 systemd[1]: Started sshd@4-143.198.155.16:22-139.178.89.65:45604.service - OpenSSH per-connection server daemon (139.178.89.65:45604). Mar 25 01:28:33.459597 systemd-logind[1463]: Removed session 4. Mar 25 01:28:33.528236 sshd[1640]: Accepted publickey for core from 139.178.89.65 port 45604 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:28:33.529942 sshd-session[1640]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:28:33.536711 systemd-logind[1463]: New session 5 of user core. Mar 25 01:28:33.543939 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 25 01:28:33.618519 sudo[1644]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 25 01:28:33.619092 sudo[1644]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 25 01:28:33.635166 sudo[1644]: pam_unix(sudo:session): session closed for user root Mar 25 01:28:33.638886 sshd[1643]: Connection closed by 139.178.89.65 port 45604 Mar 25 01:28:33.640193 sshd-session[1640]: pam_unix(sshd:session): session closed for user core Mar 25 01:28:33.657734 systemd[1]: sshd@4-143.198.155.16:22-139.178.89.65:45604.service: Deactivated successfully. Mar 25 01:28:33.660409 systemd[1]: session-5.scope: Deactivated successfully. Mar 25 01:28:33.661449 systemd-logind[1463]: Session 5 logged out. Waiting for processes to exit. Mar 25 01:28:33.665716 systemd[1]: Started sshd@5-143.198.155.16:22-139.178.89.65:45612.service - OpenSSH per-connection server daemon (139.178.89.65:45612). Mar 25 01:28:33.667754 systemd-logind[1463]: Removed session 5. Mar 25 01:28:33.748930 sshd[1649]: Accepted publickey for core from 139.178.89.65 port 45612 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:28:33.751462 sshd-session[1649]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:28:33.758851 systemd-logind[1463]: New session 6 of user core. Mar 25 01:28:33.770952 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 25 01:28:33.836819 sudo[1654]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 25 01:28:33.838444 sudo[1654]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 25 01:28:33.844337 sudo[1654]: pam_unix(sudo:session): session closed for user root Mar 25 01:28:33.854340 sudo[1653]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 25 01:28:33.854861 sudo[1653]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 25 01:28:33.874718 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 25 01:28:33.943659 augenrules[1676]: No rules Mar 25 01:28:33.944538 systemd[1]: audit-rules.service: Deactivated successfully. Mar 25 01:28:33.944876 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 25 01:28:33.946605 sudo[1653]: pam_unix(sudo:session): session closed for user root Mar 25 01:28:33.951398 sshd[1652]: Connection closed by 139.178.89.65 port 45612 Mar 25 01:28:33.952099 sshd-session[1649]: pam_unix(sshd:session): session closed for user core Mar 25 01:28:33.966852 systemd[1]: sshd@5-143.198.155.16:22-139.178.89.65:45612.service: Deactivated successfully. Mar 25 01:28:33.969617 systemd[1]: session-6.scope: Deactivated successfully. Mar 25 01:28:33.971853 systemd-logind[1463]: Session 6 logged out. Waiting for processes to exit. Mar 25 01:28:33.973725 systemd[1]: Started sshd@6-143.198.155.16:22-139.178.89.65:45618.service - OpenSSH per-connection server daemon (139.178.89.65:45618). Mar 25 01:28:33.975833 systemd-logind[1463]: Removed session 6. Mar 25 01:28:34.043041 sshd[1684]: Accepted publickey for core from 139.178.89.65 port 45618 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:28:34.045134 sshd-session[1684]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:28:34.051774 systemd-logind[1463]: New session 7 of user core. Mar 25 01:28:34.061083 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 25 01:28:34.123477 sudo[1688]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 25 01:28:34.124606 sudo[1688]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 25 01:28:34.640330 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 25 01:28:34.655370 (dockerd)[1705]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 25 01:28:35.220113 dockerd[1705]: time="2025-03-25T01:28:35.219690954Z" level=info msg="Starting up" Mar 25 01:28:35.224845 dockerd[1705]: time="2025-03-25T01:28:35.224773266Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Mar 25 01:28:35.273921 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2878817208-merged.mount: Deactivated successfully. Mar 25 01:28:35.334503 dockerd[1705]: time="2025-03-25T01:28:35.333751556Z" level=info msg="Loading containers: start." Mar 25 01:28:35.596769 kernel: Initializing XFRM netlink socket Mar 25 01:28:35.706433 systemd-networkd[1366]: docker0: Link UP Mar 25 01:28:35.767134 dockerd[1705]: time="2025-03-25T01:28:35.767027243Z" level=info msg="Loading containers: done." Mar 25 01:28:35.793736 dockerd[1705]: time="2025-03-25T01:28:35.793579251Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 25 01:28:35.794031 dockerd[1705]: time="2025-03-25T01:28:35.793796967Z" level=info msg="Docker daemon" commit=c710b88579fcb5e0d53f96dcae976d79323b9166 containerd-snapshotter=false storage-driver=overlay2 version=27.4.1 Mar 25 01:28:35.794031 dockerd[1705]: time="2025-03-25T01:28:35.793995258Z" level=info msg="Daemon has completed initialization" Mar 25 01:28:35.796799 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck781599647-merged.mount: Deactivated successfully. Mar 25 01:28:35.861115 dockerd[1705]: time="2025-03-25T01:28:35.860666408Z" level=info msg="API listen on /run/docker.sock" Mar 25 01:28:35.861373 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 25 01:28:36.872844 containerd[1487]: time="2025-03-25T01:28:36.872275372Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.3\"" Mar 25 01:28:37.395135 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3236637083.mount: Deactivated successfully. Mar 25 01:28:39.303707 containerd[1487]: time="2025-03-25T01:28:39.303577193Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:39.304718 containerd[1487]: time="2025-03-25T01:28:39.304595756Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.3: active requests=0, bytes read=28682430" Mar 25 01:28:39.307841 containerd[1487]: time="2025-03-25T01:28:39.307741688Z" level=info msg="ImageCreate event name:\"sha256:f8bdc4cfa0651e2d7edb4678d2b90129aef82a19249b37dc8d4705e8bd604295\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:39.309668 containerd[1487]: time="2025-03-25T01:28:39.308933722Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.3\" with image id \"sha256:f8bdc4cfa0651e2d7edb4678d2b90129aef82a19249b37dc8d4705e8bd604295\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:279e45cf07e4f56925c3c5237179eb63616788426a96e94df5fedf728b18926e\", size \"28679230\" in 2.436604812s" Mar 25 01:28:39.309668 containerd[1487]: time="2025-03-25T01:28:39.309004222Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.3\" returns image reference \"sha256:f8bdc4cfa0651e2d7edb4678d2b90129aef82a19249b37dc8d4705e8bd604295\"" Mar 25 01:28:39.310299 containerd[1487]: time="2025-03-25T01:28:39.310258928Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.3\"" Mar 25 01:28:39.311331 containerd[1487]: time="2025-03-25T01:28:39.310729748Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:279e45cf07e4f56925c3c5237179eb63616788426a96e94df5fedf728b18926e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:41.207764 containerd[1487]: time="2025-03-25T01:28:41.207699809Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:41.208939 containerd[1487]: time="2025-03-25T01:28:41.208879600Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.3: active requests=0, bytes read=24779684" Mar 25 01:28:41.211679 containerd[1487]: time="2025-03-25T01:28:41.209783506Z" level=info msg="ImageCreate event name:\"sha256:085818208a5213f37ef6d103caaf8e1e243816a614eb5b87a98bfffe79c687b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:41.213129 containerd[1487]: time="2025-03-25T01:28:41.213078763Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:54456a96a1bbdc35dcc2e70fcc1355bf655af67694e40b650ac12e83521f6411\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:41.214620 containerd[1487]: time="2025-03-25T01:28:41.214568808Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.3\" with image id \"sha256:085818208a5213f37ef6d103caaf8e1e243816a614eb5b87a98bfffe79c687b5\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:54456a96a1bbdc35dcc2e70fcc1355bf655af67694e40b650ac12e83521f6411\", size \"26267292\" in 1.903770879s" Mar 25 01:28:41.214758 containerd[1487]: time="2025-03-25T01:28:41.214620026Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.3\" returns image reference \"sha256:085818208a5213f37ef6d103caaf8e1e243816a614eb5b87a98bfffe79c687b5\"" Mar 25 01:28:41.216104 containerd[1487]: time="2025-03-25T01:28:41.216043402Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.3\"" Mar 25 01:28:42.757477 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 25 01:28:42.761509 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:28:42.931674 containerd[1487]: time="2025-03-25T01:28:42.931157085Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:42.933671 containerd[1487]: time="2025-03-25T01:28:42.933193867Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.3: active requests=0, bytes read=19171419" Mar 25 01:28:42.934553 containerd[1487]: time="2025-03-25T01:28:42.934489566Z" level=info msg="ImageCreate event name:\"sha256:b4260bf5078ab1b01dd05fb05015fc436b7100b7b9b5ea738e247a86008b16b8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:42.940862 containerd[1487]: time="2025-03-25T01:28:42.939214824Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.3\" with image id \"sha256:b4260bf5078ab1b01dd05fb05015fc436b7100b7b9b5ea738e247a86008b16b8\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:aafae2e3a8d65bc6dc3a0c6095c24bc72b1ff608e1417f0f5e860ce4a61c27df\", size \"20659045\" in 1.722945958s" Mar 25 01:28:42.940862 containerd[1487]: time="2025-03-25T01:28:42.939272718Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.3\" returns image reference \"sha256:b4260bf5078ab1b01dd05fb05015fc436b7100b7b9b5ea738e247a86008b16b8\"" Mar 25 01:28:42.940862 containerd[1487]: time="2025-03-25T01:28:42.939795132Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:aafae2e3a8d65bc6dc3a0c6095c24bc72b1ff608e1417f0f5e860ce4a61c27df\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:42.941122 containerd[1487]: time="2025-03-25T01:28:42.941013085Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.3\"" Mar 25 01:28:43.006323 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:28:43.019471 (kubelet)[1981]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 25 01:28:43.091347 systemd-resolved[1332]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. Mar 25 01:28:43.093739 kubelet[1981]: E0325 01:28:43.093604 1981 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 25 01:28:43.099059 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 25 01:28:43.099535 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 25 01:28:43.100240 systemd[1]: kubelet.service: Consumed 263ms CPU time, 104M memory peak. Mar 25 01:28:44.153462 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2599770463.mount: Deactivated successfully. Mar 25 01:28:44.858173 containerd[1487]: time="2025-03-25T01:28:44.858054828Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:44.859481 containerd[1487]: time="2025-03-25T01:28:44.859393905Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.3: active requests=0, bytes read=30918185" Mar 25 01:28:44.860551 containerd[1487]: time="2025-03-25T01:28:44.860228266Z" level=info msg="ImageCreate event name:\"sha256:a1ae78fd2f9d8fc345928378dc947c7f1e95f01c1a552781827071867a95d09c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:44.865536 containerd[1487]: time="2025-03-25T01:28:44.863842521Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:5015269547a0b7dd2c062758e9a64467b58978ff2502cad4c3f5cdf4aa554ad3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:44.865536 containerd[1487]: time="2025-03-25T01:28:44.865322048Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.3\" with image id \"sha256:a1ae78fd2f9d8fc345928378dc947c7f1e95f01c1a552781827071867a95d09c\", repo tag \"registry.k8s.io/kube-proxy:v1.32.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:5015269547a0b7dd2c062758e9a64467b58978ff2502cad4c3f5cdf4aa554ad3\", size \"30917204\" in 1.924270292s" Mar 25 01:28:44.865536 containerd[1487]: time="2025-03-25T01:28:44.865377305Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.3\" returns image reference \"sha256:a1ae78fd2f9d8fc345928378dc947c7f1e95f01c1a552781827071867a95d09c\"" Mar 25 01:28:44.866581 containerd[1487]: time="2025-03-25T01:28:44.866528382Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Mar 25 01:28:45.425541 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1671255665.mount: Deactivated successfully. Mar 25 01:28:46.146204 systemd-resolved[1332]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Mar 25 01:28:46.621706 containerd[1487]: time="2025-03-25T01:28:46.621417743Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:46.623354 containerd[1487]: time="2025-03-25T01:28:46.623257672Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Mar 25 01:28:46.625660 containerd[1487]: time="2025-03-25T01:28:46.624150298Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:46.627930 containerd[1487]: time="2025-03-25T01:28:46.627871915Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:46.629540 containerd[1487]: time="2025-03-25T01:28:46.629482251Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.762696346s" Mar 25 01:28:46.629826 containerd[1487]: time="2025-03-25T01:28:46.629799875Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Mar 25 01:28:46.630703 containerd[1487]: time="2025-03-25T01:28:46.630467668Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 25 01:28:47.130856 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2373997101.mount: Deactivated successfully. Mar 25 01:28:47.137127 containerd[1487]: time="2025-03-25T01:28:47.137065883Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 25 01:28:47.139640 containerd[1487]: time="2025-03-25T01:28:47.139489230Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Mar 25 01:28:47.140288 containerd[1487]: time="2025-03-25T01:28:47.140212434Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 25 01:28:47.146045 containerd[1487]: time="2025-03-25T01:28:47.145111478Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 25 01:28:47.146045 containerd[1487]: time="2025-03-25T01:28:47.145807636Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 514.872973ms" Mar 25 01:28:47.146045 containerd[1487]: time="2025-03-25T01:28:47.145846736Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Mar 25 01:28:47.147238 containerd[1487]: time="2025-03-25T01:28:47.147171551Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Mar 25 01:28:47.772624 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2340644301.mount: Deactivated successfully. Mar 25 01:28:50.511472 containerd[1487]: time="2025-03-25T01:28:50.511398971Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:50.513777 containerd[1487]: time="2025-03-25T01:28:50.512239008Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551320" Mar 25 01:28:50.514072 containerd[1487]: time="2025-03-25T01:28:50.514023279Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:50.524927 containerd[1487]: time="2025-03-25T01:28:50.524866626Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:28:50.526279 containerd[1487]: time="2025-03-25T01:28:50.526230637Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 3.378763537s" Mar 25 01:28:50.526543 containerd[1487]: time="2025-03-25T01:28:50.526504193Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Mar 25 01:28:53.255983 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 25 01:28:53.261005 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:28:53.481179 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:28:53.495348 (kubelet)[2135]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 25 01:28:53.563734 kubelet[2135]: E0325 01:28:53.562987 2135 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 25 01:28:53.568219 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 25 01:28:53.568481 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 25 01:28:53.571714 systemd[1]: kubelet.service: Consumed 219ms CPU time, 106.1M memory peak. Mar 25 01:28:53.733820 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:28:53.734239 systemd[1]: kubelet.service: Consumed 219ms CPU time, 106.1M memory peak. Mar 25 01:28:53.738057 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:28:53.799305 systemd[1]: Reload requested from client PID 2150 ('systemctl') (unit session-7.scope)... Mar 25 01:28:53.799328 systemd[1]: Reloading... Mar 25 01:28:53.975663 zram_generator::config[2194]: No configuration found. Mar 25 01:28:54.153186 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 25 01:28:54.329823 systemd[1]: Reloading finished in 529 ms. Mar 25 01:28:54.412706 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:28:54.418433 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:28:54.420821 systemd[1]: kubelet.service: Deactivated successfully. Mar 25 01:28:54.421285 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:28:54.421453 systemd[1]: kubelet.service: Consumed 147ms CPU time, 91.8M memory peak. Mar 25 01:28:54.424534 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:28:54.602900 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:28:54.617755 (kubelet)[2250]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 25 01:28:54.696298 kubelet[2250]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 25 01:28:54.696298 kubelet[2250]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 25 01:28:54.696298 kubelet[2250]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 25 01:28:54.697523 kubelet[2250]: I0325 01:28:54.696394 2250 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 25 01:28:55.460407 kubelet[2250]: I0325 01:28:55.460335 2250 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Mar 25 01:28:55.460407 kubelet[2250]: I0325 01:28:55.460391 2250 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 25 01:28:55.461665 kubelet[2250]: I0325 01:28:55.461130 2250 server.go:954] "Client rotation is on, will bootstrap in background" Mar 25 01:28:55.504805 kubelet[2250]: I0325 01:28:55.504754 2250 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 25 01:28:55.511536 kubelet[2250]: E0325 01:28:55.511456 2250 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://143.198.155.16:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:55.533060 kubelet[2250]: I0325 01:28:55.532917 2250 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 25 01:28:55.541022 kubelet[2250]: I0325 01:28:55.540498 2250 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 25 01:28:55.547242 kubelet[2250]: I0325 01:28:55.546770 2250 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 25 01:28:55.547242 kubelet[2250]: I0325 01:28:55.546879 2250 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4284.0.0-c-20cdc03f82","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 25 01:28:55.550196 kubelet[2250]: I0325 01:28:55.550121 2250 topology_manager.go:138] "Creating topology manager with none policy" Mar 25 01:28:55.551079 kubelet[2250]: I0325 01:28:55.550417 2250 container_manager_linux.go:304] "Creating device plugin manager" Mar 25 01:28:55.551079 kubelet[2250]: I0325 01:28:55.550753 2250 state_mem.go:36] "Initialized new in-memory state store" Mar 25 01:28:55.558158 kubelet[2250]: I0325 01:28:55.557399 2250 kubelet.go:446] "Attempting to sync node with API server" Mar 25 01:28:55.558158 kubelet[2250]: I0325 01:28:55.557456 2250 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 25 01:28:55.558158 kubelet[2250]: I0325 01:28:55.557503 2250 kubelet.go:352] "Adding apiserver pod source" Mar 25 01:28:55.558158 kubelet[2250]: I0325 01:28:55.557520 2250 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 25 01:28:55.568850 kubelet[2250]: I0325 01:28:55.568803 2250 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" Mar 25 01:28:55.572930 kubelet[2250]: W0325 01:28:55.572855 2250 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://143.198.155.16:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 143.198.155.16:6443: connect: connection refused Mar 25 01:28:55.573577 kubelet[2250]: E0325 01:28:55.573540 2250 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://143.198.155.16:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:55.574419 kubelet[2250]: W0325 01:28:55.574352 2250 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://143.198.155.16:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4284.0.0-c-20cdc03f82&limit=500&resourceVersion=0": dial tcp 143.198.155.16:6443: connect: connection refused Mar 25 01:28:55.575715 kubelet[2250]: E0325 01:28:55.574590 2250 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://143.198.155.16:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4284.0.0-c-20cdc03f82&limit=500&resourceVersion=0\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:55.575715 kubelet[2250]: I0325 01:28:55.575059 2250 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 25 01:28:55.575715 kubelet[2250]: W0325 01:28:55.575221 2250 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 25 01:28:55.577509 kubelet[2250]: I0325 01:28:55.577481 2250 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 25 01:28:55.577745 kubelet[2250]: I0325 01:28:55.577728 2250 server.go:1287] "Started kubelet" Mar 25 01:28:55.582282 kubelet[2250]: I0325 01:28:55.581227 2250 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Mar 25 01:28:55.584448 kubelet[2250]: I0325 01:28:55.584396 2250 server.go:490] "Adding debug handlers to kubelet server" Mar 25 01:28:55.589595 kubelet[2250]: I0325 01:28:55.588920 2250 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 25 01:28:55.589595 kubelet[2250]: I0325 01:28:55.589470 2250 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 25 01:28:55.595133 kubelet[2250]: E0325 01:28:55.592900 2250 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://143.198.155.16:6443/api/v1/namespaces/default/events\": dial tcp 143.198.155.16:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4284.0.0-c-20cdc03f82.182fe784262a61d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4284.0.0-c-20cdc03f82,UID:ci-4284.0.0-c-20cdc03f82,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4284.0.0-c-20cdc03f82,},FirstTimestamp:2025-03-25 01:28:55.577682387 +0000 UTC m=+0.952720780,LastTimestamp:2025-03-25 01:28:55.577682387 +0000 UTC m=+0.952720780,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4284.0.0-c-20cdc03f82,}" Mar 25 01:28:55.597918 kubelet[2250]: I0325 01:28:55.597800 2250 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 25 01:28:55.598708 kubelet[2250]: I0325 01:28:55.598677 2250 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 25 01:28:55.602589 kubelet[2250]: E0325 01:28:55.602017 2250 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" Mar 25 01:28:55.602589 kubelet[2250]: I0325 01:28:55.602078 2250 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 25 01:28:55.603895 kubelet[2250]: I0325 01:28:55.603860 2250 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Mar 25 01:28:55.604204 kubelet[2250]: I0325 01:28:55.604183 2250 reconciler.go:26] "Reconciler: start to sync state" Mar 25 01:28:55.604991 kubelet[2250]: W0325 01:28:55.604931 2250 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://143.198.155.16:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.155.16:6443: connect: connection refused Mar 25 01:28:55.605184 kubelet[2250]: E0325 01:28:55.605155 2250 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://143.198.155.16:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:55.605863 kubelet[2250]: E0325 01:28:55.605816 2250 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.155.16:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-c-20cdc03f82?timeout=10s\": dial tcp 143.198.155.16:6443: connect: connection refused" interval="200ms" Mar 25 01:28:55.612219 kubelet[2250]: I0325 01:28:55.611581 2250 factory.go:221] Registration of the systemd container factory successfully Mar 25 01:28:55.612219 kubelet[2250]: I0325 01:28:55.611804 2250 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 25 01:28:55.616665 kubelet[2250]: I0325 01:28:55.616583 2250 factory.go:221] Registration of the containerd container factory successfully Mar 25 01:28:55.634070 kubelet[2250]: I0325 01:28:55.634028 2250 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 25 01:28:55.634070 kubelet[2250]: I0325 01:28:55.634056 2250 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 25 01:28:55.634316 kubelet[2250]: I0325 01:28:55.634088 2250 state_mem.go:36] "Initialized new in-memory state store" Mar 25 01:28:55.637086 kubelet[2250]: I0325 01:28:55.637042 2250 policy_none.go:49] "None policy: Start" Mar 25 01:28:55.637086 kubelet[2250]: I0325 01:28:55.637097 2250 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 25 01:28:55.637304 kubelet[2250]: I0325 01:28:55.637121 2250 state_mem.go:35] "Initializing new in-memory state store" Mar 25 01:28:55.656841 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 25 01:28:55.660904 kubelet[2250]: I0325 01:28:55.660571 2250 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 25 01:28:55.663716 kubelet[2250]: I0325 01:28:55.663671 2250 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 25 01:28:55.663716 kubelet[2250]: I0325 01:28:55.663762 2250 status_manager.go:227] "Starting to sync pod status with apiserver" Mar 25 01:28:55.663716 kubelet[2250]: I0325 01:28:55.663795 2250 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 25 01:28:55.664161 kubelet[2250]: I0325 01:28:55.664093 2250 kubelet.go:2388] "Starting kubelet main sync loop" Mar 25 01:28:55.664307 kubelet[2250]: E0325 01:28:55.664268 2250 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 25 01:28:55.670467 kubelet[2250]: W0325 01:28:55.670320 2250 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://143.198.155.16:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.155.16:6443: connect: connection refused Mar 25 01:28:55.670467 kubelet[2250]: E0325 01:28:55.670411 2250 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://143.198.155.16:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:55.680571 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 25 01:28:55.688893 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 25 01:28:55.699684 kubelet[2250]: I0325 01:28:55.699571 2250 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 25 01:28:55.701550 kubelet[2250]: I0325 01:28:55.701060 2250 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 25 01:28:55.701550 kubelet[2250]: I0325 01:28:55.701093 2250 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 25 01:28:55.701550 kubelet[2250]: I0325 01:28:55.701556 2250 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 25 01:28:55.706414 kubelet[2250]: E0325 01:28:55.706361 2250 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 25 01:28:55.706837 kubelet[2250]: E0325 01:28:55.706794 2250 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4284.0.0-c-20cdc03f82\" not found" Mar 25 01:28:55.779669 systemd[1]: Created slice kubepods-burstable-pod5dd43591be4d9acd85e69e7f9100fc57.slice - libcontainer container kubepods-burstable-pod5dd43591be4d9acd85e69e7f9100fc57.slice. Mar 25 01:28:55.803454 kubelet[2250]: I0325 01:28:55.802666 2250 kubelet_node_status.go:76] "Attempting to register node" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.803454 kubelet[2250]: E0325 01:28:55.803319 2250 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://143.198.155.16:6443/api/v1/nodes\": dial tcp 143.198.155.16:6443: connect: connection refused" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805223 kubelet[2250]: I0325 01:28:55.804815 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4ae26885669d92b3428c37e4f6d2d17a-kubeconfig\") pod \"kube-scheduler-ci-4284.0.0-c-20cdc03f82\" (UID: \"4ae26885669d92b3428c37e4f6d2d17a\") " pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805223 kubelet[2250]: I0325 01:28:55.804901 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5dd43591be4d9acd85e69e7f9100fc57-ca-certs\") pod \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" (UID: \"5dd43591be4d9acd85e69e7f9100fc57\") " pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805223 kubelet[2250]: I0325 01:28:55.804928 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5dd43591be4d9acd85e69e7f9100fc57-k8s-certs\") pod \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" (UID: \"5dd43591be4d9acd85e69e7f9100fc57\") " pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805223 kubelet[2250]: I0325 01:28:55.804952 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5dd43591be4d9acd85e69e7f9100fc57-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" (UID: \"5dd43591be4d9acd85e69e7f9100fc57\") " pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805223 kubelet[2250]: I0325 01:28:55.805006 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-k8s-certs\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805740 kubelet[2250]: I0325 01:28:55.805033 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805740 kubelet[2250]: I0325 01:28:55.805078 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-ca-certs\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805740 kubelet[2250]: I0325 01:28:55.805100 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-flexvolume-dir\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805740 kubelet[2250]: I0325 01:28:55.805150 2250 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-kubeconfig\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.805740 kubelet[2250]: E0325 01:28:55.805584 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.806754 kubelet[2250]: E0325 01:28:55.806673 2250 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.155.16:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-c-20cdc03f82?timeout=10s\": dial tcp 143.198.155.16:6443: connect: connection refused" interval="400ms" Mar 25 01:28:55.811004 systemd[1]: Created slice kubepods-burstable-podaff1e0550efee4d3ea6a70f023943fac.slice - libcontainer container kubepods-burstable-podaff1e0550efee4d3ea6a70f023943fac.slice. Mar 25 01:28:55.820816 kubelet[2250]: E0325 01:28:55.820302 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:55.830477 systemd[1]: Created slice kubepods-burstable-pod4ae26885669d92b3428c37e4f6d2d17a.slice - libcontainer container kubepods-burstable-pod4ae26885669d92b3428c37e4f6d2d17a.slice. Mar 25 01:28:55.833380 kubelet[2250]: E0325 01:28:55.833337 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:56.011336 kubelet[2250]: I0325 01:28:56.011293 2250 kubelet_node_status.go:76] "Attempting to register node" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:56.012260 kubelet[2250]: E0325 01:28:56.012224 2250 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://143.198.155.16:6443/api/v1/nodes\": dial tcp 143.198.155.16:6443: connect: connection refused" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:56.106751 kubelet[2250]: E0325 01:28:56.106607 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:56.107969 containerd[1487]: time="2025-03-25T01:28:56.107921070Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4284.0.0-c-20cdc03f82,Uid:5dd43591be4d9acd85e69e7f9100fc57,Namespace:kube-system,Attempt:0,}" Mar 25 01:28:56.122015 kubelet[2250]: E0325 01:28:56.121232 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:56.130403 containerd[1487]: time="2025-03-25T01:28:56.130024041Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4284.0.0-c-20cdc03f82,Uid:aff1e0550efee4d3ea6a70f023943fac,Namespace:kube-system,Attempt:0,}" Mar 25 01:28:56.136756 kubelet[2250]: E0325 01:28:56.136710 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:56.140458 containerd[1487]: time="2025-03-25T01:28:56.138797316Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4284.0.0-c-20cdc03f82,Uid:4ae26885669d92b3428c37e4f6d2d17a,Namespace:kube-system,Attempt:0,}" Mar 25 01:28:56.208701 kubelet[2250]: E0325 01:28:56.207888 2250 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.155.16:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-c-20cdc03f82?timeout=10s\": dial tcp 143.198.155.16:6443: connect: connection refused" interval="800ms" Mar 25 01:28:56.257036 containerd[1487]: time="2025-03-25T01:28:56.256495750Z" level=info msg="connecting to shim 24d331d20d1020fe9ae6517dbe6f683b4bc4c1ae07a42c7081853cd272743dfb" address="unix:///run/containerd/s/c107543f6f4af8098b70ea559cdfb9c2c012fa21ff67a65b49352ab53eb36fc7" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:28:56.269528 containerd[1487]: time="2025-03-25T01:28:56.269450501Z" level=info msg="connecting to shim e0a7b8bd90a70f8bf9b676bde3b0988258c8f5872d1a813c3905b91cb1c58191" address="unix:///run/containerd/s/29fea89eee7ea55c13a346588383dd300db416a2e7bc1b58027303d24799b109" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:28:56.271520 containerd[1487]: time="2025-03-25T01:28:56.271334332Z" level=info msg="connecting to shim 6af16da2d4d7fee3476efe75f1c61610c5e4512fc34d666ad34033521a0ec94e" address="unix:///run/containerd/s/b87a0b6b6986c2c1caf90f54dfdb69fb6304a51673a0bec3f65a716052c4d246" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:28:56.388481 kubelet[2250]: W0325 01:28:56.387547 2250 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://143.198.155.16:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 143.198.155.16:6443: connect: connection refused Mar 25 01:28:56.388481 kubelet[2250]: E0325 01:28:56.387652 2250 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://143.198.155.16:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:56.404912 systemd[1]: Started cri-containerd-24d331d20d1020fe9ae6517dbe6f683b4bc4c1ae07a42c7081853cd272743dfb.scope - libcontainer container 24d331d20d1020fe9ae6517dbe6f683b4bc4c1ae07a42c7081853cd272743dfb. Mar 25 01:28:56.406514 systemd[1]: Started cri-containerd-6af16da2d4d7fee3476efe75f1c61610c5e4512fc34d666ad34033521a0ec94e.scope - libcontainer container 6af16da2d4d7fee3476efe75f1c61610c5e4512fc34d666ad34033521a0ec94e. Mar 25 01:28:56.408515 systemd[1]: Started cri-containerd-e0a7b8bd90a70f8bf9b676bde3b0988258c8f5872d1a813c3905b91cb1c58191.scope - libcontainer container e0a7b8bd90a70f8bf9b676bde3b0988258c8f5872d1a813c3905b91cb1c58191. Mar 25 01:28:56.417937 kubelet[2250]: I0325 01:28:56.417142 2250 kubelet_node_status.go:76] "Attempting to register node" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:56.420312 kubelet[2250]: E0325 01:28:56.420058 2250 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://143.198.155.16:6443/api/v1/nodes\": dial tcp 143.198.155.16:6443: connect: connection refused" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:56.469924 kubelet[2250]: W0325 01:28:56.469795 2250 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://143.198.155.16:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.155.16:6443: connect: connection refused Mar 25 01:28:56.469924 kubelet[2250]: E0325 01:28:56.469872 2250 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://143.198.155.16:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:56.545146 containerd[1487]: time="2025-03-25T01:28:56.545084813Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4284.0.0-c-20cdc03f82,Uid:5dd43591be4d9acd85e69e7f9100fc57,Namespace:kube-system,Attempt:0,} returns sandbox id \"e0a7b8bd90a70f8bf9b676bde3b0988258c8f5872d1a813c3905b91cb1c58191\"" Mar 25 01:28:56.549368 containerd[1487]: time="2025-03-25T01:28:56.548314296Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4284.0.0-c-20cdc03f82,Uid:aff1e0550efee4d3ea6a70f023943fac,Namespace:kube-system,Attempt:0,} returns sandbox id \"24d331d20d1020fe9ae6517dbe6f683b4bc4c1ae07a42c7081853cd272743dfb\"" Mar 25 01:28:56.552075 kubelet[2250]: E0325 01:28:56.551987 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:56.555951 kubelet[2250]: E0325 01:28:56.555262 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:56.561231 containerd[1487]: time="2025-03-25T01:28:56.558734747Z" level=info msg="CreateContainer within sandbox \"e0a7b8bd90a70f8bf9b676bde3b0988258c8f5872d1a813c3905b91cb1c58191\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 25 01:28:56.563349 containerd[1487]: time="2025-03-25T01:28:56.563258442Z" level=info msg="CreateContainer within sandbox \"24d331d20d1020fe9ae6517dbe6f683b4bc4c1ae07a42c7081853cd272743dfb\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 25 01:28:56.589858 containerd[1487]: time="2025-03-25T01:28:56.589057845Z" level=info msg="Container 221adfc128d6d78274aee339e9fc1ff00fbc7d06b97e2e4aa3a182b748a0c673: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:28:56.590980 containerd[1487]: time="2025-03-25T01:28:56.589079793Z" level=info msg="Container 3c40de6c3fac4e65f59ab9156160e8cc038f04d09d4fcda525a0fe7479c7465a: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:28:56.595713 containerd[1487]: time="2025-03-25T01:28:56.595590779Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4284.0.0-c-20cdc03f82,Uid:4ae26885669d92b3428c37e4f6d2d17a,Namespace:kube-system,Attempt:0,} returns sandbox id \"6af16da2d4d7fee3476efe75f1c61610c5e4512fc34d666ad34033521a0ec94e\"" Mar 25 01:28:56.597560 kubelet[2250]: E0325 01:28:56.596934 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:56.612047 containerd[1487]: time="2025-03-25T01:28:56.611829347Z" level=info msg="CreateContainer within sandbox \"e0a7b8bd90a70f8bf9b676bde3b0988258c8f5872d1a813c3905b91cb1c58191\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3c40de6c3fac4e65f59ab9156160e8cc038f04d09d4fcda525a0fe7479c7465a\"" Mar 25 01:28:56.613387 containerd[1487]: time="2025-03-25T01:28:56.613260748Z" level=info msg="CreateContainer within sandbox \"24d331d20d1020fe9ae6517dbe6f683b4bc4c1ae07a42c7081853cd272743dfb\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"221adfc128d6d78274aee339e9fc1ff00fbc7d06b97e2e4aa3a182b748a0c673\"" Mar 25 01:28:56.614670 containerd[1487]: time="2025-03-25T01:28:56.613721044Z" level=info msg="CreateContainer within sandbox \"6af16da2d4d7fee3476efe75f1c61610c5e4512fc34d666ad34033521a0ec94e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 25 01:28:56.615520 containerd[1487]: time="2025-03-25T01:28:56.615478100Z" level=info msg="StartContainer for \"221adfc128d6d78274aee339e9fc1ff00fbc7d06b97e2e4aa3a182b748a0c673\"" Mar 25 01:28:56.617450 containerd[1487]: time="2025-03-25T01:28:56.615760392Z" level=info msg="StartContainer for \"3c40de6c3fac4e65f59ab9156160e8cc038f04d09d4fcda525a0fe7479c7465a\"" Mar 25 01:28:56.618731 containerd[1487]: time="2025-03-25T01:28:56.618681495Z" level=info msg="connecting to shim 221adfc128d6d78274aee339e9fc1ff00fbc7d06b97e2e4aa3a182b748a0c673" address="unix:///run/containerd/s/c107543f6f4af8098b70ea559cdfb9c2c012fa21ff67a65b49352ab53eb36fc7" protocol=ttrpc version=3 Mar 25 01:28:56.625598 containerd[1487]: time="2025-03-25T01:28:56.625528404Z" level=info msg="connecting to shim 3c40de6c3fac4e65f59ab9156160e8cc038f04d09d4fcda525a0fe7479c7465a" address="unix:///run/containerd/s/29fea89eee7ea55c13a346588383dd300db416a2e7bc1b58027303d24799b109" protocol=ttrpc version=3 Mar 25 01:28:56.638446 containerd[1487]: time="2025-03-25T01:28:56.638372579Z" level=info msg="Container 8df0e50d82bac6fa6c992b624ac3866ddd414e98d72023300b3d4ab9561ea825: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:28:56.656670 containerd[1487]: time="2025-03-25T01:28:56.655619548Z" level=info msg="CreateContainer within sandbox \"6af16da2d4d7fee3476efe75f1c61610c5e4512fc34d666ad34033521a0ec94e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"8df0e50d82bac6fa6c992b624ac3866ddd414e98d72023300b3d4ab9561ea825\"" Mar 25 01:28:56.656670 containerd[1487]: time="2025-03-25T01:28:56.656535151Z" level=info msg="StartContainer for \"8df0e50d82bac6fa6c992b624ac3866ddd414e98d72023300b3d4ab9561ea825\"" Mar 25 01:28:56.658277 systemd[1]: Started cri-containerd-221adfc128d6d78274aee339e9fc1ff00fbc7d06b97e2e4aa3a182b748a0c673.scope - libcontainer container 221adfc128d6d78274aee339e9fc1ff00fbc7d06b97e2e4aa3a182b748a0c673. Mar 25 01:28:56.663164 kubelet[2250]: W0325 01:28:56.661354 2250 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://143.198.155.16:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.155.16:6443: connect: connection refused Mar 25 01:28:56.663164 kubelet[2250]: E0325 01:28:56.662218 2250 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://143.198.155.16:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 143.198.155.16:6443: connect: connection refused" logger="UnhandledError" Mar 25 01:28:56.663362 containerd[1487]: time="2025-03-25T01:28:56.663164565Z" level=info msg="connecting to shim 8df0e50d82bac6fa6c992b624ac3866ddd414e98d72023300b3d4ab9561ea825" address="unix:///run/containerd/s/b87a0b6b6986c2c1caf90f54dfdb69fb6304a51673a0bec3f65a716052c4d246" protocol=ttrpc version=3 Mar 25 01:28:56.698010 systemd[1]: Started cri-containerd-3c40de6c3fac4e65f59ab9156160e8cc038f04d09d4fcda525a0fe7479c7465a.scope - libcontainer container 3c40de6c3fac4e65f59ab9156160e8cc038f04d09d4fcda525a0fe7479c7465a. Mar 25 01:28:56.699758 systemd[1]: Started cri-containerd-8df0e50d82bac6fa6c992b624ac3866ddd414e98d72023300b3d4ab9561ea825.scope - libcontainer container 8df0e50d82bac6fa6c992b624ac3866ddd414e98d72023300b3d4ab9561ea825. Mar 25 01:28:56.778543 containerd[1487]: time="2025-03-25T01:28:56.778274869Z" level=info msg="StartContainer for \"221adfc128d6d78274aee339e9fc1ff00fbc7d06b97e2e4aa3a182b748a0c673\" returns successfully" Mar 25 01:28:56.838811 containerd[1487]: time="2025-03-25T01:28:56.838488039Z" level=info msg="StartContainer for \"3c40de6c3fac4e65f59ab9156160e8cc038f04d09d4fcda525a0fe7479c7465a\" returns successfully" Mar 25 01:28:56.864668 containerd[1487]: time="2025-03-25T01:28:56.864547262Z" level=info msg="StartContainer for \"8df0e50d82bac6fa6c992b624ac3866ddd414e98d72023300b3d4ab9561ea825\" returns successfully" Mar 25 01:28:57.009848 kubelet[2250]: E0325 01:28:57.009786 2250 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.155.16:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-c-20cdc03f82?timeout=10s\": dial tcp 143.198.155.16:6443: connect: connection refused" interval="1.6s" Mar 25 01:28:57.223673 kubelet[2250]: I0325 01:28:57.221823 2250 kubelet_node_status.go:76] "Attempting to register node" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:57.713995 kubelet[2250]: E0325 01:28:57.713844 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:57.714249 kubelet[2250]: E0325 01:28:57.714119 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:57.718067 kubelet[2250]: E0325 01:28:57.718004 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:57.718224 kubelet[2250]: E0325 01:28:57.718149 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:57.722096 kubelet[2250]: E0325 01:28:57.722052 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:57.722275 kubelet[2250]: E0325 01:28:57.722255 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:58.724181 kubelet[2250]: E0325 01:28:58.724129 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:58.725937 kubelet[2250]: E0325 01:28:58.724265 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:58.725937 kubelet[2250]: E0325 01:28:58.724664 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:58.725937 kubelet[2250]: E0325 01:28:58.724764 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:58.725937 kubelet[2250]: E0325 01:28:58.725179 2250 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:58.725937 kubelet[2250]: E0325 01:28:58.725316 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:59.133621 kubelet[2250]: E0325 01:28:59.133565 2250 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4284.0.0-c-20cdc03f82\" not found" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.268807 kubelet[2250]: I0325 01:28:59.268736 2250 kubelet_node_status.go:79] "Successfully registered node" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.305883 kubelet[2250]: I0325 01:28:59.305825 2250 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.314275 kubelet[2250]: E0325 01:28:59.314209 2250 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.314275 kubelet[2250]: I0325 01:28:59.314256 2250 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.316470 kubelet[2250]: E0325 01:28:59.316424 2250 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4284.0.0-c-20cdc03f82\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.316470 kubelet[2250]: I0325 01:28:59.316457 2250 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.321062 kubelet[2250]: E0325 01:28:59.320997 2250 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.576308 kubelet[2250]: I0325 01:28:59.576245 2250 apiserver.go:52] "Watching apiserver" Mar 25 01:28:59.604957 kubelet[2250]: I0325 01:28:59.604889 2250 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Mar 25 01:28:59.726286 kubelet[2250]: I0325 01:28:59.725417 2250 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.726286 kubelet[2250]: I0325 01:28:59.726052 2250 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.729105 kubelet[2250]: E0325 01:28:59.729057 2250 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4284.0.0-c-20cdc03f82\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.729760 kubelet[2250]: E0325 01:28:59.729537 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:28:59.731869 kubelet[2250]: E0325 01:28:59.731793 2250 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:28:59.732117 kubelet[2250]: E0325 01:28:59.732092 2250 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:01.588419 systemd[1]: Reload requested from client PID 2516 ('systemctl') (unit session-7.scope)... Mar 25 01:29:01.588444 systemd[1]: Reloading... Mar 25 01:29:01.840223 zram_generator::config[2563]: No configuration found. Mar 25 01:29:02.057165 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 25 01:29:02.242031 systemd[1]: Reloading finished in 652 ms. Mar 25 01:29:02.283804 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:29:02.300243 systemd[1]: kubelet.service: Deactivated successfully. Mar 25 01:29:02.301493 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:29:02.301592 systemd[1]: kubelet.service: Consumed 1.459s CPU time, 123.1M memory peak. Mar 25 01:29:02.305811 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 25 01:29:02.500853 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 25 01:29:02.514778 (kubelet)[2611]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 25 01:29:02.611556 kubelet[2611]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 25 01:29:02.612707 kubelet[2611]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 25 01:29:02.612707 kubelet[2611]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 25 01:29:02.613204 kubelet[2611]: I0325 01:29:02.612622 2611 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 25 01:29:02.624714 kubelet[2611]: I0325 01:29:02.623875 2611 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Mar 25 01:29:02.624714 kubelet[2611]: I0325 01:29:02.623914 2611 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 25 01:29:02.624714 kubelet[2611]: I0325 01:29:02.624299 2611 server.go:954] "Client rotation is on, will bootstrap in background" Mar 25 01:29:02.628703 kubelet[2611]: I0325 01:29:02.628398 2611 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 25 01:29:02.632605 kubelet[2611]: I0325 01:29:02.632413 2611 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 25 01:29:02.661712 kubelet[2611]: I0325 01:29:02.660850 2611 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 25 01:29:02.666209 sudo[2625]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 25 01:29:02.667844 sudo[2625]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Mar 25 01:29:02.674788 kubelet[2611]: I0325 01:29:02.674670 2611 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 25 01:29:02.675686 kubelet[2611]: I0325 01:29:02.675153 2611 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 25 01:29:02.675686 kubelet[2611]: I0325 01:29:02.675210 2611 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4284.0.0-c-20cdc03f82","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 25 01:29:02.675686 kubelet[2611]: I0325 01:29:02.675483 2611 topology_manager.go:138] "Creating topology manager with none policy" Mar 25 01:29:02.675686 kubelet[2611]: I0325 01:29:02.675500 2611 container_manager_linux.go:304] "Creating device plugin manager" Mar 25 01:29:02.676079 kubelet[2611]: I0325 01:29:02.675571 2611 state_mem.go:36] "Initialized new in-memory state store" Mar 25 01:29:02.676670 kubelet[2611]: I0325 01:29:02.676364 2611 kubelet.go:446] "Attempting to sync node with API server" Mar 25 01:29:02.676670 kubelet[2611]: I0325 01:29:02.676388 2611 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 25 01:29:02.677867 kubelet[2611]: I0325 01:29:02.677850 2611 kubelet.go:352] "Adding apiserver pod source" Mar 25 01:29:02.678164 kubelet[2611]: I0325 01:29:02.678079 2611 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 25 01:29:02.688892 kubelet[2611]: I0325 01:29:02.688857 2611 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" Mar 25 01:29:02.691654 kubelet[2611]: I0325 01:29:02.691122 2611 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 25 01:29:02.691818 kubelet[2611]: I0325 01:29:02.691803 2611 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 25 01:29:02.691912 kubelet[2611]: I0325 01:29:02.691903 2611 server.go:1287] "Started kubelet" Mar 25 01:29:02.698305 kubelet[2611]: I0325 01:29:02.696208 2611 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Mar 25 01:29:02.700655 kubelet[2611]: I0325 01:29:02.700101 2611 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 25 01:29:02.714532 kubelet[2611]: I0325 01:29:02.714201 2611 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 25 01:29:02.722525 kubelet[2611]: I0325 01:29:02.702001 2611 server.go:490] "Adding debug handlers to kubelet server" Mar 25 01:29:02.724110 kubelet[2611]: I0325 01:29:02.724076 2611 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 25 01:29:02.728586 kubelet[2611]: I0325 01:29:02.702044 2611 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 25 01:29:02.728586 kubelet[2611]: I0325 01:29:02.727771 2611 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 25 01:29:02.728586 kubelet[2611]: I0325 01:29:02.728208 2611 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Mar 25 01:29:02.728586 kubelet[2611]: I0325 01:29:02.728362 2611 reconciler.go:26] "Reconciler: start to sync state" Mar 25 01:29:02.736692 kubelet[2611]: I0325 01:29:02.736595 2611 factory.go:221] Registration of the systemd container factory successfully Mar 25 01:29:02.736860 kubelet[2611]: I0325 01:29:02.736795 2611 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 25 01:29:02.742134 kubelet[2611]: E0325 01:29:02.741526 2611 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 25 01:29:02.751432 kubelet[2611]: I0325 01:29:02.750769 2611 factory.go:221] Registration of the containerd container factory successfully Mar 25 01:29:02.758677 kubelet[2611]: I0325 01:29:02.757242 2611 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 25 01:29:02.773799 kubelet[2611]: I0325 01:29:02.772143 2611 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 25 01:29:02.773799 kubelet[2611]: I0325 01:29:02.772196 2611 status_manager.go:227] "Starting to sync pod status with apiserver" Mar 25 01:29:02.773799 kubelet[2611]: I0325 01:29:02.772223 2611 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 25 01:29:02.773799 kubelet[2611]: I0325 01:29:02.772233 2611 kubelet.go:2388] "Starting kubelet main sync loop" Mar 25 01:29:02.773799 kubelet[2611]: E0325 01:29:02.772314 2611 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 25 01:29:02.872430 kubelet[2611]: E0325 01:29:02.872374 2611 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 25 01:29:02.880912 kubelet[2611]: I0325 01:29:02.880713 2611 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 25 01:29:02.880912 kubelet[2611]: I0325 01:29:02.880752 2611 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 25 01:29:02.880912 kubelet[2611]: I0325 01:29:02.880777 2611 state_mem.go:36] "Initialized new in-memory state store" Mar 25 01:29:02.881785 kubelet[2611]: I0325 01:29:02.881380 2611 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 25 01:29:02.881785 kubelet[2611]: I0325 01:29:02.881400 2611 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 25 01:29:02.881785 kubelet[2611]: I0325 01:29:02.881431 2611 policy_none.go:49] "None policy: Start" Mar 25 01:29:02.881785 kubelet[2611]: I0325 01:29:02.881446 2611 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 25 01:29:02.881785 kubelet[2611]: I0325 01:29:02.881466 2611 state_mem.go:35] "Initializing new in-memory state store" Mar 25 01:29:02.881785 kubelet[2611]: I0325 01:29:02.881616 2611 state_mem.go:75] "Updated machine memory state" Mar 25 01:29:02.897920 kubelet[2611]: I0325 01:29:02.894302 2611 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 25 01:29:02.897920 kubelet[2611]: I0325 01:29:02.894547 2611 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 25 01:29:02.897920 kubelet[2611]: I0325 01:29:02.894565 2611 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 25 01:29:02.897920 kubelet[2611]: I0325 01:29:02.895259 2611 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 25 01:29:02.911797 kubelet[2611]: E0325 01:29:02.911750 2611 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 25 01:29:03.018677 kubelet[2611]: I0325 01:29:03.016381 2611 kubelet_node_status.go:76] "Attempting to register node" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.034672 kubelet[2611]: I0325 01:29:03.033792 2611 kubelet_node_status.go:125] "Node was previously registered" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.035082 kubelet[2611]: I0325 01:29:03.034866 2611 kubelet_node_status.go:79] "Successfully registered node" node="ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.074704 kubelet[2611]: I0325 01:29:03.074374 2611 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.074948 kubelet[2611]: I0325 01:29:03.074929 2611 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.075265 kubelet[2611]: I0325 01:29:03.075248 2611 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.103835 kubelet[2611]: W0325 01:29:03.103783 2611 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 25 01:29:03.111171 kubelet[2611]: W0325 01:29:03.111134 2611 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 25 01:29:03.112719 kubelet[2611]: W0325 01:29:03.112497 2611 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 25 01:29:03.130664 kubelet[2611]: I0325 01:29:03.129836 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4ae26885669d92b3428c37e4f6d2d17a-kubeconfig\") pod \"kube-scheduler-ci-4284.0.0-c-20cdc03f82\" (UID: \"4ae26885669d92b3428c37e4f6d2d17a\") " pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130664 kubelet[2611]: I0325 01:29:03.129916 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5dd43591be4d9acd85e69e7f9100fc57-k8s-certs\") pod \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" (UID: \"5dd43591be4d9acd85e69e7f9100fc57\") " pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130664 kubelet[2611]: I0325 01:29:03.129973 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5dd43591be4d9acd85e69e7f9100fc57-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" (UID: \"5dd43591be4d9acd85e69e7f9100fc57\") " pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130664 kubelet[2611]: I0325 01:29:03.130013 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-ca-certs\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130664 kubelet[2611]: I0325 01:29:03.130046 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-flexvolume-dir\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130957 kubelet[2611]: I0325 01:29:03.130076 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5dd43591be4d9acd85e69e7f9100fc57-ca-certs\") pod \"kube-apiserver-ci-4284.0.0-c-20cdc03f82\" (UID: \"5dd43591be4d9acd85e69e7f9100fc57\") " pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130957 kubelet[2611]: I0325 01:29:03.130107 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-k8s-certs\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130957 kubelet[2611]: I0325 01:29:03.130138 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-kubeconfig\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.130957 kubelet[2611]: I0325 01:29:03.130171 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aff1e0550efee4d3ea6a70f023943fac-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4284.0.0-c-20cdc03f82\" (UID: \"aff1e0550efee4d3ea6a70f023943fac\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" Mar 25 01:29:03.408281 kubelet[2611]: E0325 01:29:03.408149 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:03.412372 kubelet[2611]: E0325 01:29:03.412248 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:03.413581 kubelet[2611]: E0325 01:29:03.413547 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:03.490791 sudo[2625]: pam_unix(sudo:session): session closed for user root Mar 25 01:29:03.679995 kubelet[2611]: I0325 01:29:03.679797 2611 apiserver.go:52] "Watching apiserver" Mar 25 01:29:03.729042 kubelet[2611]: I0325 01:29:03.728822 2611 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Mar 25 01:29:03.852745 kubelet[2611]: E0325 01:29:03.852062 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:03.856071 kubelet[2611]: E0325 01:29:03.855673 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:03.857375 kubelet[2611]: E0325 01:29:03.856829 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:03.901288 kubelet[2611]: I0325 01:29:03.901180 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4284.0.0-c-20cdc03f82" podStartSLOduration=0.901142405 podStartE2EDuration="901.142405ms" podCreationTimestamp="2025-03-25 01:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-25 01:29:03.901123723 +0000 UTC m=+1.375701927" watchObservedRunningTime="2025-03-25 01:29:03.901142405 +0000 UTC m=+1.375720615" Mar 25 01:29:03.901501 kubelet[2611]: I0325 01:29:03.901357 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4284.0.0-c-20cdc03f82" podStartSLOduration=0.90134381 podStartE2EDuration="901.34381ms" podCreationTimestamp="2025-03-25 01:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-25 01:29:03.885748278 +0000 UTC m=+1.360326486" watchObservedRunningTime="2025-03-25 01:29:03.90134381 +0000 UTC m=+1.375922017" Mar 25 01:29:03.934668 kubelet[2611]: I0325 01:29:03.934420 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4284.0.0-c-20cdc03f82" podStartSLOduration=0.934402765 podStartE2EDuration="934.402765ms" podCreationTimestamp="2025-03-25 01:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-25 01:29:03.921585363 +0000 UTC m=+1.396163570" watchObservedRunningTime="2025-03-25 01:29:03.934402765 +0000 UTC m=+1.408980971" Mar 25 01:29:04.858172 kubelet[2611]: E0325 01:29:04.857566 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:04.858172 kubelet[2611]: E0325 01:29:04.857721 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:05.299116 sudo[1688]: pam_unix(sudo:session): session closed for user root Mar 25 01:29:05.307348 sshd[1687]: Connection closed by 139.178.89.65 port 45618 Mar 25 01:29:05.310432 sshd-session[1684]: pam_unix(sshd:session): session closed for user core Mar 25 01:29:05.316653 systemd[1]: sshd@6-143.198.155.16:22-139.178.89.65:45618.service: Deactivated successfully. Mar 25 01:29:05.321128 systemd[1]: session-7.scope: Deactivated successfully. Mar 25 01:29:05.322017 systemd[1]: session-7.scope: Consumed 5.735s CPU time, 217.2M memory peak. Mar 25 01:29:05.324827 systemd-logind[1463]: Session 7 logged out. Waiting for processes to exit. Mar 25 01:29:05.327041 systemd-logind[1463]: Removed session 7. Mar 25 01:29:05.861738 kubelet[2611]: E0325 01:29:05.861698 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:06.001888 kubelet[2611]: I0325 01:29:06.001844 2611 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 25 01:29:06.002769 containerd[1487]: time="2025-03-25T01:29:06.002712006Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 25 01:29:06.003563 kubelet[2611]: I0325 01:29:06.003542 2611 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 25 01:29:06.698736 systemd[1]: Created slice kubepods-besteffort-podf5287082_fd9f_473d_a452_192fe61168b2.slice - libcontainer container kubepods-besteffort-podf5287082_fd9f_473d_a452_192fe61168b2.slice. Mar 25 01:29:06.720333 systemd[1]: Created slice kubepods-burstable-podaf367a1f_cf63_4ea2_9d0d_322d41aeb2ca.slice - libcontainer container kubepods-burstable-podaf367a1f_cf63_4ea2_9d0d_322d41aeb2ca.slice. Mar 25 01:29:06.762263 kubelet[2611]: I0325 01:29:06.762192 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqv9m\" (UniqueName: \"kubernetes.io/projected/f5287082-fd9f-473d-a452-192fe61168b2-kube-api-access-jqv9m\") pod \"kube-proxy-zkg2h\" (UID: \"f5287082-fd9f-473d-a452-192fe61168b2\") " pod="kube-system/kube-proxy-zkg2h" Mar 25 01:29:06.762263 kubelet[2611]: I0325 01:29:06.762240 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-xtables-lock\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762263 kubelet[2611]: I0325 01:29:06.762259 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-etc-cni-netd\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762488 kubelet[2611]: I0325 01:29:06.762281 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-run\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762488 kubelet[2611]: I0325 01:29:06.762298 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-bpf-maps\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762488 kubelet[2611]: I0325 01:29:06.762313 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hostproc\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762488 kubelet[2611]: I0325 01:29:06.762329 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-lib-modules\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762488 kubelet[2611]: I0325 01:29:06.762345 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvh4b\" (UniqueName: \"kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-kube-api-access-mvh4b\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762488 kubelet[2611]: I0325 01:29:06.762365 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f5287082-fd9f-473d-a452-192fe61168b2-lib-modules\") pod \"kube-proxy-zkg2h\" (UID: \"f5287082-fd9f-473d-a452-192fe61168b2\") " pod="kube-system/kube-proxy-zkg2h" Mar 25 01:29:06.762649 kubelet[2611]: I0325 01:29:06.762380 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-clustermesh-secrets\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762649 kubelet[2611]: I0325 01:29:06.762395 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-kernel\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762649 kubelet[2611]: I0325 01:29:06.762409 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/f5287082-fd9f-473d-a452-192fe61168b2-kube-proxy\") pod \"kube-proxy-zkg2h\" (UID: \"f5287082-fd9f-473d-a452-192fe61168b2\") " pod="kube-system/kube-proxy-zkg2h" Mar 25 01:29:06.762649 kubelet[2611]: I0325 01:29:06.762448 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-config-path\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762649 kubelet[2611]: I0325 01:29:06.762471 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cni-path\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762775 kubelet[2611]: I0325 01:29:06.762486 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-net\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762775 kubelet[2611]: I0325 01:29:06.762501 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f5287082-fd9f-473d-a452-192fe61168b2-xtables-lock\") pod \"kube-proxy-zkg2h\" (UID: \"f5287082-fd9f-473d-a452-192fe61168b2\") " pod="kube-system/kube-proxy-zkg2h" Mar 25 01:29:06.762775 kubelet[2611]: I0325 01:29:06.762518 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-cgroup\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:06.762775 kubelet[2611]: I0325 01:29:06.762587 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hubble-tls\") pod \"cilium-8896b\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " pod="kube-system/cilium-8896b" Mar 25 01:29:07.006970 kubelet[2611]: E0325 01:29:07.006918 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:07.010108 containerd[1487]: time="2025-03-25T01:29:07.010052319Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zkg2h,Uid:f5287082-fd9f-473d-a452-192fe61168b2,Namespace:kube-system,Attempt:0,}" Mar 25 01:29:07.030897 kubelet[2611]: E0325 01:29:07.029698 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:07.031123 containerd[1487]: time="2025-03-25T01:29:07.030240579Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-8896b,Uid:af367a1f-cf63-4ea2-9d0d-322d41aeb2ca,Namespace:kube-system,Attempt:0,}" Mar 25 01:29:07.042118 containerd[1487]: time="2025-03-25T01:29:07.042050068Z" level=info msg="connecting to shim 9fa8dd008b5bb0fc7f259d1ab1eff5346835039f90b53c7d9dd71d5e4b40132e" address="unix:///run/containerd/s/8a48fa5b0ccf3b65461acdb42bf8156a4e3e9581d2f9c1247b3b2982848c14c8" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:29:07.102979 systemd[1]: Started cri-containerd-9fa8dd008b5bb0fc7f259d1ab1eff5346835039f90b53c7d9dd71d5e4b40132e.scope - libcontainer container 9fa8dd008b5bb0fc7f259d1ab1eff5346835039f90b53c7d9dd71d5e4b40132e. Mar 25 01:29:07.108819 containerd[1487]: time="2025-03-25T01:29:07.107823819Z" level=info msg="connecting to shim d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638" address="unix:///run/containerd/s/29e8fde34aba6f6bc6d6187a1c6aec479e91c6fc315517a095752c11921797ff" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:29:07.145977 systemd[1]: Started cri-containerd-d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638.scope - libcontainer container d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638. Mar 25 01:29:07.182887 systemd[1]: Created slice kubepods-besteffort-pod9ca7be5e_1545_4171_a3c5_a6faa9445e65.slice - libcontainer container kubepods-besteffort-pod9ca7be5e_1545_4171_a3c5_a6faa9445e65.slice. Mar 25 01:29:07.251504 containerd[1487]: time="2025-03-25T01:29:07.251444257Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-8896b,Uid:af367a1f-cf63-4ea2-9d0d-322d41aeb2ca,Namespace:kube-system,Attempt:0,} returns sandbox id \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\"" Mar 25 01:29:07.259591 kubelet[2611]: E0325 01:29:07.258487 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:07.266611 containerd[1487]: time="2025-03-25T01:29:07.266286044Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 25 01:29:07.268175 kubelet[2611]: I0325 01:29:07.267887 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9ca7be5e-1545-4171-a3c5-a6faa9445e65-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-nd256\" (UID: \"9ca7be5e-1545-4171-a3c5-a6faa9445e65\") " pod="kube-system/cilium-operator-6c4d7847fc-nd256" Mar 25 01:29:07.268175 kubelet[2611]: I0325 01:29:07.268030 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/9ca7be5e-1545-4171-a3c5-a6faa9445e65-kube-api-access-85z98\") pod \"cilium-operator-6c4d7847fc-nd256\" (UID: \"9ca7be5e-1545-4171-a3c5-a6faa9445e65\") " pod="kube-system/cilium-operator-6c4d7847fc-nd256" Mar 25 01:29:07.272295 systemd-resolved[1332]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. Mar 25 01:29:07.277077 containerd[1487]: time="2025-03-25T01:29:07.275436214Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zkg2h,Uid:f5287082-fd9f-473d-a452-192fe61168b2,Namespace:kube-system,Attempt:0,} returns sandbox id \"9fa8dd008b5bb0fc7f259d1ab1eff5346835039f90b53c7d9dd71d5e4b40132e\"" Mar 25 01:29:07.278709 kubelet[2611]: E0325 01:29:07.278572 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:07.283748 containerd[1487]: time="2025-03-25T01:29:07.283065989Z" level=info msg="CreateContainer within sandbox \"9fa8dd008b5bb0fc7f259d1ab1eff5346835039f90b53c7d9dd71d5e4b40132e\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 25 01:29:07.298649 containerd[1487]: time="2025-03-25T01:29:07.298027307Z" level=info msg="Container ab8f0f6512f96ed91e6840f01cec1a70fcc8a5c3a9cd5502e1751af0b0990436: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:07.307222 containerd[1487]: time="2025-03-25T01:29:07.307153482Z" level=info msg="CreateContainer within sandbox \"9fa8dd008b5bb0fc7f259d1ab1eff5346835039f90b53c7d9dd71d5e4b40132e\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"ab8f0f6512f96ed91e6840f01cec1a70fcc8a5c3a9cd5502e1751af0b0990436\"" Mar 25 01:29:07.310686 containerd[1487]: time="2025-03-25T01:29:07.309182786Z" level=info msg="StartContainer for \"ab8f0f6512f96ed91e6840f01cec1a70fcc8a5c3a9cd5502e1751af0b0990436\"" Mar 25 01:29:07.313414 containerd[1487]: time="2025-03-25T01:29:07.313341413Z" level=info msg="connecting to shim ab8f0f6512f96ed91e6840f01cec1a70fcc8a5c3a9cd5502e1751af0b0990436" address="unix:///run/containerd/s/8a48fa5b0ccf3b65461acdb42bf8156a4e3e9581d2f9c1247b3b2982848c14c8" protocol=ttrpc version=3 Mar 25 01:29:07.343987 systemd[1]: Started cri-containerd-ab8f0f6512f96ed91e6840f01cec1a70fcc8a5c3a9cd5502e1751af0b0990436.scope - libcontainer container ab8f0f6512f96ed91e6840f01cec1a70fcc8a5c3a9cd5502e1751af0b0990436. Mar 25 01:29:07.481912 containerd[1487]: time="2025-03-25T01:29:07.481845978Z" level=info msg="StartContainer for \"ab8f0f6512f96ed91e6840f01cec1a70fcc8a5c3a9cd5502e1751af0b0990436\" returns successfully" Mar 25 01:29:07.487676 kubelet[2611]: E0325 01:29:07.487274 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:07.491408 containerd[1487]: time="2025-03-25T01:29:07.490499492Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-nd256,Uid:9ca7be5e-1545-4171-a3c5-a6faa9445e65,Namespace:kube-system,Attempt:0,}" Mar 25 01:29:07.519848 containerd[1487]: time="2025-03-25T01:29:07.519374495Z" level=info msg="connecting to shim 851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca" address="unix:///run/containerd/s/28146dff866a384331989927a86f5812d5f4f773d84e57c690b6ed6b1c8d573b" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:29:07.570015 systemd[1]: Started cri-containerd-851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca.scope - libcontainer container 851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca. Mar 25 01:29:07.671719 containerd[1487]: time="2025-03-25T01:29:07.671493485Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-nd256,Uid:9ca7be5e-1545-4171-a3c5-a6faa9445e65,Namespace:kube-system,Attempt:0,} returns sandbox id \"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\"" Mar 25 01:29:07.672801 kubelet[2611]: E0325 01:29:07.672769 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:07.901860 kubelet[2611]: E0325 01:29:07.899141 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:07.921326 kubelet[2611]: I0325 01:29:07.921244 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-zkg2h" podStartSLOduration=1.921217781 podStartE2EDuration="1.921217781s" podCreationTimestamp="2025-03-25 01:29:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-25 01:29:07.920778756 +0000 UTC m=+5.395356963" watchObservedRunningTime="2025-03-25 01:29:07.921217781 +0000 UTC m=+5.395795990" Mar 25 01:29:11.654539 kubelet[2611]: E0325 01:29:11.654454 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:11.829974 kubelet[2611]: E0325 01:29:11.829398 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:11.906376 kubelet[2611]: E0325 01:29:11.905965 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:11.907588 kubelet[2611]: E0325 01:29:11.907283 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:12.907541 kubelet[2611]: E0325 01:29:12.907048 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:13.079678 kubelet[2611]: E0325 01:29:13.077979 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:13.917793 kubelet[2611]: E0325 01:29:13.917557 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:14.604824 update_engine[1464]: I20250325 01:29:14.604624 1464 update_attempter.cc:509] Updating boot flags... Mar 25 01:29:14.672887 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2987) Mar 25 01:29:14.893677 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2986) Mar 25 01:29:15.083876 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2986) Mar 25 01:29:19.039349 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount698568853.mount: Deactivated successfully. Mar 25 01:29:21.608926 containerd[1487]: time="2025-03-25T01:29:21.608847886Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:29:21.610471 containerd[1487]: time="2025-03-25T01:29:21.610366240Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Mar 25 01:29:21.611368 containerd[1487]: time="2025-03-25T01:29:21.611305474Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:29:21.612710 containerd[1487]: time="2025-03-25T01:29:21.612666907Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 14.346330122s" Mar 25 01:29:21.612710 containerd[1487]: time="2025-03-25T01:29:21.612707642Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Mar 25 01:29:21.614976 containerd[1487]: time="2025-03-25T01:29:21.614942322Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 25 01:29:21.619501 containerd[1487]: time="2025-03-25T01:29:21.618683567Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 25 01:29:21.647664 containerd[1487]: time="2025-03-25T01:29:21.647091076Z" level=info msg="Container 4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:21.664546 containerd[1487]: time="2025-03-25T01:29:21.664498707Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\"" Mar 25 01:29:21.666260 containerd[1487]: time="2025-03-25T01:29:21.666217071Z" level=info msg="StartContainer for \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\"" Mar 25 01:29:21.669302 containerd[1487]: time="2025-03-25T01:29:21.668976828Z" level=info msg="connecting to shim 4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396" address="unix:///run/containerd/s/29e8fde34aba6f6bc6d6187a1c6aec479e91c6fc315517a095752c11921797ff" protocol=ttrpc version=3 Mar 25 01:29:21.732756 systemd[1]: Started cri-containerd-4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396.scope - libcontainer container 4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396. Mar 25 01:29:21.821418 containerd[1487]: time="2025-03-25T01:29:21.821342971Z" level=info msg="StartContainer for \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" returns successfully" Mar 25 01:29:21.832422 systemd[1]: cri-containerd-4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396.scope: Deactivated successfully. Mar 25 01:29:21.940848 containerd[1487]: time="2025-03-25T01:29:21.939815809Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" id:\"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" pid:3036 exited_at:{seconds:1742866161 nanos:861717841}" Mar 25 01:29:21.941194 containerd[1487]: time="2025-03-25T01:29:21.941072880Z" level=info msg="received exit event container_id:\"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" id:\"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" pid:3036 exited_at:{seconds:1742866161 nanos:861717841}" Mar 25 01:29:21.992251 kubelet[2611]: E0325 01:29:21.992153 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:22.036477 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396-rootfs.mount: Deactivated successfully. Mar 25 01:29:22.996057 kubelet[2611]: E0325 01:29:22.995959 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:23.004550 containerd[1487]: time="2025-03-25T01:29:23.003617711Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 25 01:29:23.044595 containerd[1487]: time="2025-03-25T01:29:23.042797642Z" level=info msg="Container 6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:23.066151 containerd[1487]: time="2025-03-25T01:29:23.065992938Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\"" Mar 25 01:29:23.070057 containerd[1487]: time="2025-03-25T01:29:23.069729827Z" level=info msg="StartContainer for \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\"" Mar 25 01:29:23.073270 containerd[1487]: time="2025-03-25T01:29:23.073214530Z" level=info msg="connecting to shim 6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab" address="unix:///run/containerd/s/29e8fde34aba6f6bc6d6187a1c6aec479e91c6fc315517a095752c11921797ff" protocol=ttrpc version=3 Mar 25 01:29:23.108022 systemd[1]: Started cri-containerd-6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab.scope - libcontainer container 6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab. Mar 25 01:29:23.154377 containerd[1487]: time="2025-03-25T01:29:23.154337637Z" level=info msg="StartContainer for \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" returns successfully" Mar 25 01:29:23.179871 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 25 01:29:23.181082 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 25 01:29:23.181884 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Mar 25 01:29:23.185035 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 25 01:29:23.190025 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 25 01:29:23.191507 systemd[1]: cri-containerd-6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab.scope: Deactivated successfully. Mar 25 01:29:23.191912 systemd[1]: cri-containerd-6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab.scope: Consumed 32ms CPU time, 7.7M memory peak, 16K read from disk, 2.2M written to disk. Mar 25 01:29:23.201745 containerd[1487]: time="2025-03-25T01:29:23.198869516Z" level=info msg="received exit event container_id:\"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" id:\"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" pid:3083 exited_at:{seconds:1742866163 nanos:197327536}" Mar 25 01:29:23.201997 containerd[1487]: time="2025-03-25T01:29:23.201874495Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" id:\"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" pid:3083 exited_at:{seconds:1742866163 nanos:197327536}" Mar 25 01:29:23.236916 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 25 01:29:24.006764 kubelet[2611]: E0325 01:29:24.006720 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:24.017590 containerd[1487]: time="2025-03-25T01:29:24.017012426Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 25 01:29:24.037664 containerd[1487]: time="2025-03-25T01:29:24.031987056Z" level=info msg="Container 9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:24.041287 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab-rootfs.mount: Deactivated successfully. Mar 25 01:29:24.052946 containerd[1487]: time="2025-03-25T01:29:24.052887527Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\"" Mar 25 01:29:24.053716 containerd[1487]: time="2025-03-25T01:29:24.053677752Z" level=info msg="StartContainer for \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\"" Mar 25 01:29:24.056444 containerd[1487]: time="2025-03-25T01:29:24.056276195Z" level=info msg="connecting to shim 9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d" address="unix:///run/containerd/s/29e8fde34aba6f6bc6d6187a1c6aec479e91c6fc315517a095752c11921797ff" protocol=ttrpc version=3 Mar 25 01:29:24.095355 systemd[1]: Started cri-containerd-9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d.scope - libcontainer container 9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d. Mar 25 01:29:24.169837 systemd[1]: cri-containerd-9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d.scope: Deactivated successfully. Mar 25 01:29:24.174824 containerd[1487]: time="2025-03-25T01:29:24.174778816Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" id:\"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" pid:3131 exited_at:{seconds:1742866164 nanos:171313538}" Mar 25 01:29:24.175557 containerd[1487]: time="2025-03-25T01:29:24.175479999Z" level=info msg="received exit event container_id:\"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" id:\"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" pid:3131 exited_at:{seconds:1742866164 nanos:171313538}" Mar 25 01:29:24.188687 containerd[1487]: time="2025-03-25T01:29:24.188306313Z" level=info msg="StartContainer for \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" returns successfully" Mar 25 01:29:24.219342 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d-rootfs.mount: Deactivated successfully. Mar 25 01:29:25.018755 kubelet[2611]: E0325 01:29:25.018420 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:25.027331 containerd[1487]: time="2025-03-25T01:29:25.027006246Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 25 01:29:25.058995 containerd[1487]: time="2025-03-25T01:29:25.058948312Z" level=info msg="Container 878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:25.071256 containerd[1487]: time="2025-03-25T01:29:25.071196570Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\"" Mar 25 01:29:25.074414 containerd[1487]: time="2025-03-25T01:29:25.074358326Z" level=info msg="StartContainer for \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\"" Mar 25 01:29:25.075418 containerd[1487]: time="2025-03-25T01:29:25.075357102Z" level=info msg="connecting to shim 878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a" address="unix:///run/containerd/s/29e8fde34aba6f6bc6d6187a1c6aec479e91c6fc315517a095752c11921797ff" protocol=ttrpc version=3 Mar 25 01:29:25.121952 systemd[1]: Started cri-containerd-878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a.scope - libcontainer container 878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a. Mar 25 01:29:25.192403 systemd[1]: cri-containerd-878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a.scope: Deactivated successfully. Mar 25 01:29:25.194573 containerd[1487]: time="2025-03-25T01:29:25.194229893Z" level=info msg="TaskExit event in podsandbox handler container_id:\"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" id:\"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" pid:3179 exited_at:{seconds:1742866165 nanos:192807274}" Mar 25 01:29:25.197837 containerd[1487]: time="2025-03-25T01:29:25.197688910Z" level=info msg="received exit event container_id:\"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" id:\"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" pid:3179 exited_at:{seconds:1742866165 nanos:192807274}" Mar 25 01:29:25.200880 containerd[1487]: time="2025-03-25T01:29:25.200822914Z" level=info msg="StartContainer for \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" returns successfully" Mar 25 01:29:25.259843 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a-rootfs.mount: Deactivated successfully. Mar 25 01:29:25.485046 containerd[1487]: time="2025-03-25T01:29:25.484012947Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:29:25.486478 containerd[1487]: time="2025-03-25T01:29:25.486385784Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Mar 25 01:29:25.487124 containerd[1487]: time="2025-03-25T01:29:25.487097472Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 25 01:29:25.489711 containerd[1487]: time="2025-03-25T01:29:25.489662751Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 3.874656022s" Mar 25 01:29:25.490043 containerd[1487]: time="2025-03-25T01:29:25.489894836Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Mar 25 01:29:25.494742 containerd[1487]: time="2025-03-25T01:29:25.493092178Z" level=info msg="CreateContainer within sandbox \"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 25 01:29:25.503691 containerd[1487]: time="2025-03-25T01:29:25.503626053Z" level=info msg="Container 3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:25.512955 containerd[1487]: time="2025-03-25T01:29:25.512905448Z" level=info msg="CreateContainer within sandbox \"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\"" Mar 25 01:29:25.514663 containerd[1487]: time="2025-03-25T01:29:25.514168500Z" level=info msg="StartContainer for \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\"" Mar 25 01:29:25.517821 containerd[1487]: time="2025-03-25T01:29:25.517694170Z" level=info msg="connecting to shim 3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124" address="unix:///run/containerd/s/28146dff866a384331989927a86f5812d5f4f773d84e57c690b6ed6b1c8d573b" protocol=ttrpc version=3 Mar 25 01:29:25.541919 systemd[1]: Started cri-containerd-3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124.scope - libcontainer container 3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124. Mar 25 01:29:25.583444 containerd[1487]: time="2025-03-25T01:29:25.583054724Z" level=info msg="StartContainer for \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" returns successfully" Mar 25 01:29:26.022113 kubelet[2611]: E0325 01:29:26.021249 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:26.039488 kubelet[2611]: E0325 01:29:26.039420 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:26.045854 containerd[1487]: time="2025-03-25T01:29:26.043619747Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 25 01:29:26.092919 containerd[1487]: time="2025-03-25T01:29:26.092863342Z" level=info msg="Container 0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:26.103148 containerd[1487]: time="2025-03-25T01:29:26.103011415Z" level=info msg="CreateContainer within sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\"" Mar 25 01:29:26.104358 containerd[1487]: time="2025-03-25T01:29:26.104279585Z" level=info msg="StartContainer for \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\"" Mar 25 01:29:26.105468 containerd[1487]: time="2025-03-25T01:29:26.105433478Z" level=info msg="connecting to shim 0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc" address="unix:///run/containerd/s/29e8fde34aba6f6bc6d6187a1c6aec479e91c6fc315517a095752c11921797ff" protocol=ttrpc version=3 Mar 25 01:29:26.151888 systemd[1]: Started cri-containerd-0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc.scope - libcontainer container 0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc. Mar 25 01:29:26.258168 containerd[1487]: time="2025-03-25T01:29:26.258115884Z" level=info msg="StartContainer for \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" returns successfully" Mar 25 01:29:26.326191 kubelet[2611]: I0325 01:29:26.323147 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-nd256" podStartSLOduration=1.505713063 podStartE2EDuration="19.32310948s" podCreationTimestamp="2025-03-25 01:29:07 +0000 UTC" firstStartedPulling="2025-03-25 01:29:07.673625188 +0000 UTC m=+5.148203374" lastFinishedPulling="2025-03-25 01:29:25.491021594 +0000 UTC m=+22.965599791" observedRunningTime="2025-03-25 01:29:26.097126062 +0000 UTC m=+23.571704269" watchObservedRunningTime="2025-03-25 01:29:26.32310948 +0000 UTC m=+23.797687689" Mar 25 01:29:26.647107 containerd[1487]: time="2025-03-25T01:29:26.646969992Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" id:\"39b8ba833c60c88e4bab35eea4f848f51c5a58a8bb270a23615cde2bfdd35dc5\" pid:3280 exited_at:{seconds:1742866166 nanos:644164618}" Mar 25 01:29:26.693276 kubelet[2611]: I0325 01:29:26.693242 2611 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Mar 25 01:29:26.851763 systemd[1]: Created slice kubepods-burstable-pod13e2dcd3_dabd_4088_81c2_151afa7dc3a1.slice - libcontainer container kubepods-burstable-pod13e2dcd3_dabd_4088_81c2_151afa7dc3a1.slice. Mar 25 01:29:26.863252 kubelet[2611]: W0325 01:29:26.861389 2611 reflector.go:569] object-"kube-system"/"coredns": failed to list *v1.ConfigMap: configmaps "coredns" is forbidden: User "system:node:ci-4284.0.0-c-20cdc03f82" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-4284.0.0-c-20cdc03f82' and this object Mar 25 01:29:26.863654 kubelet[2611]: I0325 01:29:26.862825 2611 status_manager.go:890] "Failed to get status for pod" podUID="13e2dcd3-dabd-4088-81c2-151afa7dc3a1" pod="kube-system/coredns-668d6bf9bc-pcv79" err="pods \"coredns-668d6bf9bc-pcv79\" is forbidden: User \"system:node:ci-4284.0.0-c-20cdc03f82\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4284.0.0-c-20cdc03f82' and this object" Mar 25 01:29:26.867044 kubelet[2611]: E0325 01:29:26.866363 2611 reflector.go:166] "Unhandled Error" err="object-\"kube-system\"/\"coredns\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"coredns\" is forbidden: User \"system:node:ci-4284.0.0-c-20cdc03f82\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4284.0.0-c-20cdc03f82' and this object" logger="UnhandledError" Mar 25 01:29:26.871182 systemd[1]: Created slice kubepods-burstable-podc28c1cb2_20c5_43cf_a014_7e9eff387ef1.slice - libcontainer container kubepods-burstable-podc28c1cb2_20c5_43cf_a014_7e9eff387ef1.slice. Mar 25 01:29:27.023557 kubelet[2611]: I0325 01:29:27.023248 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13e2dcd3-dabd-4088-81c2-151afa7dc3a1-config-volume\") pod \"coredns-668d6bf9bc-pcv79\" (UID: \"13e2dcd3-dabd-4088-81c2-151afa7dc3a1\") " pod="kube-system/coredns-668d6bf9bc-pcv79" Mar 25 01:29:27.023557 kubelet[2611]: I0325 01:29:27.023304 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5wn6\" (UniqueName: \"kubernetes.io/projected/c28c1cb2-20c5-43cf-a014-7e9eff387ef1-kube-api-access-b5wn6\") pod \"coredns-668d6bf9bc-7jh8f\" (UID: \"c28c1cb2-20c5-43cf-a014-7e9eff387ef1\") " pod="kube-system/coredns-668d6bf9bc-7jh8f" Mar 25 01:29:27.023557 kubelet[2611]: I0325 01:29:27.023328 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5lht\" (UniqueName: \"kubernetes.io/projected/13e2dcd3-dabd-4088-81c2-151afa7dc3a1-kube-api-access-x5lht\") pod \"coredns-668d6bf9bc-pcv79\" (UID: \"13e2dcd3-dabd-4088-81c2-151afa7dc3a1\") " pod="kube-system/coredns-668d6bf9bc-pcv79" Mar 25 01:29:27.023557 kubelet[2611]: I0325 01:29:27.023344 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c28c1cb2-20c5-43cf-a014-7e9eff387ef1-config-volume\") pod \"coredns-668d6bf9bc-7jh8f\" (UID: \"c28c1cb2-20c5-43cf-a014-7e9eff387ef1\") " pod="kube-system/coredns-668d6bf9bc-7jh8f" Mar 25 01:29:27.050272 kubelet[2611]: E0325 01:29:27.050226 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:27.051207 kubelet[2611]: E0325 01:29:27.051176 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:28.053166 kubelet[2611]: E0325 01:29:28.052657 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:28.058143 kubelet[2611]: E0325 01:29:28.058098 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:28.060312 containerd[1487]: time="2025-03-25T01:29:28.060015468Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-pcv79,Uid:13e2dcd3-dabd-4088-81c2-151afa7dc3a1,Namespace:kube-system,Attempt:0,}" Mar 25 01:29:28.079305 kubelet[2611]: E0325 01:29:28.077300 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:28.085248 containerd[1487]: time="2025-03-25T01:29:28.080812883Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7jh8f,Uid:c28c1cb2-20c5-43cf-a014-7e9eff387ef1,Namespace:kube-system,Attempt:0,}" Mar 25 01:29:29.022950 systemd-networkd[1366]: cilium_host: Link UP Mar 25 01:29:29.027728 systemd-networkd[1366]: cilium_net: Link UP Mar 25 01:29:29.028081 systemd-networkd[1366]: cilium_net: Gained carrier Mar 25 01:29:29.028290 systemd-networkd[1366]: cilium_host: Gained carrier Mar 25 01:29:29.062286 kubelet[2611]: E0325 01:29:29.060706 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:29.199961 systemd-networkd[1366]: cilium_vxlan: Link UP Mar 25 01:29:29.200373 systemd-networkd[1366]: cilium_vxlan: Gained carrier Mar 25 01:29:29.669736 kernel: NET: Registered PF_ALG protocol family Mar 25 01:29:29.922805 systemd-networkd[1366]: cilium_host: Gained IPv6LL Mar 25 01:29:29.936362 systemd[1]: Started sshd@7-143.198.155.16:22-218.92.0.156:61211.service - OpenSSH per-connection server daemon (218.92.0.156:61211). Mar 25 01:29:30.050791 systemd-networkd[1366]: cilium_net: Gained IPv6LL Mar 25 01:29:30.626941 systemd-networkd[1366]: cilium_vxlan: Gained IPv6LL Mar 25 01:29:30.680578 systemd-networkd[1366]: lxc_health: Link UP Mar 25 01:29:30.690329 systemd-networkd[1366]: lxc_health: Gained carrier Mar 25 01:29:30.996109 sshd-session[3731]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.156 user=root Mar 25 01:29:31.033286 kubelet[2611]: E0325 01:29:31.032397 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:31.061185 kubelet[2611]: I0325 01:29:31.058538 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-8896b" podStartSLOduration=10.706301424 podStartE2EDuration="25.058516497s" podCreationTimestamp="2025-03-25 01:29:06 +0000 UTC" firstStartedPulling="2025-03-25 01:29:07.262528305 +0000 UTC m=+4.737106505" lastFinishedPulling="2025-03-25 01:29:21.61474338 +0000 UTC m=+19.089321578" observedRunningTime="2025-03-25 01:29:27.114296202 +0000 UTC m=+24.588874432" watchObservedRunningTime="2025-03-25 01:29:31.058516497 +0000 UTC m=+28.533094700" Mar 25 01:29:31.066075 kubelet[2611]: E0325 01:29:31.066028 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:31.158668 kernel: eth0: renamed from tmpa64d8 Mar 25 01:29:31.155857 systemd-networkd[1366]: lxc2d1a05c176fc: Link UP Mar 25 01:29:31.169824 systemd-networkd[1366]: lxc2d1a05c176fc: Gained carrier Mar 25 01:29:31.181376 systemd-networkd[1366]: lxcecc54c109156: Link UP Mar 25 01:29:31.189689 kernel: eth0: renamed from tmp9d207 Mar 25 01:29:31.196555 systemd-networkd[1366]: lxcecc54c109156: Gained carrier Mar 25 01:29:32.067806 kubelet[2611]: E0325 01:29:32.067767 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:32.482002 systemd-networkd[1366]: lxc_health: Gained IPv6LL Mar 25 01:29:32.801990 systemd-networkd[1366]: lxc2d1a05c176fc: Gained IPv6LL Mar 25 01:29:32.866030 systemd-networkd[1366]: lxcecc54c109156: Gained IPv6LL Mar 25 01:29:32.876003 sshd[3500]: PAM: Permission denied for root from 218.92.0.156 Mar 25 01:29:33.152672 sshd-session[3758]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.156 user=root Mar 25 01:29:34.446366 sshd[3500]: PAM: Permission denied for root from 218.92.0.156 Mar 25 01:29:34.722366 sshd-session[3762]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.156 user=root Mar 25 01:29:36.151293 sshd[3500]: PAM: Permission denied for root from 218.92.0.156 Mar 25 01:29:36.289255 sshd[3500]: Received disconnect from 218.92.0.156 port 61211:11: [preauth] Mar 25 01:29:36.289255 sshd[3500]: Disconnected from authenticating user root 218.92.0.156 port 61211 [preauth] Mar 25 01:29:36.294011 systemd[1]: sshd@7-143.198.155.16:22-218.92.0.156:61211.service: Deactivated successfully. Mar 25 01:29:37.838577 containerd[1487]: time="2025-03-25T01:29:37.838510431Z" level=info msg="connecting to shim 9d2073cd9a7f94b705b8887459e129ed5a681dd5f893a5b5a2bbcbc67ddf92bd" address="unix:///run/containerd/s/92099161a9a0b88592afebcd7062ec81e725517198b885e9a2dbef6742407e3f" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:29:37.855611 containerd[1487]: time="2025-03-25T01:29:37.855544791Z" level=info msg="connecting to shim a64d88cc0132aea9ef6a5a96032891a97d85c510dc2f1eff22bfdbd26b158ba0" address="unix:///run/containerd/s/bf01dd3bc31745df4f30ac105fa405c2d3875a6ee67b28cdd681a3e3a1d6f1af" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:29:37.990403 systemd[1]: Started cri-containerd-a64d88cc0132aea9ef6a5a96032891a97d85c510dc2f1eff22bfdbd26b158ba0.scope - libcontainer container a64d88cc0132aea9ef6a5a96032891a97d85c510dc2f1eff22bfdbd26b158ba0. Mar 25 01:29:38.003482 systemd[1]: Started cri-containerd-9d2073cd9a7f94b705b8887459e129ed5a681dd5f893a5b5a2bbcbc67ddf92bd.scope - libcontainer container 9d2073cd9a7f94b705b8887459e129ed5a681dd5f893a5b5a2bbcbc67ddf92bd. Mar 25 01:29:38.167169 containerd[1487]: time="2025-03-25T01:29:38.166510182Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-pcv79,Uid:13e2dcd3-dabd-4088-81c2-151afa7dc3a1,Namespace:kube-system,Attempt:0,} returns sandbox id \"a64d88cc0132aea9ef6a5a96032891a97d85c510dc2f1eff22bfdbd26b158ba0\"" Mar 25 01:29:38.168642 kubelet[2611]: E0325 01:29:38.168591 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:38.172981 containerd[1487]: time="2025-03-25T01:29:38.172925914Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7jh8f,Uid:c28c1cb2-20c5-43cf-a014-7e9eff387ef1,Namespace:kube-system,Attempt:0,} returns sandbox id \"9d2073cd9a7f94b705b8887459e129ed5a681dd5f893a5b5a2bbcbc67ddf92bd\"" Mar 25 01:29:38.176821 kubelet[2611]: E0325 01:29:38.176177 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:38.177871 containerd[1487]: time="2025-03-25T01:29:38.177825428Z" level=info msg="CreateContainer within sandbox \"a64d88cc0132aea9ef6a5a96032891a97d85c510dc2f1eff22bfdbd26b158ba0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 25 01:29:38.179965 containerd[1487]: time="2025-03-25T01:29:38.179776825Z" level=info msg="CreateContainer within sandbox \"9d2073cd9a7f94b705b8887459e129ed5a681dd5f893a5b5a2bbcbc67ddf92bd\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 25 01:29:38.208302 containerd[1487]: time="2025-03-25T01:29:38.207854405Z" level=info msg="Container c1f069538e3883db97016567331a1f272a28569342dc741d801020e8e4a9af02: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:38.211094 containerd[1487]: time="2025-03-25T01:29:38.210033665Z" level=info msg="Container 063aba047338fe76b4ed6e2f568071ac78760a08c576ba4544b20f620242172c: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:29:38.225586 containerd[1487]: time="2025-03-25T01:29:38.225481333Z" level=info msg="CreateContainer within sandbox \"a64d88cc0132aea9ef6a5a96032891a97d85c510dc2f1eff22bfdbd26b158ba0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c1f069538e3883db97016567331a1f272a28569342dc741d801020e8e4a9af02\"" Mar 25 01:29:38.231940 containerd[1487]: time="2025-03-25T01:29:38.229617962Z" level=info msg="CreateContainer within sandbox \"9d2073cd9a7f94b705b8887459e129ed5a681dd5f893a5b5a2bbcbc67ddf92bd\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"063aba047338fe76b4ed6e2f568071ac78760a08c576ba4544b20f620242172c\"" Mar 25 01:29:38.231940 containerd[1487]: time="2025-03-25T01:29:38.229761451Z" level=info msg="StartContainer for \"c1f069538e3883db97016567331a1f272a28569342dc741d801020e8e4a9af02\"" Mar 25 01:29:38.231940 containerd[1487]: time="2025-03-25T01:29:38.230996519Z" level=info msg="connecting to shim c1f069538e3883db97016567331a1f272a28569342dc741d801020e8e4a9af02" address="unix:///run/containerd/s/bf01dd3bc31745df4f30ac105fa405c2d3875a6ee67b28cdd681a3e3a1d6f1af" protocol=ttrpc version=3 Mar 25 01:29:38.232552 containerd[1487]: time="2025-03-25T01:29:38.232007058Z" level=info msg="StartContainer for \"063aba047338fe76b4ed6e2f568071ac78760a08c576ba4544b20f620242172c\"" Mar 25 01:29:38.234708 containerd[1487]: time="2025-03-25T01:29:38.233894367Z" level=info msg="connecting to shim 063aba047338fe76b4ed6e2f568071ac78760a08c576ba4544b20f620242172c" address="unix:///run/containerd/s/92099161a9a0b88592afebcd7062ec81e725517198b885e9a2dbef6742407e3f" protocol=ttrpc version=3 Mar 25 01:29:38.280395 systemd[1]: Started cri-containerd-063aba047338fe76b4ed6e2f568071ac78760a08c576ba4544b20f620242172c.scope - libcontainer container 063aba047338fe76b4ed6e2f568071ac78760a08c576ba4544b20f620242172c. Mar 25 01:29:38.282571 systemd[1]: Started cri-containerd-c1f069538e3883db97016567331a1f272a28569342dc741d801020e8e4a9af02.scope - libcontainer container c1f069538e3883db97016567331a1f272a28569342dc741d801020e8e4a9af02. Mar 25 01:29:38.347156 containerd[1487]: time="2025-03-25T01:29:38.346796243Z" level=info msg="StartContainer for \"063aba047338fe76b4ed6e2f568071ac78760a08c576ba4544b20f620242172c\" returns successfully" Mar 25 01:29:38.354222 containerd[1487]: time="2025-03-25T01:29:38.354011302Z" level=info msg="StartContainer for \"c1f069538e3883db97016567331a1f272a28569342dc741d801020e8e4a9af02\" returns successfully" Mar 25 01:29:38.748987 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1044300310.mount: Deactivated successfully. Mar 25 01:29:39.098413 kubelet[2611]: E0325 01:29:39.097875 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:39.103492 kubelet[2611]: E0325 01:29:39.102983 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:39.117838 kubelet[2611]: I0325 01:29:39.117589 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-7jh8f" podStartSLOduration=32.117564141 podStartE2EDuration="32.117564141s" podCreationTimestamp="2025-03-25 01:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-25 01:29:39.115666535 +0000 UTC m=+36.590244742" watchObservedRunningTime="2025-03-25 01:29:39.117564141 +0000 UTC m=+36.592142348" Mar 25 01:29:39.140057 kubelet[2611]: I0325 01:29:39.139929 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-pcv79" podStartSLOduration=32.13990409 podStartE2EDuration="32.13990409s" podCreationTimestamp="2025-03-25 01:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-25 01:29:39.138981946 +0000 UTC m=+36.613560155" watchObservedRunningTime="2025-03-25 01:29:39.13990409 +0000 UTC m=+36.614482299" Mar 25 01:29:40.107560 kubelet[2611]: E0325 01:29:40.107365 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:40.107560 kubelet[2611]: E0325 01:29:40.107439 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:41.110081 kubelet[2611]: E0325 01:29:41.109839 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:29:41.110081 kubelet[2611]: E0325 01:29:41.109944 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:02.455916 systemd[1]: Started sshd@8-143.198.155.16:22-139.178.89.65:50386.service - OpenSSH per-connection server daemon (139.178.89.65:50386). Mar 25 01:30:02.745475 sshd[3945]: Accepted publickey for core from 139.178.89.65 port 50386 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:02.762406 sshd-session[3945]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:02.857615 systemd-logind[1463]: New session 8 of user core. Mar 25 01:30:02.860468 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 25 01:30:04.029930 sshd[3949]: Connection closed by 139.178.89.65 port 50386 Mar 25 01:30:04.031236 sshd-session[3945]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:04.039465 systemd[1]: sshd@8-143.198.155.16:22-139.178.89.65:50386.service: Deactivated successfully. Mar 25 01:30:04.051588 systemd[1]: session-8.scope: Deactivated successfully. Mar 25 01:30:04.056488 systemd-logind[1463]: Session 8 logged out. Waiting for processes to exit. Mar 25 01:30:04.060501 systemd-logind[1463]: Removed session 8. Mar 25 01:30:09.054840 systemd[1]: Started sshd@9-143.198.155.16:22-139.178.89.65:55848.service - OpenSSH per-connection server daemon (139.178.89.65:55848). Mar 25 01:30:09.129326 sshd[3965]: Accepted publickey for core from 139.178.89.65 port 55848 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:09.130704 sshd-session[3965]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:09.138897 systemd-logind[1463]: New session 9 of user core. Mar 25 01:30:09.150201 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 25 01:30:09.359333 sshd[3967]: Connection closed by 139.178.89.65 port 55848 Mar 25 01:30:09.360344 sshd-session[3965]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:09.367168 systemd[1]: sshd@9-143.198.155.16:22-139.178.89.65:55848.service: Deactivated successfully. Mar 25 01:30:09.367510 systemd-logind[1463]: Session 9 logged out. Waiting for processes to exit. Mar 25 01:30:09.371276 systemd[1]: session-9.scope: Deactivated successfully. Mar 25 01:30:09.374239 systemd-logind[1463]: Removed session 9. Mar 25 01:30:12.981576 systemd[1]: Started sshd@10-143.198.155.16:22-62.234.184.59:60334.service - OpenSSH per-connection server daemon (62.234.184.59:60334). Mar 25 01:30:13.666382 sshd[3980]: Invalid user from 62.234.184.59 port 60334 Mar 25 01:30:14.378910 systemd[1]: Started sshd@11-143.198.155.16:22-139.178.89.65:55852.service - OpenSSH per-connection server daemon (139.178.89.65:55852). Mar 25 01:30:14.448678 sshd[3983]: Accepted publickey for core from 139.178.89.65 port 55852 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:14.451237 sshd-session[3983]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:14.462666 systemd-logind[1463]: New session 10 of user core. Mar 25 01:30:14.468225 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 25 01:30:14.625266 sshd[3985]: Connection closed by 139.178.89.65 port 55852 Mar 25 01:30:14.626356 sshd-session[3983]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:14.633302 systemd[1]: sshd@11-143.198.155.16:22-139.178.89.65:55852.service: Deactivated successfully. Mar 25 01:30:14.639292 systemd[1]: session-10.scope: Deactivated successfully. Mar 25 01:30:14.640981 systemd-logind[1463]: Session 10 logged out. Waiting for processes to exit. Mar 25 01:30:14.642481 systemd-logind[1463]: Removed session 10. Mar 25 01:30:17.773952 kubelet[2611]: E0325 01:30:17.773889 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:19.645619 systemd[1]: Started sshd@12-143.198.155.16:22-139.178.89.65:44838.service - OpenSSH per-connection server daemon (139.178.89.65:44838). Mar 25 01:30:19.715384 sshd[3997]: Accepted publickey for core from 139.178.89.65 port 44838 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:19.716526 sshd-session[3997]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:19.724560 systemd-logind[1463]: New session 11 of user core. Mar 25 01:30:19.732025 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 25 01:30:19.893172 sshd[3999]: Connection closed by 139.178.89.65 port 44838 Mar 25 01:30:19.893672 sshd-session[3997]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:19.901527 systemd[1]: sshd@12-143.198.155.16:22-139.178.89.65:44838.service: Deactivated successfully. Mar 25 01:30:19.906325 systemd[1]: session-11.scope: Deactivated successfully. Mar 25 01:30:19.908351 systemd-logind[1463]: Session 11 logged out. Waiting for processes to exit. Mar 25 01:30:19.912331 systemd-logind[1463]: Removed session 11. Mar 25 01:30:21.641654 sshd[3980]: Connection closed by invalid user 62.234.184.59 port 60334 [preauth] Mar 25 01:30:21.645006 systemd[1]: sshd@10-143.198.155.16:22-62.234.184.59:60334.service: Deactivated successfully. Mar 25 01:30:24.917786 systemd[1]: Started sshd@13-143.198.155.16:22-139.178.89.65:44844.service - OpenSSH per-connection server daemon (139.178.89.65:44844). Mar 25 01:30:25.000958 sshd[4014]: Accepted publickey for core from 139.178.89.65 port 44844 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:25.003849 sshd-session[4014]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:25.013163 systemd-logind[1463]: New session 12 of user core. Mar 25 01:30:25.026000 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 25 01:30:25.215042 sshd[4016]: Connection closed by 139.178.89.65 port 44844 Mar 25 01:30:25.217293 sshd-session[4014]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:25.226808 systemd-logind[1463]: Session 12 logged out. Waiting for processes to exit. Mar 25 01:30:25.227815 systemd[1]: sshd@13-143.198.155.16:22-139.178.89.65:44844.service: Deactivated successfully. Mar 25 01:30:25.231522 systemd[1]: session-12.scope: Deactivated successfully. Mar 25 01:30:25.234078 systemd-logind[1463]: Removed session 12. Mar 25 01:30:30.244313 systemd[1]: Started sshd@14-143.198.155.16:22-139.178.89.65:35738.service - OpenSSH per-connection server daemon (139.178.89.65:35738). Mar 25 01:30:30.356701 sshd[4029]: Accepted publickey for core from 139.178.89.65 port 35738 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:30.358159 sshd-session[4029]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:30.365809 systemd-logind[1463]: New session 13 of user core. Mar 25 01:30:30.373013 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 25 01:30:30.573978 sshd[4031]: Connection closed by 139.178.89.65 port 35738 Mar 25 01:30:30.576657 sshd-session[4029]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:30.584145 systemd[1]: sshd@14-143.198.155.16:22-139.178.89.65:35738.service: Deactivated successfully. Mar 25 01:30:30.588082 systemd[1]: session-13.scope: Deactivated successfully. Mar 25 01:30:30.590887 systemd-logind[1463]: Session 13 logged out. Waiting for processes to exit. Mar 25 01:30:30.593037 systemd-logind[1463]: Removed session 13. Mar 25 01:30:33.773896 kubelet[2611]: E0325 01:30:33.773829 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:35.597978 systemd[1]: Started sshd@15-143.198.155.16:22-139.178.89.65:35746.service - OpenSSH per-connection server daemon (139.178.89.65:35746). Mar 25 01:30:35.665018 sshd[4044]: Accepted publickey for core from 139.178.89.65 port 35746 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:35.666971 sshd-session[4044]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:35.674414 systemd-logind[1463]: New session 14 of user core. Mar 25 01:30:35.678894 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 25 01:30:35.773953 kubelet[2611]: E0325 01:30:35.773807 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:35.774958 kubelet[2611]: E0325 01:30:35.774917 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:35.854919 sshd[4046]: Connection closed by 139.178.89.65 port 35746 Mar 25 01:30:35.855927 sshd-session[4044]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:35.870002 systemd[1]: sshd@15-143.198.155.16:22-139.178.89.65:35746.service: Deactivated successfully. Mar 25 01:30:35.875168 systemd[1]: session-14.scope: Deactivated successfully. Mar 25 01:30:35.880950 systemd-logind[1463]: Session 14 logged out. Waiting for processes to exit. Mar 25 01:30:35.889074 systemd[1]: Started sshd@16-143.198.155.16:22-139.178.89.65:35754.service - OpenSSH per-connection server daemon (139.178.89.65:35754). Mar 25 01:30:35.898435 systemd-logind[1463]: Removed session 14. Mar 25 01:30:35.970998 sshd[4058]: Accepted publickey for core from 139.178.89.65 port 35754 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:35.982606 sshd-session[4058]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:36.003016 systemd-logind[1463]: New session 15 of user core. Mar 25 01:30:36.007953 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 25 01:30:36.254775 sshd[4061]: Connection closed by 139.178.89.65 port 35754 Mar 25 01:30:36.255596 sshd-session[4058]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:36.271391 systemd[1]: sshd@16-143.198.155.16:22-139.178.89.65:35754.service: Deactivated successfully. Mar 25 01:30:36.278918 systemd[1]: session-15.scope: Deactivated successfully. Mar 25 01:30:36.283291 systemd-logind[1463]: Session 15 logged out. Waiting for processes to exit. Mar 25 01:30:36.293064 systemd[1]: Started sshd@17-143.198.155.16:22-139.178.89.65:35764.service - OpenSSH per-connection server daemon (139.178.89.65:35764). Mar 25 01:30:36.305470 systemd-logind[1463]: Removed session 15. Mar 25 01:30:36.368311 sshd[4070]: Accepted publickey for core from 139.178.89.65 port 35764 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:36.369261 sshd-session[4070]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:36.379899 systemd-logind[1463]: New session 16 of user core. Mar 25 01:30:36.390964 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 25 01:30:36.551686 sshd[4073]: Connection closed by 139.178.89.65 port 35764 Mar 25 01:30:36.550331 sshd-session[4070]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:36.554954 systemd[1]: sshd@17-143.198.155.16:22-139.178.89.65:35764.service: Deactivated successfully. Mar 25 01:30:36.559669 systemd[1]: session-16.scope: Deactivated successfully. Mar 25 01:30:36.564137 systemd-logind[1463]: Session 16 logged out. Waiting for processes to exit. Mar 25 01:30:36.567359 systemd-logind[1463]: Removed session 16. Mar 25 01:30:36.775063 kubelet[2611]: E0325 01:30:36.775016 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:39.773053 kubelet[2611]: E0325 01:30:39.772955 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:41.566776 systemd[1]: Started sshd@18-143.198.155.16:22-139.178.89.65:37282.service - OpenSSH per-connection server daemon (139.178.89.65:37282). Mar 25 01:30:41.652939 sshd[4089]: Accepted publickey for core from 139.178.89.65 port 37282 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:41.656251 sshd-session[4089]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:41.663996 systemd-logind[1463]: New session 17 of user core. Mar 25 01:30:41.669979 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 25 01:30:41.837349 sshd[4091]: Connection closed by 139.178.89.65 port 37282 Mar 25 01:30:41.836199 sshd-session[4089]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:41.844768 systemd[1]: sshd@18-143.198.155.16:22-139.178.89.65:37282.service: Deactivated successfully. Mar 25 01:30:41.851485 systemd[1]: session-17.scope: Deactivated successfully. Mar 25 01:30:41.853904 systemd-logind[1463]: Session 17 logged out. Waiting for processes to exit. Mar 25 01:30:41.855795 systemd-logind[1463]: Removed session 17. Mar 25 01:30:46.864460 systemd[1]: Started sshd@19-143.198.155.16:22-139.178.89.65:37288.service - OpenSSH per-connection server daemon (139.178.89.65:37288). Mar 25 01:30:46.950528 sshd[4103]: Accepted publickey for core from 139.178.89.65 port 37288 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:46.951947 sshd-session[4103]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:46.961189 systemd-logind[1463]: New session 18 of user core. Mar 25 01:30:46.967925 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 25 01:30:47.143563 sshd[4105]: Connection closed by 139.178.89.65 port 37288 Mar 25 01:30:47.145982 sshd-session[4103]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:47.152252 systemd[1]: sshd@19-143.198.155.16:22-139.178.89.65:37288.service: Deactivated successfully. Mar 25 01:30:47.160364 systemd[1]: session-18.scope: Deactivated successfully. Mar 25 01:30:47.165114 systemd-logind[1463]: Session 18 logged out. Waiting for processes to exit. Mar 25 01:30:47.167569 systemd-logind[1463]: Removed session 18. Mar 25 01:30:52.165255 systemd[1]: Started sshd@20-143.198.155.16:22-139.178.89.65:60438.service - OpenSSH per-connection server daemon (139.178.89.65:60438). Mar 25 01:30:52.247705 sshd[4117]: Accepted publickey for core from 139.178.89.65 port 60438 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:52.250216 sshd-session[4117]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:52.258132 systemd-logind[1463]: New session 19 of user core. Mar 25 01:30:52.261941 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 25 01:30:52.433267 sshd[4119]: Connection closed by 139.178.89.65 port 60438 Mar 25 01:30:52.434040 sshd-session[4117]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:52.438337 systemd[1]: sshd@20-143.198.155.16:22-139.178.89.65:60438.service: Deactivated successfully. Mar 25 01:30:52.442932 systemd[1]: session-19.scope: Deactivated successfully. Mar 25 01:30:52.445580 systemd-logind[1463]: Session 19 logged out. Waiting for processes to exit. Mar 25 01:30:52.447572 systemd-logind[1463]: Removed session 19. Mar 25 01:30:56.774709 kubelet[2611]: E0325 01:30:56.773774 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:57.454423 systemd[1]: Started sshd@21-143.198.155.16:22-139.178.89.65:60444.service - OpenSSH per-connection server daemon (139.178.89.65:60444). Mar 25 01:30:57.527657 sshd[4131]: Accepted publickey for core from 139.178.89.65 port 60444 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:57.529768 sshd-session[4131]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:57.540457 systemd-logind[1463]: New session 20 of user core. Mar 25 01:30:57.546034 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 25 01:30:57.699148 sshd[4133]: Connection closed by 139.178.89.65 port 60444 Mar 25 01:30:57.699924 sshd-session[4131]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:57.711775 systemd[1]: sshd@21-143.198.155.16:22-139.178.89.65:60444.service: Deactivated successfully. Mar 25 01:30:57.714397 systemd[1]: session-20.scope: Deactivated successfully. Mar 25 01:30:57.717013 systemd-logind[1463]: Session 20 logged out. Waiting for processes to exit. Mar 25 01:30:57.720196 systemd[1]: Started sshd@22-143.198.155.16:22-139.178.89.65:60452.service - OpenSSH per-connection server daemon (139.178.89.65:60452). Mar 25 01:30:57.722384 systemd-logind[1463]: Removed session 20. Mar 25 01:30:57.788197 sshd[4144]: Accepted publickey for core from 139.178.89.65 port 60452 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:57.790047 sshd-session[4144]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:57.798545 systemd-logind[1463]: New session 21 of user core. Mar 25 01:30:57.804981 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 25 01:30:58.222246 sshd[4147]: Connection closed by 139.178.89.65 port 60452 Mar 25 01:30:58.226648 sshd-session[4144]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:58.243810 systemd[1]: Started sshd@23-143.198.155.16:22-139.178.89.65:60070.service - OpenSSH per-connection server daemon (139.178.89.65:60070). Mar 25 01:30:58.246603 systemd[1]: sshd@22-143.198.155.16:22-139.178.89.65:60452.service: Deactivated successfully. Mar 25 01:30:58.251072 systemd[1]: session-21.scope: Deactivated successfully. Mar 25 01:30:58.255839 systemd-logind[1463]: Session 21 logged out. Waiting for processes to exit. Mar 25 01:30:58.259054 systemd-logind[1463]: Removed session 21. Mar 25 01:30:58.359924 sshd[4159]: Accepted publickey for core from 139.178.89.65 port 60070 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:58.362036 sshd-session[4159]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:58.372536 systemd-logind[1463]: New session 22 of user core. Mar 25 01:30:58.378062 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 25 01:30:58.776215 kubelet[2611]: E0325 01:30:58.776143 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:30:59.463241 sshd[4164]: Connection closed by 139.178.89.65 port 60070 Mar 25 01:30:59.463839 sshd-session[4159]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:59.483698 systemd[1]: sshd@23-143.198.155.16:22-139.178.89.65:60070.service: Deactivated successfully. Mar 25 01:30:59.487972 systemd[1]: session-22.scope: Deactivated successfully. Mar 25 01:30:59.495443 systemd-logind[1463]: Session 22 logged out. Waiting for processes to exit. Mar 25 01:30:59.499551 systemd[1]: Started sshd@24-143.198.155.16:22-139.178.89.65:60084.service - OpenSSH per-connection server daemon (139.178.89.65:60084). Mar 25 01:30:59.503731 systemd-logind[1463]: Removed session 22. Mar 25 01:30:59.585023 sshd[4178]: Accepted publickey for core from 139.178.89.65 port 60084 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:30:59.586656 sshd-session[4178]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:30:59.597132 systemd-logind[1463]: New session 23 of user core. Mar 25 01:30:59.601961 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 25 01:30:59.976027 sshd[4183]: Connection closed by 139.178.89.65 port 60084 Mar 25 01:30:59.979082 sshd-session[4178]: pam_unix(sshd:session): session closed for user core Mar 25 01:30:59.993548 systemd[1]: sshd@24-143.198.155.16:22-139.178.89.65:60084.service: Deactivated successfully. Mar 25 01:30:59.998524 systemd[1]: session-23.scope: Deactivated successfully. Mar 25 01:31:00.004705 systemd-logind[1463]: Session 23 logged out. Waiting for processes to exit. Mar 25 01:31:00.007727 systemd[1]: Started sshd@25-143.198.155.16:22-139.178.89.65:60100.service - OpenSSH per-connection server daemon (139.178.89.65:60100). Mar 25 01:31:00.012238 systemd-logind[1463]: Removed session 23. Mar 25 01:31:00.086744 sshd[4192]: Accepted publickey for core from 139.178.89.65 port 60100 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:00.088018 sshd-session[4192]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:00.097603 systemd-logind[1463]: New session 24 of user core. Mar 25 01:31:00.160058 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 25 01:31:00.336768 sshd[4195]: Connection closed by 139.178.89.65 port 60100 Mar 25 01:31:00.337835 sshd-session[4192]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:00.346267 systemd[1]: sshd@25-143.198.155.16:22-139.178.89.65:60100.service: Deactivated successfully. Mar 25 01:31:00.350201 systemd[1]: session-24.scope: Deactivated successfully. Mar 25 01:31:00.351779 systemd-logind[1463]: Session 24 logged out. Waiting for processes to exit. Mar 25 01:31:00.353829 systemd-logind[1463]: Removed session 24. Mar 25 01:31:02.091764 systemd[1]: Started sshd@26-143.198.155.16:22-218.92.0.156:62968.service - OpenSSH per-connection server daemon (218.92.0.156:62968). Mar 25 01:31:03.277074 sshd-session[4211]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.156 user=root Mar 25 01:31:04.590709 sshd[4207]: PAM: Permission denied for root from 218.92.0.156 Mar 25 01:31:04.910482 sshd-session[4212]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.156 user=root Mar 25 01:31:05.356858 systemd[1]: Started sshd@27-143.198.155.16:22-139.178.89.65:60104.service - OpenSSH per-connection server daemon (139.178.89.65:60104). Mar 25 01:31:05.423924 sshd[4214]: Accepted publickey for core from 139.178.89.65 port 60104 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:05.427253 sshd-session[4214]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:05.433384 systemd-logind[1463]: New session 25 of user core. Mar 25 01:31:05.443127 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 25 01:31:05.596514 sshd[4216]: Connection closed by 139.178.89.65 port 60104 Mar 25 01:31:05.597878 sshd-session[4214]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:05.604638 systemd[1]: sshd@27-143.198.155.16:22-139.178.89.65:60104.service: Deactivated successfully. Mar 25 01:31:05.608901 systemd[1]: session-25.scope: Deactivated successfully. Mar 25 01:31:05.611600 systemd-logind[1463]: Session 25 logged out. Waiting for processes to exit. Mar 25 01:31:05.613457 systemd-logind[1463]: Removed session 25. Mar 25 01:31:06.695503 sshd[4207]: PAM: Permission denied for root from 218.92.0.156 Mar 25 01:31:07.011543 sshd-session[4226]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.156 user=root Mar 25 01:31:08.736851 sshd[4207]: PAM: Permission denied for root from 218.92.0.156 Mar 25 01:31:09.094192 sshd[4207]: Received disconnect from 218.92.0.156 port 62968:11: [preauth] Mar 25 01:31:09.094192 sshd[4207]: Disconnected from authenticating user root 218.92.0.156 port 62968 [preauth] Mar 25 01:31:09.097326 systemd[1]: sshd@26-143.198.155.16:22-218.92.0.156:62968.service: Deactivated successfully. Mar 25 01:31:10.643965 systemd[1]: Started sshd@28-143.198.155.16:22-139.178.89.65:59964.service - OpenSSH per-connection server daemon (139.178.89.65:59964). Mar 25 01:31:10.703407 sshd[4234]: Accepted publickey for core from 139.178.89.65 port 59964 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:10.706527 sshd-session[4234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:10.718992 systemd-logind[1463]: New session 26 of user core. Mar 25 01:31:10.724003 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 25 01:31:10.924620 sshd[4236]: Connection closed by 139.178.89.65 port 59964 Mar 25 01:31:10.925827 sshd-session[4234]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:10.932049 systemd[1]: sshd@28-143.198.155.16:22-139.178.89.65:59964.service: Deactivated successfully. Mar 25 01:31:10.936008 systemd[1]: session-26.scope: Deactivated successfully. Mar 25 01:31:10.940555 systemd-logind[1463]: Session 26 logged out. Waiting for processes to exit. Mar 25 01:31:10.948611 systemd-logind[1463]: Removed session 26. Mar 25 01:31:15.943694 systemd[1]: Started sshd@29-143.198.155.16:22-139.178.89.65:59972.service - OpenSSH per-connection server daemon (139.178.89.65:59972). Mar 25 01:31:16.014179 sshd[4247]: Accepted publickey for core from 139.178.89.65 port 59972 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:16.017094 sshd-session[4247]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:16.026501 systemd-logind[1463]: New session 27 of user core. Mar 25 01:31:16.031159 systemd[1]: Started session-27.scope - Session 27 of User core. Mar 25 01:31:16.191332 sshd[4249]: Connection closed by 139.178.89.65 port 59972 Mar 25 01:31:16.190138 sshd-session[4247]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:16.195930 systemd[1]: sshd@29-143.198.155.16:22-139.178.89.65:59972.service: Deactivated successfully. Mar 25 01:31:16.199548 systemd[1]: session-27.scope: Deactivated successfully. Mar 25 01:31:16.203005 systemd-logind[1463]: Session 27 logged out. Waiting for processes to exit. Mar 25 01:31:16.205271 systemd-logind[1463]: Removed session 27. Mar 25 01:31:21.207064 systemd[1]: Started sshd@30-143.198.155.16:22-139.178.89.65:49630.service - OpenSSH per-connection server daemon (139.178.89.65:49630). Mar 25 01:31:21.273018 sshd[4260]: Accepted publickey for core from 139.178.89.65 port 49630 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:21.275263 sshd-session[4260]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:21.282135 systemd-logind[1463]: New session 28 of user core. Mar 25 01:31:21.293036 systemd[1]: Started session-28.scope - Session 28 of User core. Mar 25 01:31:21.434891 sshd[4262]: Connection closed by 139.178.89.65 port 49630 Mar 25 01:31:21.435796 sshd-session[4260]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:21.447209 systemd[1]: sshd@30-143.198.155.16:22-139.178.89.65:49630.service: Deactivated successfully. Mar 25 01:31:21.450168 systemd[1]: session-28.scope: Deactivated successfully. Mar 25 01:31:21.452898 systemd-logind[1463]: Session 28 logged out. Waiting for processes to exit. Mar 25 01:31:21.461124 systemd[1]: Started sshd@31-143.198.155.16:22-139.178.89.65:49642.service - OpenSSH per-connection server daemon (139.178.89.65:49642). Mar 25 01:31:21.463262 systemd-logind[1463]: Removed session 28. Mar 25 01:31:21.523258 sshd[4272]: Accepted publickey for core from 139.178.89.65 port 49642 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:21.525486 sshd-session[4272]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:21.533967 systemd-logind[1463]: New session 29 of user core. Mar 25 01:31:21.543021 systemd[1]: Started session-29.scope - Session 29 of User core. Mar 25 01:31:23.335945 containerd[1487]: time="2025-03-25T01:31:23.335816459Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 25 01:31:23.340106 containerd[1487]: time="2025-03-25T01:31:23.339786192Z" level=info msg="StopContainer for \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" with timeout 30 (s)" Mar 25 01:31:23.343739 containerd[1487]: time="2025-03-25T01:31:23.343673106Z" level=info msg="Stop container \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" with signal terminated" Mar 25 01:31:23.346925 containerd[1487]: time="2025-03-25T01:31:23.346865381Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" id:\"990e2581b070acd91d1d69c6d91ecc6b7226f20bd9e422d35d6a3e7e40b95ae8\" pid:4293 exited_at:{seconds:1742866283 nanos:346224671}" Mar 25 01:31:23.350576 containerd[1487]: time="2025-03-25T01:31:23.350388617Z" level=info msg="StopContainer for \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" with timeout 2 (s)" Mar 25 01:31:23.351585 containerd[1487]: time="2025-03-25T01:31:23.351436258Z" level=info msg="Stop container \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" with signal terminated" Mar 25 01:31:23.368489 systemd[1]: cri-containerd-3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124.scope: Deactivated successfully. Mar 25 01:31:23.373275 containerd[1487]: time="2025-03-25T01:31:23.373230519Z" level=info msg="received exit event container_id:\"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" id:\"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" pid:3220 exited_at:{seconds:1742866283 nanos:372577247}" Mar 25 01:31:23.374862 containerd[1487]: time="2025-03-25T01:31:23.374777773Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" id:\"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" pid:3220 exited_at:{seconds:1742866283 nanos:372577247}" Mar 25 01:31:23.377428 systemd-networkd[1366]: lxc_health: Link DOWN Mar 25 01:31:23.377976 systemd-networkd[1366]: lxc_health: Lost carrier Mar 25 01:31:23.418055 systemd[1]: cri-containerd-0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc.scope: Deactivated successfully. Mar 25 01:31:23.418920 systemd[1]: cri-containerd-0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc.scope: Consumed 10.500s CPU time, 188M memory peak, 69.3M read from disk, 13.3M written to disk. Mar 25 01:31:23.430967 containerd[1487]: time="2025-03-25T01:31:23.430900111Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" id:\"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" pid:3252 exited_at:{seconds:1742866283 nanos:430314961}" Mar 25 01:31:23.432704 containerd[1487]: time="2025-03-25T01:31:23.431545612Z" level=info msg="received exit event container_id:\"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" id:\"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" pid:3252 exited_at:{seconds:1742866283 nanos:430314961}" Mar 25 01:31:23.459351 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124-rootfs.mount: Deactivated successfully. Mar 25 01:31:23.481376 containerd[1487]: time="2025-03-25T01:31:23.481092701Z" level=error msg="collecting metrics for 3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124" error="ttrpc: closed" Mar 25 01:31:23.481376 containerd[1487]: time="2025-03-25T01:31:23.481221910Z" level=info msg="StopContainer for \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" returns successfully" Mar 25 01:31:23.488813 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc-rootfs.mount: Deactivated successfully. Mar 25 01:31:23.499616 containerd[1487]: time="2025-03-25T01:31:23.498819140Z" level=info msg="StopPodSandbox for \"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\"" Mar 25 01:31:23.508815 containerd[1487]: time="2025-03-25T01:31:23.508719509Z" level=info msg="StopContainer for \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" returns successfully" Mar 25 01:31:23.511620 containerd[1487]: time="2025-03-25T01:31:23.511117313Z" level=info msg="StopPodSandbox for \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\"" Mar 25 01:31:23.511620 containerd[1487]: time="2025-03-25T01:31:23.511242880Z" level=info msg="Container to stop \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 25 01:31:23.511620 containerd[1487]: time="2025-03-25T01:31:23.511268785Z" level=info msg="Container to stop \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 25 01:31:23.511620 containerd[1487]: time="2025-03-25T01:31:23.511285737Z" level=info msg="Container to stop \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 25 01:31:23.511620 containerd[1487]: time="2025-03-25T01:31:23.511313716Z" level=info msg="Container to stop \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 25 01:31:23.511620 containerd[1487]: time="2025-03-25T01:31:23.511327773Z" level=info msg="Container to stop \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 25 01:31:23.516748 containerd[1487]: time="2025-03-25T01:31:23.516670633Z" level=info msg="Container to stop \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 25 01:31:23.524197 systemd[1]: cri-containerd-d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638.scope: Deactivated successfully. Mar 25 01:31:23.526454 containerd[1487]: time="2025-03-25T01:31:23.526404618Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" id:\"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" pid:2754 exit_status:137 exited_at:{seconds:1742866283 nanos:526042101}" Mar 25 01:31:23.534445 systemd[1]: cri-containerd-851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca.scope: Deactivated successfully. Mar 25 01:31:23.580982 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638-rootfs.mount: Deactivated successfully. Mar 25 01:31:23.586603 containerd[1487]: time="2025-03-25T01:31:23.586217905Z" level=info msg="shim disconnected" id=d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638 namespace=k8s.io Mar 25 01:31:23.586603 containerd[1487]: time="2025-03-25T01:31:23.586248412Z" level=warning msg="cleaning up after shim disconnected" id=d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638 namespace=k8s.io Mar 25 01:31:23.586603 containerd[1487]: time="2025-03-25T01:31:23.586256164Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 25 01:31:23.596996 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca-rootfs.mount: Deactivated successfully. Mar 25 01:31:23.601113 containerd[1487]: time="2025-03-25T01:31:23.601035328Z" level=info msg="shim disconnected" id=851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca namespace=k8s.io Mar 25 01:31:23.601113 containerd[1487]: time="2025-03-25T01:31:23.601081791Z" level=warning msg="cleaning up after shim disconnected" id=851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca namespace=k8s.io Mar 25 01:31:23.601736 containerd[1487]: time="2025-03-25T01:31:23.601112012Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 25 01:31:23.628694 containerd[1487]: time="2025-03-25T01:31:23.626709528Z" level=info msg="TaskExit event in podsandbox handler container_id:\"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\" id:\"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\" pid:2837 exit_status:137 exited_at:{seconds:1742866283 nanos:542187088}" Mar 25 01:31:23.628881 containerd[1487]: time="2025-03-25T01:31:23.628831621Z" level=info msg="received exit event sandbox_id:\"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\" exit_status:137 exited_at:{seconds:1742866283 nanos:542187088}" Mar 25 01:31:23.631214 containerd[1487]: time="2025-03-25T01:31:23.629182890Z" level=info msg="TearDown network for sandbox \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" successfully" Mar 25 01:31:23.631214 containerd[1487]: time="2025-03-25T01:31:23.629220775Z" level=info msg="StopPodSandbox for \"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" returns successfully" Mar 25 01:31:23.631214 containerd[1487]: time="2025-03-25T01:31:23.629374206Z" level=info msg="received exit event sandbox_id:\"d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638\" exit_status:137 exited_at:{seconds:1742866283 nanos:526042101}" Mar 25 01:31:23.630868 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d97002c5a7b4061af05ac13d169c17ec02a096f6829b90c546b5ef8d44489638-shm.mount: Deactivated successfully. Mar 25 01:31:23.637489 containerd[1487]: time="2025-03-25T01:31:23.637403100Z" level=info msg="TearDown network for sandbox \"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\" successfully" Mar 25 01:31:23.640445 containerd[1487]: time="2025-03-25T01:31:23.637436147Z" level=info msg="StopPodSandbox for \"851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca\" returns successfully" Mar 25 01:31:23.775014 kubelet[2611]: E0325 01:31:23.773257 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:23.795234 kubelet[2611]: I0325 01:31:23.794723 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-kernel\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795234 kubelet[2611]: I0325 01:31:23.794853 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-etc-cni-netd\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795234 kubelet[2611]: I0325 01:31:23.794873 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-lib-modules\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795234 kubelet[2611]: I0325 01:31:23.794894 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-clustermesh-secrets\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795234 kubelet[2611]: I0325 01:31:23.794920 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9ca7be5e-1545-4171-a3c5-a6faa9445e65-cilium-config-path\") pod \"9ca7be5e-1545-4171-a3c5-a6faa9445e65\" (UID: \"9ca7be5e-1545-4171-a3c5-a6faa9445e65\") " Mar 25 01:31:23.795234 kubelet[2611]: I0325 01:31:23.794941 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-bpf-maps\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795541 kubelet[2611]: I0325 01:31:23.794957 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-run\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795541 kubelet[2611]: I0325 01:31:23.794972 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-config-path\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795541 kubelet[2611]: I0325 01:31:23.794986 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-net\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795541 kubelet[2611]: I0325 01:31:23.795000 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cni-path\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795541 kubelet[2611]: I0325 01:31:23.795015 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hostproc\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795541 kubelet[2611]: I0325 01:31:23.795034 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvh4b\" (UniqueName: \"kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-kube-api-access-mvh4b\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795745 kubelet[2611]: I0325 01:31:23.795051 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hubble-tls\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795745 kubelet[2611]: I0325 01:31:23.795067 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-cgroup\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.795745 kubelet[2611]: I0325 01:31:23.795084 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/9ca7be5e-1545-4171-a3c5-a6faa9445e65-kube-api-access-85z98\") pod \"9ca7be5e-1545-4171-a3c5-a6faa9445e65\" (UID: \"9ca7be5e-1545-4171-a3c5-a6faa9445e65\") " Mar 25 01:31:23.795745 kubelet[2611]: I0325 01:31:23.795097 2611 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-xtables-lock\") pod \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\" (UID: \"af367a1f-cf63-4ea2-9d0d-322d41aeb2ca\") " Mar 25 01:31:23.799207 kubelet[2611]: I0325 01:31:23.799142 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.799861 kubelet[2611]: I0325 01:31:23.799470 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.799861 kubelet[2611]: I0325 01:31:23.799466 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.799861 kubelet[2611]: I0325 01:31:23.799503 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.799861 kubelet[2611]: I0325 01:31:23.799565 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cni-path" (OuterVolumeSpecName: "cni-path") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.802036 kubelet[2611]: I0325 01:31:23.801980 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ca7be5e-1545-4171-a3c5-a6faa9445e65-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "9ca7be5e-1545-4171-a3c5-a6faa9445e65" (UID: "9ca7be5e-1545-4171-a3c5-a6faa9445e65"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 25 01:31:23.802195 kubelet[2611]: I0325 01:31:23.802069 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.802195 kubelet[2611]: I0325 01:31:23.802096 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.804777 kubelet[2611]: I0325 01:31:23.804137 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 25 01:31:23.804777 kubelet[2611]: I0325 01:31:23.804340 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hostproc" (OuterVolumeSpecName: "hostproc") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.805059 kubelet[2611]: I0325 01:31:23.805026 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 25 01:31:23.805096 kubelet[2611]: I0325 01:31:23.805080 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.805125 kubelet[2611]: I0325 01:31:23.805108 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 25 01:31:23.819445 kubelet[2611]: I0325 01:31:23.819331 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-kube-api-access-mvh4b" (OuterVolumeSpecName: "kube-api-access-mvh4b") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "kube-api-access-mvh4b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 25 01:31:23.819445 kubelet[2611]: I0325 01:31:23.819428 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" (UID: "af367a1f-cf63-4ea2-9d0d-322d41aeb2ca"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 25 01:31:23.819686 kubelet[2611]: I0325 01:31:23.819485 2611 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ca7be5e-1545-4171-a3c5-a6faa9445e65-kube-api-access-85z98" (OuterVolumeSpecName: "kube-api-access-85z98") pod "9ca7be5e-1545-4171-a3c5-a6faa9445e65" (UID: "9ca7be5e-1545-4171-a3c5-a6faa9445e65"). InnerVolumeSpecName "kube-api-access-85z98". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.895941 2611 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-bpf-maps\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.895995 2611 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cni-path\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.896005 2611 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-run\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.896016 2611 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-config-path\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.896027 2611 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-net\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.896037 2611 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hubble-tls\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.896046 2611 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-hostproc\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.896721 kubelet[2611]: I0325 01:31:23.896055 2611 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mvh4b\" (UniqueName: \"kubernetes.io/projected/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-kube-api-access-mvh4b\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896065 2611 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-cilium-cgroup\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896076 2611 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-85z98\" (UniqueName: \"kubernetes.io/projected/9ca7be5e-1545-4171-a3c5-a6faa9445e65-kube-api-access-85z98\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896086 2611 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-xtables-lock\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896095 2611 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-lib-modules\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896103 2611 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-clustermesh-secrets\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896111 2611 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-host-proc-sys-kernel\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896120 2611 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca-etc-cni-netd\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:23.897063 kubelet[2611]: I0325 01:31:23.896127 2611 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9ca7be5e-1545-4171-a3c5-a6faa9445e65-cilium-config-path\") on node \"ci-4284.0.0-c-20cdc03f82\" DevicePath \"\"" Mar 25 01:31:24.438661 kubelet[2611]: I0325 01:31:24.437725 2611 scope.go:117] "RemoveContainer" containerID="3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124" Mar 25 01:31:24.447546 containerd[1487]: time="2025-03-25T01:31:24.447281757Z" level=info msg="RemoveContainer for \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\"" Mar 25 01:31:24.448915 systemd[1]: Removed slice kubepods-besteffort-pod9ca7be5e_1545_4171_a3c5_a6faa9445e65.slice - libcontainer container kubepods-besteffort-pod9ca7be5e_1545_4171_a3c5_a6faa9445e65.slice. Mar 25 01:31:24.460185 containerd[1487]: time="2025-03-25T01:31:24.460126349Z" level=info msg="RemoveContainer for \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" returns successfully" Mar 25 01:31:24.460777 kubelet[2611]: I0325 01:31:24.460620 2611 scope.go:117] "RemoveContainer" containerID="3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124" Mar 25 01:31:24.460736 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-851f810dd0bc76feaef68d2e08ef16c8183e5055ae4fe55af1421fe1fae8a5ca-shm.mount: Deactivated successfully. Mar 25 01:31:24.460923 systemd[1]: var-lib-kubelet-pods-9ca7be5e\x2d1545\x2d4171\x2da3c5\x2da6faa9445e65-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d85z98.mount: Deactivated successfully. Mar 25 01:31:24.461098 systemd[1]: var-lib-kubelet-pods-af367a1f\x2dcf63\x2d4ea2\x2d9d0d\x2d322d41aeb2ca-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dmvh4b.mount: Deactivated successfully. Mar 25 01:31:24.461224 systemd[1]: var-lib-kubelet-pods-af367a1f\x2dcf63\x2d4ea2\x2d9d0d\x2d322d41aeb2ca-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 25 01:31:24.461522 systemd[1]: var-lib-kubelet-pods-af367a1f\x2dcf63\x2d4ea2\x2d9d0d\x2d322d41aeb2ca-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 25 01:31:24.488168 containerd[1487]: time="2025-03-25T01:31:24.464463023Z" level=error msg="ContainerStatus for \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\": not found" Mar 25 01:31:24.491467 systemd[1]: Removed slice kubepods-burstable-podaf367a1f_cf63_4ea2_9d0d_322d41aeb2ca.slice - libcontainer container kubepods-burstable-podaf367a1f_cf63_4ea2_9d0d_322d41aeb2ca.slice. Mar 25 01:31:24.492426 systemd[1]: kubepods-burstable-podaf367a1f_cf63_4ea2_9d0d_322d41aeb2ca.slice: Consumed 10.629s CPU time, 188.3M memory peak, 69.4M read from disk, 15.6M written to disk. Mar 25 01:31:24.494383 kubelet[2611]: E0325 01:31:24.493951 2611 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\": not found" containerID="3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124" Mar 25 01:31:24.512948 kubelet[2611]: I0325 01:31:24.494033 2611 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124"} err="failed to get container status \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\": rpc error: code = NotFound desc = an error occurred when try to find container \"3ce77233c90b7e82a5e3618e03423a26ece2ac3651457041895956ca03c4f124\": not found" Mar 25 01:31:24.515867 kubelet[2611]: I0325 01:31:24.512910 2611 scope.go:117] "RemoveContainer" containerID="0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc" Mar 25 01:31:24.520040 containerd[1487]: time="2025-03-25T01:31:24.519974987Z" level=info msg="RemoveContainer for \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\"" Mar 25 01:31:24.529673 containerd[1487]: time="2025-03-25T01:31:24.529570571Z" level=info msg="RemoveContainer for \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" returns successfully" Mar 25 01:31:24.529970 kubelet[2611]: I0325 01:31:24.529937 2611 scope.go:117] "RemoveContainer" containerID="878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a" Mar 25 01:31:24.540610 containerd[1487]: time="2025-03-25T01:31:24.540543564Z" level=info msg="RemoveContainer for \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\"" Mar 25 01:31:24.548500 containerd[1487]: time="2025-03-25T01:31:24.548413563Z" level=info msg="RemoveContainer for \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" returns successfully" Mar 25 01:31:24.549261 kubelet[2611]: I0325 01:31:24.549224 2611 scope.go:117] "RemoveContainer" containerID="9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d" Mar 25 01:31:24.553056 containerd[1487]: time="2025-03-25T01:31:24.552994469Z" level=info msg="RemoveContainer for \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\"" Mar 25 01:31:24.557312 containerd[1487]: time="2025-03-25T01:31:24.557182415Z" level=info msg="RemoveContainer for \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" returns successfully" Mar 25 01:31:24.557975 kubelet[2611]: I0325 01:31:24.557715 2611 scope.go:117] "RemoveContainer" containerID="6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab" Mar 25 01:31:24.561557 containerd[1487]: time="2025-03-25T01:31:24.560875985Z" level=info msg="RemoveContainer for \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\"" Mar 25 01:31:24.564823 containerd[1487]: time="2025-03-25T01:31:24.564760789Z" level=info msg="RemoveContainer for \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" returns successfully" Mar 25 01:31:24.565358 kubelet[2611]: I0325 01:31:24.565332 2611 scope.go:117] "RemoveContainer" containerID="4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396" Mar 25 01:31:24.567965 containerd[1487]: time="2025-03-25T01:31:24.567914294Z" level=info msg="RemoveContainer for \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\"" Mar 25 01:31:24.572352 containerd[1487]: time="2025-03-25T01:31:24.572295305Z" level=info msg="RemoveContainer for \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" returns successfully" Mar 25 01:31:24.572884 kubelet[2611]: I0325 01:31:24.572844 2611 scope.go:117] "RemoveContainer" containerID="0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc" Mar 25 01:31:24.573924 containerd[1487]: time="2025-03-25T01:31:24.573304090Z" level=error msg="ContainerStatus for \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\": not found" Mar 25 01:31:24.574096 kubelet[2611]: E0325 01:31:24.573561 2611 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\": not found" containerID="0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc" Mar 25 01:31:24.574096 kubelet[2611]: I0325 01:31:24.573602 2611 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc"} err="failed to get container status \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\": rpc error: code = NotFound desc = an error occurred when try to find container \"0c6bc57f174a89bbedf86cebbd09d662dcaab4730adb648e78eba75fcf9740dc\": not found" Mar 25 01:31:24.574096 kubelet[2611]: I0325 01:31:24.573665 2611 scope.go:117] "RemoveContainer" containerID="878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a" Mar 25 01:31:24.574621 containerd[1487]: time="2025-03-25T01:31:24.574454309Z" level=error msg="ContainerStatus for \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\": not found" Mar 25 01:31:24.574824 kubelet[2611]: E0325 01:31:24.574779 2611 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\": not found" containerID="878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a" Mar 25 01:31:24.574948 kubelet[2611]: I0325 01:31:24.574827 2611 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a"} err="failed to get container status \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\": rpc error: code = NotFound desc = an error occurred when try to find container \"878eab7a66ca377c865432e2cb307d7894b1c9723d56fcb5931425dd269bf75a\": not found" Mar 25 01:31:24.574948 kubelet[2611]: I0325 01:31:24.574860 2611 scope.go:117] "RemoveContainer" containerID="9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d" Mar 25 01:31:24.575400 kubelet[2611]: E0325 01:31:24.575329 2611 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\": not found" containerID="9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d" Mar 25 01:31:24.575400 kubelet[2611]: I0325 01:31:24.575361 2611 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d"} err="failed to get container status \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\": rpc error: code = NotFound desc = an error occurred when try to find container \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\": not found" Mar 25 01:31:24.575400 kubelet[2611]: I0325 01:31:24.575392 2611 scope.go:117] "RemoveContainer" containerID="6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab" Mar 25 01:31:24.575530 containerd[1487]: time="2025-03-25T01:31:24.575171142Z" level=error msg="ContainerStatus for \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9f036f06dc8afdaba7aeca390c6bc1afe72f0881cc95d89571878f2e5f92918d\": not found" Mar 25 01:31:24.575726 containerd[1487]: time="2025-03-25T01:31:24.575654838Z" level=error msg="ContainerStatus for \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\": not found" Mar 25 01:31:24.575984 kubelet[2611]: E0325 01:31:24.575874 2611 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\": not found" containerID="6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab" Mar 25 01:31:24.575984 kubelet[2611]: I0325 01:31:24.575905 2611 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab"} err="failed to get container status \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\": rpc error: code = NotFound desc = an error occurred when try to find container \"6fccc7151a1821451c05227c4d7aba7f30fece439d7fc40645fa798b324700ab\": not found" Mar 25 01:31:24.575984 kubelet[2611]: I0325 01:31:24.575927 2611 scope.go:117] "RemoveContainer" containerID="4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396" Mar 25 01:31:24.576564 containerd[1487]: time="2025-03-25T01:31:24.576331603Z" level=error msg="ContainerStatus for \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\": not found" Mar 25 01:31:24.576675 kubelet[2611]: E0325 01:31:24.576496 2611 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\": not found" containerID="4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396" Mar 25 01:31:24.576675 kubelet[2611]: I0325 01:31:24.576525 2611 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396"} err="failed to get container status \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\": rpc error: code = NotFound desc = an error occurred when try to find container \"4e884eac1fa0af1f55fc18c6f5053d4a8428670ae0b7ac56f759defce04f5396\": not found" Mar 25 01:31:24.776466 kubelet[2611]: I0325 01:31:24.776402 2611 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ca7be5e-1545-4171-a3c5-a6faa9445e65" path="/var/lib/kubelet/pods/9ca7be5e-1545-4171-a3c5-a6faa9445e65/volumes" Mar 25 01:31:24.777044 kubelet[2611]: I0325 01:31:24.776879 2611 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" path="/var/lib/kubelet/pods/af367a1f-cf63-4ea2-9d0d-322d41aeb2ca/volumes" Mar 25 01:31:25.219593 sshd[4275]: Connection closed by 139.178.89.65 port 49642 Mar 25 01:31:25.223054 sshd-session[4272]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:25.240719 systemd[1]: sshd@31-143.198.155.16:22-139.178.89.65:49642.service: Deactivated successfully. Mar 25 01:31:25.244747 systemd[1]: session-29.scope: Deactivated successfully. Mar 25 01:31:25.245442 systemd[1]: session-29.scope: Consumed 1.000s CPU time, 28.3M memory peak. Mar 25 01:31:25.248257 systemd-logind[1463]: Session 29 logged out. Waiting for processes to exit. Mar 25 01:31:25.252882 systemd[1]: Started sshd@32-143.198.155.16:22-139.178.89.65:49644.service - OpenSSH per-connection server daemon (139.178.89.65:49644). Mar 25 01:31:25.255183 systemd-logind[1463]: Removed session 29. Mar 25 01:31:25.321498 sshd[4422]: Accepted publickey for core from 139.178.89.65 port 49644 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:25.323795 sshd-session[4422]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:25.331459 systemd-logind[1463]: New session 30 of user core. Mar 25 01:31:25.345988 systemd[1]: Started session-30.scope - Session 30 of User core. Mar 25 01:31:25.992626 sshd[4425]: Connection closed by 139.178.89.65 port 49644 Mar 25 01:31:25.993100 sshd-session[4422]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:26.009139 systemd[1]: sshd@32-143.198.155.16:22-139.178.89.65:49644.service: Deactivated successfully. Mar 25 01:31:26.014220 systemd[1]: session-30.scope: Deactivated successfully. Mar 25 01:31:26.020653 systemd-logind[1463]: Session 30 logged out. Waiting for processes to exit. Mar 25 01:31:26.024136 systemd[1]: Started sshd@33-143.198.155.16:22-139.178.89.65:49658.service - OpenSSH per-connection server daemon (139.178.89.65:49658). Mar 25 01:31:26.030367 systemd-logind[1463]: Removed session 30. Mar 25 01:31:26.058936 kubelet[2611]: I0325 01:31:26.058360 2611 memory_manager.go:355] "RemoveStaleState removing state" podUID="af367a1f-cf63-4ea2-9d0d-322d41aeb2ca" containerName="cilium-agent" Mar 25 01:31:26.058936 kubelet[2611]: I0325 01:31:26.058399 2611 memory_manager.go:355] "RemoveStaleState removing state" podUID="9ca7be5e-1545-4171-a3c5-a6faa9445e65" containerName="cilium-operator" Mar 25 01:31:26.141484 systemd[1]: Created slice kubepods-burstable-pod77fa7cf6_965a_471b_85d0_6f8190073ba5.slice - libcontainer container kubepods-burstable-pod77fa7cf6_965a_471b_85d0_6f8190073ba5.slice. Mar 25 01:31:26.171865 sshd[4434]: Accepted publickey for core from 139.178.89.65 port 49658 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:26.173431 sshd-session[4434]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:26.190188 systemd-logind[1463]: New session 31 of user core. Mar 25 01:31:26.194884 systemd[1]: Started session-31.scope - Session 31 of User core. Mar 25 01:31:26.221431 kubelet[2611]: I0325 01:31:26.220914 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/77fa7cf6-965a-471b-85d0-6f8190073ba5-cilium-config-path\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221431 kubelet[2611]: I0325 01:31:26.220977 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/77fa7cf6-965a-471b-85d0-6f8190073ba5-clustermesh-secrets\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221431 kubelet[2611]: I0325 01:31:26.221002 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-cilium-cgroup\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221431 kubelet[2611]: I0325 01:31:26.221026 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-etc-cni-netd\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221431 kubelet[2611]: I0325 01:31:26.221051 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzxwq\" (UniqueName: \"kubernetes.io/projected/77fa7cf6-965a-471b-85d0-6f8190073ba5-kube-api-access-fzxwq\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221431 kubelet[2611]: I0325 01:31:26.221080 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-bpf-maps\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221864 kubelet[2611]: I0325 01:31:26.221104 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/77fa7cf6-965a-471b-85d0-6f8190073ba5-hubble-tls\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221864 kubelet[2611]: I0325 01:31:26.221124 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-xtables-lock\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221864 kubelet[2611]: I0325 01:31:26.221147 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/77fa7cf6-965a-471b-85d0-6f8190073ba5-cilium-ipsec-secrets\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221864 kubelet[2611]: I0325 01:31:26.221173 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-host-proc-sys-net\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221864 kubelet[2611]: I0325 01:31:26.221200 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-hostproc\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.221864 kubelet[2611]: I0325 01:31:26.221222 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-host-proc-sys-kernel\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.222098 kubelet[2611]: I0325 01:31:26.221246 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-cilium-run\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.222098 kubelet[2611]: I0325 01:31:26.221280 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-lib-modules\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.222098 kubelet[2611]: I0325 01:31:26.221304 2611 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/77fa7cf6-965a-471b-85d0-6f8190073ba5-cni-path\") pod \"cilium-p6xtq\" (UID: \"77fa7cf6-965a-471b-85d0-6f8190073ba5\") " pod="kube-system/cilium-p6xtq" Mar 25 01:31:26.261041 sshd[4437]: Connection closed by 139.178.89.65 port 49658 Mar 25 01:31:26.259897 sshd-session[4434]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:26.273705 systemd[1]: sshd@33-143.198.155.16:22-139.178.89.65:49658.service: Deactivated successfully. Mar 25 01:31:26.279874 systemd[1]: session-31.scope: Deactivated successfully. Mar 25 01:31:26.287036 systemd-logind[1463]: Session 31 logged out. Waiting for processes to exit. Mar 25 01:31:26.291209 systemd[1]: Started sshd@34-143.198.155.16:22-139.178.89.65:49674.service - OpenSSH per-connection server daemon (139.178.89.65:49674). Mar 25 01:31:26.294083 systemd-logind[1463]: Removed session 31. Mar 25 01:31:26.405591 sshd[4444]: Accepted publickey for core from 139.178.89.65 port 49674 ssh2: RSA SHA256:PEWrDSaOsqsnd5IJWLSBxoLVDP2Pw8xv31tYLVecIBo Mar 25 01:31:26.407683 sshd-session[4444]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 25 01:31:26.414608 systemd-logind[1463]: New session 32 of user core. Mar 25 01:31:26.420943 systemd[1]: Started session-32.scope - Session 32 of User core. Mar 25 01:31:26.446013 kubelet[2611]: E0325 01:31:26.445841 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:26.447278 containerd[1487]: time="2025-03-25T01:31:26.447076429Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-p6xtq,Uid:77fa7cf6-965a-471b-85d0-6f8190073ba5,Namespace:kube-system,Attempt:0,}" Mar 25 01:31:26.473692 containerd[1487]: time="2025-03-25T01:31:26.473146457Z" level=info msg="connecting to shim 6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201" address="unix:///run/containerd/s/faeabbc52218faebff60dafa7cfa3e773a72d82943edf015c1962422a06bbe21" namespace=k8s.io protocol=ttrpc version=3 Mar 25 01:31:26.520031 systemd[1]: Started cri-containerd-6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201.scope - libcontainer container 6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201. Mar 25 01:31:26.577675 containerd[1487]: time="2025-03-25T01:31:26.574616803Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-p6xtq,Uid:77fa7cf6-965a-471b-85d0-6f8190073ba5,Namespace:kube-system,Attempt:0,} returns sandbox id \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\"" Mar 25 01:31:26.577953 kubelet[2611]: E0325 01:31:26.577202 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:26.583741 containerd[1487]: time="2025-03-25T01:31:26.583181510Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 25 01:31:26.594460 containerd[1487]: time="2025-03-25T01:31:26.594401079Z" level=info msg="Container 9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:31:26.612338 containerd[1487]: time="2025-03-25T01:31:26.609743032Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730\"" Mar 25 01:31:26.615930 containerd[1487]: time="2025-03-25T01:31:26.615817541Z" level=info msg="StartContainer for \"9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730\"" Mar 25 01:31:26.618418 containerd[1487]: time="2025-03-25T01:31:26.618158899Z" level=info msg="connecting to shim 9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730" address="unix:///run/containerd/s/faeabbc52218faebff60dafa7cfa3e773a72d82943edf015c1962422a06bbe21" protocol=ttrpc version=3 Mar 25 01:31:26.652026 systemd[1]: Started cri-containerd-9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730.scope - libcontainer container 9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730. Mar 25 01:31:26.700398 containerd[1487]: time="2025-03-25T01:31:26.700342436Z" level=info msg="StartContainer for \"9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730\" returns successfully" Mar 25 01:31:26.714536 systemd[1]: cri-containerd-9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730.scope: Deactivated successfully. Mar 25 01:31:26.716184 containerd[1487]: time="2025-03-25T01:31:26.716133075Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730\" id:\"9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730\" pid:4516 exited_at:{seconds:1742866286 nanos:715608830}" Mar 25 01:31:26.716789 containerd[1487]: time="2025-03-25T01:31:26.716409872Z" level=info msg="received exit event container_id:\"9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730\" id:\"9a57e7195cdfd6c55ca00f7d633ea011d33ba3a5093774666b4749eb4e0f8730\" pid:4516 exited_at:{seconds:1742866286 nanos:715608830}" Mar 25 01:31:27.487739 kubelet[2611]: E0325 01:31:27.487699 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:27.499972 containerd[1487]: time="2025-03-25T01:31:27.497385588Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 25 01:31:27.509983 containerd[1487]: time="2025-03-25T01:31:27.509925335Z" level=info msg="Container 9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:31:27.518872 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2800512161.mount: Deactivated successfully. Mar 25 01:31:27.530219 containerd[1487]: time="2025-03-25T01:31:27.530142578Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420\"" Mar 25 01:31:27.533755 containerd[1487]: time="2025-03-25T01:31:27.532971426Z" level=info msg="StartContainer for \"9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420\"" Mar 25 01:31:27.537433 containerd[1487]: time="2025-03-25T01:31:27.537381433Z" level=info msg="connecting to shim 9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420" address="unix:///run/containerd/s/faeabbc52218faebff60dafa7cfa3e773a72d82943edf015c1962422a06bbe21" protocol=ttrpc version=3 Mar 25 01:31:27.567049 systemd[1]: Started cri-containerd-9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420.scope - libcontainer container 9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420. Mar 25 01:31:27.614124 containerd[1487]: time="2025-03-25T01:31:27.614075983Z" level=info msg="StartContainer for \"9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420\" returns successfully" Mar 25 01:31:27.624151 systemd[1]: cri-containerd-9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420.scope: Deactivated successfully. Mar 25 01:31:27.626197 containerd[1487]: time="2025-03-25T01:31:27.625901094Z" level=info msg="received exit event container_id:\"9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420\" id:\"9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420\" pid:4560 exited_at:{seconds:1742866287 nanos:625594196}" Mar 25 01:31:27.626197 containerd[1487]: time="2025-03-25T01:31:27.626137061Z" level=info msg="TaskExit event in podsandbox handler container_id:\"9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420\" id:\"9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420\" pid:4560 exited_at:{seconds:1742866287 nanos:625594196}" Mar 25 01:31:27.663945 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9a5c33e244040976a9317f2353fd9400c551be6fae957e7118378a08868de420-rootfs.mount: Deactivated successfully. Mar 25 01:31:27.987809 kubelet[2611]: E0325 01:31:27.987743 2611 kubelet.go:3008] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 25 01:31:28.494375 kubelet[2611]: E0325 01:31:28.493311 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:28.496437 containerd[1487]: time="2025-03-25T01:31:28.496290326Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 25 01:31:28.520671 containerd[1487]: time="2025-03-25T01:31:28.516114209Z" level=info msg="Container 034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:31:28.539730 containerd[1487]: time="2025-03-25T01:31:28.539672976Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885\"" Mar 25 01:31:28.540893 containerd[1487]: time="2025-03-25T01:31:28.540844974Z" level=info msg="StartContainer for \"034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885\"" Mar 25 01:31:28.543793 containerd[1487]: time="2025-03-25T01:31:28.543735811Z" level=info msg="connecting to shim 034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885" address="unix:///run/containerd/s/faeabbc52218faebff60dafa7cfa3e773a72d82943edf015c1962422a06bbe21" protocol=ttrpc version=3 Mar 25 01:31:28.582042 systemd[1]: Started cri-containerd-034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885.scope - libcontainer container 034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885. Mar 25 01:31:28.644836 containerd[1487]: time="2025-03-25T01:31:28.644601131Z" level=info msg="StartContainer for \"034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885\" returns successfully" Mar 25 01:31:28.650861 systemd[1]: cri-containerd-034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885.scope: Deactivated successfully. Mar 25 01:31:28.653561 containerd[1487]: time="2025-03-25T01:31:28.653444620Z" level=info msg="TaskExit event in podsandbox handler container_id:\"034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885\" id:\"034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885\" pid:4604 exited_at:{seconds:1742866288 nanos:652585111}" Mar 25 01:31:28.653561 containerd[1487]: time="2025-03-25T01:31:28.653565598Z" level=info msg="received exit event container_id:\"034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885\" id:\"034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885\" pid:4604 exited_at:{seconds:1742866288 nanos:652585111}" Mar 25 01:31:28.686225 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-034f9212821e8a1278e5e180cd977e638d8238feb91e6375b12f5e9b4a94e885-rootfs.mount: Deactivated successfully. Mar 25 01:31:29.500892 kubelet[2611]: E0325 01:31:29.500854 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:29.508175 containerd[1487]: time="2025-03-25T01:31:29.505980150Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 25 01:31:29.533368 containerd[1487]: time="2025-03-25T01:31:29.531389014Z" level=info msg="Container be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:31:29.541779 containerd[1487]: time="2025-03-25T01:31:29.541727222Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b\"" Mar 25 01:31:29.543085 containerd[1487]: time="2025-03-25T01:31:29.543046620Z" level=info msg="StartContainer for \"be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b\"" Mar 25 01:31:29.544404 containerd[1487]: time="2025-03-25T01:31:29.544331331Z" level=info msg="connecting to shim be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b" address="unix:///run/containerd/s/faeabbc52218faebff60dafa7cfa3e773a72d82943edf015c1962422a06bbe21" protocol=ttrpc version=3 Mar 25 01:31:29.596930 systemd[1]: Started cri-containerd-be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b.scope - libcontainer container be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b. Mar 25 01:31:29.681073 systemd[1]: cri-containerd-be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b.scope: Deactivated successfully. Mar 25 01:31:29.681593 containerd[1487]: time="2025-03-25T01:31:29.681542157Z" level=info msg="received exit event container_id:\"be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b\" id:\"be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b\" pid:4643 exited_at:{seconds:1742866289 nanos:680929002}" Mar 25 01:31:29.683581 containerd[1487]: time="2025-03-25T01:31:29.683521977Z" level=info msg="TaskExit event in podsandbox handler container_id:\"be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b\" id:\"be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b\" pid:4643 exited_at:{seconds:1742866289 nanos:680929002}" Mar 25 01:31:29.697590 containerd[1487]: time="2025-03-25T01:31:29.697123570Z" level=info msg="StartContainer for \"be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b\" returns successfully" Mar 25 01:31:29.714854 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-be83c19b48dedbbb47ec324ff2bb4ce5eacb09af5a2a58089880823c6957a09b-rootfs.mount: Deactivated successfully. Mar 25 01:31:30.509681 kubelet[2611]: E0325 01:31:30.509074 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:30.514432 containerd[1487]: time="2025-03-25T01:31:30.514370334Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 25 01:31:30.532858 containerd[1487]: time="2025-03-25T01:31:30.532796339Z" level=info msg="Container 51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55: CDI devices from CRI Config.CDIDevices: []" Mar 25 01:31:30.539753 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount505470428.mount: Deactivated successfully. Mar 25 01:31:30.544544 containerd[1487]: time="2025-03-25T01:31:30.544472587Z" level=info msg="CreateContainer within sandbox \"6cfcac251d09b378229ac7a100041779a732233604f133c3844877a4f7e28201\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\"" Mar 25 01:31:30.549719 containerd[1487]: time="2025-03-25T01:31:30.547570685Z" level=info msg="StartContainer for \"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\"" Mar 25 01:31:30.550730 containerd[1487]: time="2025-03-25T01:31:30.550680882Z" level=info msg="connecting to shim 51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55" address="unix:///run/containerd/s/faeabbc52218faebff60dafa7cfa3e773a72d82943edf015c1962422a06bbe21" protocol=ttrpc version=3 Mar 25 01:31:30.595918 systemd[1]: Started cri-containerd-51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55.scope - libcontainer container 51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55. Mar 25 01:31:30.637474 containerd[1487]: time="2025-03-25T01:31:30.637342856Z" level=info msg="StartContainer for \"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" returns successfully" Mar 25 01:31:30.752179 containerd[1487]: time="2025-03-25T01:31:30.752124073Z" level=info msg="TaskExit event in podsandbox handler container_id:\"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" id:\"a1e6ea162b7f1c1ff384b31bac0c1a90fa37b8e79035ce56703b9993b02e48e7\" pid:4708 exited_at:{seconds:1742866290 nanos:751744508}" Mar 25 01:31:31.173685 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Mar 25 01:31:31.518128 kubelet[2611]: E0325 01:31:31.517782 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:31.540995 kubelet[2611]: I0325 01:31:31.539815 2611 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-p6xtq" podStartSLOduration=5.539772619 podStartE2EDuration="5.539772619s" podCreationTimestamp="2025-03-25 01:31:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-25 01:31:31.539492815 +0000 UTC m=+149.014071032" watchObservedRunningTime="2025-03-25 01:31:31.539772619 +0000 UTC m=+149.014350825" Mar 25 01:31:32.521970 kubelet[2611]: E0325 01:31:32.520965 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:33.273224 containerd[1487]: time="2025-03-25T01:31:33.273112417Z" level=info msg="TaskExit event in podsandbox handler container_id:\"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" id:\"892364fa8d0bc0625e61568183f5fbf8f75657a421e674154cbac91f0b88758e\" pid:4892 exit_status:1 exited_at:{seconds:1742866293 nanos:272581945}" Mar 25 01:31:35.183251 systemd-networkd[1366]: lxc_health: Link UP Mar 25 01:31:35.203917 systemd-networkd[1366]: lxc_health: Gained carrier Mar 25 01:31:35.521922 containerd[1487]: time="2025-03-25T01:31:35.521860333Z" level=info msg="TaskExit event in podsandbox handler container_id:\"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" id:\"8e482d0296955b6b3190c486b71f67e49d2b418f2fea85c7ff4d7fd086dd955b\" pid:5260 exit_status:1 exited_at:{seconds:1742866295 nanos:520826709}" Mar 25 01:31:36.450771 kubelet[2611]: E0325 01:31:36.448811 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:36.537403 kubelet[2611]: E0325 01:31:36.536885 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:37.090011 systemd-networkd[1366]: lxc_health: Gained IPv6LL Mar 25 01:31:37.539835 kubelet[2611]: E0325 01:31:37.539798 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:38.183434 containerd[1487]: time="2025-03-25T01:31:38.183381920Z" level=info msg="TaskExit event in podsandbox handler container_id:\"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" id:\"1fd8bc4aa68378ede874b42c35ddba20d276f7ea6b00824ee090ec06e9cd2b80\" pid:5292 exited_at:{seconds:1742866298 nanos:182558319}" Mar 25 01:31:38.774351 kubelet[2611]: E0325 01:31:38.774289 2611 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Mar 25 01:31:40.645176 containerd[1487]: time="2025-03-25T01:31:40.645036058Z" level=info msg="TaskExit event in podsandbox handler container_id:\"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" id:\"e85731ca3e0924cafd4106bb77579dc3fdaa8aca634cfa61228a67a1c138cb5d\" pid:5317 exited_at:{seconds:1742866300 nanos:643981967}" Mar 25 01:31:42.900182 containerd[1487]: time="2025-03-25T01:31:42.900010761Z" level=info msg="TaskExit event in podsandbox handler container_id:\"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" id:\"8d35a4c825ad7c973a282a8a0b27c95427dc9cf79d8c53e5c0cdb90c01bba5c7\" pid:5355 exited_at:{seconds:1742866302 nanos:896648889}" Mar 25 01:31:45.086943 containerd[1487]: time="2025-03-25T01:31:45.086796737Z" level=info msg="TaskExit event in podsandbox handler container_id:\"51fbc13e607374943cf3eb54d3c45b5ac19a47b3cbf2f416c10f8cce210e5d55\" id:\"d136e2e9f95c0d330b09b1e6ad2d8b59260d57fbc1a0880ca26cc7e8b6147fde\" pid:5383 exited_at:{seconds:1742866305 nanos:86179929}" Mar 25 01:31:45.125268 sshd[4452]: Connection closed by 139.178.89.65 port 49674 Mar 25 01:31:45.126743 sshd-session[4444]: pam_unix(sshd:session): session closed for user core Mar 25 01:31:45.131981 systemd-logind[1463]: Session 32 logged out. Waiting for processes to exit. Mar 25 01:31:45.144059 systemd[1]: sshd@34-143.198.155.16:22-139.178.89.65:49674.service: Deactivated successfully. Mar 25 01:31:45.148450 systemd[1]: session-32.scope: Deactivated successfully. Mar 25 01:31:45.150980 systemd-logind[1463]: Removed session 32.