Oct 9 07:53:24.035375 kernel: Linux version 6.6.54-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Oct 8 18:24:27 -00 2024 Oct 9 07:53:24.035462 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ed527eaf992abc270af9987554566193214d123941456fd3066b47855e5178a5 Oct 9 07:53:24.035486 kernel: BIOS-provided physical RAM map: Oct 9 07:53:24.035498 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Oct 9 07:53:24.035508 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Oct 9 07:53:24.035520 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Oct 9 07:53:24.035534 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Oct 9 07:53:24.035546 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Oct 9 07:53:24.035556 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Oct 9 07:53:24.035570 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Oct 9 07:53:24.035587 kernel: NX (Execute Disable) protection: active Oct 9 07:53:24.035599 kernel: APIC: Static calls initialized Oct 9 07:53:24.035611 kernel: SMBIOS 2.8 present. Oct 9 07:53:24.035623 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Oct 9 07:53:24.035639 kernel: Hypervisor detected: KVM Oct 9 07:53:24.035660 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Oct 9 07:53:24.035672 kernel: kvm-clock: using sched offset of 3093970794 cycles Oct 9 07:53:24.035691 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Oct 9 07:53:24.035734 kernel: tsc: Detected 2494.140 MHz processor Oct 9 07:53:24.035750 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Oct 9 07:53:24.035767 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Oct 9 07:53:24.035782 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Oct 9 07:53:24.035797 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Oct 9 07:53:24.035810 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Oct 9 07:53:24.035831 kernel: ACPI: Early table checksum verification disabled Oct 9 07:53:24.035844 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Oct 9 07:53:24.035859 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 9 07:53:24.035872 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 9 07:53:24.035883 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 9 07:53:24.035897 kernel: ACPI: FACS 0x000000007FFE0000 000040 Oct 9 07:53:24.035909 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 9 07:53:24.035921 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 9 07:53:24.035950 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 9 07:53:24.035993 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Oct 9 07:53:24.036032 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Oct 9 07:53:24.036048 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Oct 9 07:53:24.036060 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Oct 9 07:53:24.036073 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Oct 9 07:53:24.036085 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Oct 9 07:53:24.036097 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Oct 9 07:53:24.036126 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Oct 9 07:53:24.036146 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Oct 9 07:53:24.036161 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Oct 9 07:53:24.036175 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Oct 9 07:53:24.036189 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Oct 9 07:53:24.036204 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Oct 9 07:53:24.036219 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Oct 9 07:53:24.036257 kernel: Zone ranges: Oct 9 07:53:24.036272 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Oct 9 07:53:24.036287 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Oct 9 07:53:24.036302 kernel: Normal empty Oct 9 07:53:24.036316 kernel: Movable zone start for each node Oct 9 07:53:24.036330 kernel: Early memory node ranges Oct 9 07:53:24.036345 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Oct 9 07:53:24.036361 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Oct 9 07:53:24.036376 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Oct 9 07:53:24.036398 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Oct 9 07:53:24.036418 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Oct 9 07:53:24.036432 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Oct 9 07:53:24.036448 kernel: ACPI: PM-Timer IO Port: 0x608 Oct 9 07:53:24.036461 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Oct 9 07:53:24.036474 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Oct 9 07:53:24.036486 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Oct 9 07:53:24.036499 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Oct 9 07:53:24.036513 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Oct 9 07:53:24.036549 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Oct 9 07:53:24.036563 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Oct 9 07:53:24.036581 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Oct 9 07:53:24.036593 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Oct 9 07:53:24.036606 kernel: TSC deadline timer available Oct 9 07:53:24.036621 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Oct 9 07:53:24.036637 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Oct 9 07:53:24.036652 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Oct 9 07:53:24.036668 kernel: Booting paravirtualized kernel on KVM Oct 9 07:53:24.036687 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Oct 9 07:53:24.036946 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Oct 9 07:53:24.036968 kernel: percpu: Embedded 58 pages/cpu s196904 r8192 d32472 u1048576 Oct 9 07:53:24.036985 kernel: pcpu-alloc: s196904 r8192 d32472 u1048576 alloc=1*2097152 Oct 9 07:53:24.037001 kernel: pcpu-alloc: [0] 0 1 Oct 9 07:53:24.037017 kernel: kvm-guest: PV spinlocks disabled, no host support Oct 9 07:53:24.037034 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ed527eaf992abc270af9987554566193214d123941456fd3066b47855e5178a5 Oct 9 07:53:24.037050 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Oct 9 07:53:24.037065 kernel: random: crng init done Oct 9 07:53:24.037092 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Oct 9 07:53:24.037109 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Oct 9 07:53:24.037125 kernel: Fallback order for Node 0: 0 Oct 9 07:53:24.037141 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Oct 9 07:53:24.037155 kernel: Policy zone: DMA32 Oct 9 07:53:24.037169 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Oct 9 07:53:24.037185 kernel: Memory: 1971204K/2096612K available (12288K kernel code, 2305K rwdata, 22716K rodata, 42828K init, 2360K bss, 125148K reserved, 0K cma-reserved) Oct 9 07:53:24.037200 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Oct 9 07:53:24.037221 kernel: Kernel/User page tables isolation: enabled Oct 9 07:53:24.037235 kernel: ftrace: allocating 37784 entries in 148 pages Oct 9 07:53:24.037250 kernel: ftrace: allocated 148 pages with 3 groups Oct 9 07:53:24.037265 kernel: Dynamic Preempt: voluntary Oct 9 07:53:24.037281 kernel: rcu: Preemptible hierarchical RCU implementation. Oct 9 07:53:24.037298 kernel: rcu: RCU event tracing is enabled. Oct 9 07:53:24.037314 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Oct 9 07:53:24.037328 kernel: Trampoline variant of Tasks RCU enabled. Oct 9 07:53:24.037342 kernel: Rude variant of Tasks RCU enabled. Oct 9 07:53:24.037356 kernel: Tracing variant of Tasks RCU enabled. Oct 9 07:53:24.037376 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Oct 9 07:53:24.037389 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Oct 9 07:53:24.037402 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Oct 9 07:53:24.037428 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Oct 9 07:53:24.037444 kernel: Console: colour VGA+ 80x25 Oct 9 07:53:24.037457 kernel: printk: console [tty0] enabled Oct 9 07:53:24.037469 kernel: printk: console [ttyS0] enabled Oct 9 07:53:24.037482 kernel: ACPI: Core revision 20230628 Oct 9 07:53:24.037496 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Oct 9 07:53:24.037518 kernel: APIC: Switch to symmetric I/O mode setup Oct 9 07:53:24.037534 kernel: x2apic enabled Oct 9 07:53:24.037551 kernel: APIC: Switched APIC routing to: physical x2apic Oct 9 07:53:24.037564 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Oct 9 07:53:24.037580 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Oct 9 07:53:24.037597 kernel: Calibrating delay loop (skipped) preset value.. 4988.28 BogoMIPS (lpj=2494140) Oct 9 07:53:24.037614 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Oct 9 07:53:24.037631 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Oct 9 07:53:24.037669 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Oct 9 07:53:24.037686 kernel: Spectre V2 : Mitigation: Retpolines Oct 9 07:53:24.037703 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Oct 9 07:53:24.037770 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Oct 9 07:53:24.037784 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Oct 9 07:53:24.037797 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Oct 9 07:53:24.037810 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Oct 9 07:53:24.037823 kernel: MDS: Mitigation: Clear CPU buffers Oct 9 07:53:24.037838 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Oct 9 07:53:24.037870 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Oct 9 07:53:24.037894 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Oct 9 07:53:24.037929 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Oct 9 07:53:24.037964 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Oct 9 07:53:24.037989 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Oct 9 07:53:24.038007 kernel: Freeing SMP alternatives memory: 32K Oct 9 07:53:24.038024 kernel: pid_max: default: 32768 minimum: 301 Oct 9 07:53:24.038039 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Oct 9 07:53:24.038062 kernel: landlock: Up and running. Oct 9 07:53:24.038078 kernel: SELinux: Initializing. Oct 9 07:53:24.038095 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 9 07:53:24.038111 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 9 07:53:24.038127 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x4f, stepping: 0x1) Oct 9 07:53:24.038142 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Oct 9 07:53:24.038156 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Oct 9 07:53:24.038172 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Oct 9 07:53:24.038190 kernel: Performance Events: unsupported p6 CPU model 79 no PMU driver, software events only. Oct 9 07:53:24.038214 kernel: signal: max sigframe size: 1776 Oct 9 07:53:24.038230 kernel: rcu: Hierarchical SRCU implementation. Oct 9 07:53:24.038245 kernel: rcu: Max phase no-delay instances is 400. Oct 9 07:53:24.038261 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Oct 9 07:53:24.038275 kernel: smp: Bringing up secondary CPUs ... Oct 9 07:53:24.038290 kernel: smpboot: x86: Booting SMP configuration: Oct 9 07:53:24.038323 kernel: .... node #0, CPUs: #1 Oct 9 07:53:24.038349 kernel: smp: Brought up 1 node, 2 CPUs Oct 9 07:53:24.038366 kernel: smpboot: Max logical packages: 1 Oct 9 07:53:24.038389 kernel: smpboot: Total of 2 processors activated (9976.56 BogoMIPS) Oct 9 07:53:24.038404 kernel: devtmpfs: initialized Oct 9 07:53:24.038419 kernel: x86/mm: Memory block size: 128MB Oct 9 07:53:24.038435 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Oct 9 07:53:24.038451 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Oct 9 07:53:24.038465 kernel: pinctrl core: initialized pinctrl subsystem Oct 9 07:53:24.038482 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Oct 9 07:53:24.038497 kernel: audit: initializing netlink subsys (disabled) Oct 9 07:53:24.038514 kernel: thermal_sys: Registered thermal governor 'step_wise' Oct 9 07:53:24.038537 kernel: thermal_sys: Registered thermal governor 'user_space' Oct 9 07:53:24.038552 kernel: audit: type=2000 audit(1728460402.866:1): state=initialized audit_enabled=0 res=1 Oct 9 07:53:24.038568 kernel: cpuidle: using governor menu Oct 9 07:53:24.038583 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Oct 9 07:53:24.038599 kernel: dca service started, version 1.12.1 Oct 9 07:53:24.038613 kernel: PCI: Using configuration type 1 for base access Oct 9 07:53:24.038629 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Oct 9 07:53:24.038645 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Oct 9 07:53:24.038662 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Oct 9 07:53:24.038683 kernel: ACPI: Added _OSI(Module Device) Oct 9 07:53:24.038699 kernel: ACPI: Added _OSI(Processor Device) Oct 9 07:53:24.038744 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Oct 9 07:53:24.038761 kernel: ACPI: Added _OSI(Processor Aggregator Device) Oct 9 07:53:24.038777 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Oct 9 07:53:24.038793 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Oct 9 07:53:24.038810 kernel: ACPI: Interpreter enabled Oct 9 07:53:24.038826 kernel: ACPI: PM: (supports S0 S5) Oct 9 07:53:24.038843 kernel: ACPI: Using IOAPIC for interrupt routing Oct 9 07:53:24.038923 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Oct 9 07:53:24.038940 kernel: PCI: Using E820 reservations for host bridge windows Oct 9 07:53:24.038954 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Oct 9 07:53:24.038970 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Oct 9 07:53:24.039376 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Oct 9 07:53:24.039608 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Oct 9 07:53:24.039818 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Oct 9 07:53:24.039853 kernel: acpiphp: Slot [3] registered Oct 9 07:53:24.039870 kernel: acpiphp: Slot [4] registered Oct 9 07:53:24.039887 kernel: acpiphp: Slot [5] registered Oct 9 07:53:24.039903 kernel: acpiphp: Slot [6] registered Oct 9 07:53:24.039920 kernel: acpiphp: Slot [7] registered Oct 9 07:53:24.039937 kernel: acpiphp: Slot [8] registered Oct 9 07:53:24.039955 kernel: acpiphp: Slot [9] registered Oct 9 07:53:24.039971 kernel: acpiphp: Slot [10] registered Oct 9 07:53:24.039989 kernel: acpiphp: Slot [11] registered Oct 9 07:53:24.040010 kernel: acpiphp: Slot [12] registered Oct 9 07:53:24.040026 kernel: acpiphp: Slot [13] registered Oct 9 07:53:24.040043 kernel: acpiphp: Slot [14] registered Oct 9 07:53:24.040060 kernel: acpiphp: Slot [15] registered Oct 9 07:53:24.040076 kernel: acpiphp: Slot [16] registered Oct 9 07:53:24.040093 kernel: acpiphp: Slot [17] registered Oct 9 07:53:24.040124 kernel: acpiphp: Slot [18] registered Oct 9 07:53:24.040140 kernel: acpiphp: Slot [19] registered Oct 9 07:53:24.040158 kernel: acpiphp: Slot [20] registered Oct 9 07:53:24.040176 kernel: acpiphp: Slot [21] registered Oct 9 07:53:24.040199 kernel: acpiphp: Slot [22] registered Oct 9 07:53:24.040214 kernel: acpiphp: Slot [23] registered Oct 9 07:53:24.040228 kernel: acpiphp: Slot [24] registered Oct 9 07:53:24.040244 kernel: acpiphp: Slot [25] registered Oct 9 07:53:24.040262 kernel: acpiphp: Slot [26] registered Oct 9 07:53:24.040280 kernel: acpiphp: Slot [27] registered Oct 9 07:53:24.040298 kernel: acpiphp: Slot [28] registered Oct 9 07:53:24.040314 kernel: acpiphp: Slot [29] registered Oct 9 07:53:24.040328 kernel: acpiphp: Slot [30] registered Oct 9 07:53:24.040350 kernel: acpiphp: Slot [31] registered Oct 9 07:53:24.040366 kernel: PCI host bridge to bus 0000:00 Oct 9 07:53:24.040624 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Oct 9 07:53:24.040874 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Oct 9 07:53:24.044893 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Oct 9 07:53:24.045142 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Oct 9 07:53:24.045278 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Oct 9 07:53:24.045396 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Oct 9 07:53:24.045613 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Oct 9 07:53:24.045901 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Oct 9 07:53:24.046153 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Oct 9 07:53:24.046376 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Oct 9 07:53:24.046579 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Oct 9 07:53:24.047921 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Oct 9 07:53:24.048188 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Oct 9 07:53:24.048388 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Oct 9 07:53:24.048636 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Oct 9 07:53:24.050032 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Oct 9 07:53:24.050286 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Oct 9 07:53:24.050483 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Oct 9 07:53:24.050688 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Oct 9 07:53:24.052205 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Oct 9 07:53:24.052458 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Oct 9 07:53:24.052661 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Oct 9 07:53:24.052890 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Oct 9 07:53:24.053061 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Oct 9 07:53:24.053227 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Oct 9 07:53:24.053466 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Oct 9 07:53:24.053649 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Oct 9 07:53:24.054994 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Oct 9 07:53:24.055200 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Oct 9 07:53:24.055424 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Oct 9 07:53:24.055591 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Oct 9 07:53:24.055781 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Oct 9 07:53:24.055967 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Oct 9 07:53:24.056157 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Oct 9 07:53:24.056321 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Oct 9 07:53:24.056490 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Oct 9 07:53:24.056651 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Oct 9 07:53:24.058108 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Oct 9 07:53:24.058320 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Oct 9 07:53:24.058514 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Oct 9 07:53:24.058679 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Oct 9 07:53:24.058928 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Oct 9 07:53:24.059134 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Oct 9 07:53:24.059313 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Oct 9 07:53:24.059488 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Oct 9 07:53:24.059699 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Oct 9 07:53:24.061171 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Oct 9 07:53:24.061356 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Oct 9 07:53:24.061379 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Oct 9 07:53:24.061394 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Oct 9 07:53:24.061410 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Oct 9 07:53:24.061424 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Oct 9 07:53:24.061438 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Oct 9 07:53:24.061463 kernel: iommu: Default domain type: Translated Oct 9 07:53:24.061477 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Oct 9 07:53:24.061491 kernel: PCI: Using ACPI for IRQ routing Oct 9 07:53:24.061505 kernel: PCI: pci_cache_line_size set to 64 bytes Oct 9 07:53:24.061519 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Oct 9 07:53:24.061533 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Oct 9 07:53:24.061848 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Oct 9 07:53:24.062047 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Oct 9 07:53:24.062246 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Oct 9 07:53:24.062271 kernel: vgaarb: loaded Oct 9 07:53:24.062285 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Oct 9 07:53:24.062301 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Oct 9 07:53:24.062317 kernel: clocksource: Switched to clocksource kvm-clock Oct 9 07:53:24.062333 kernel: VFS: Disk quotas dquot_6.6.0 Oct 9 07:53:24.062350 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Oct 9 07:53:24.062368 kernel: pnp: PnP ACPI init Oct 9 07:53:24.062385 kernel: pnp: PnP ACPI: found 4 devices Oct 9 07:53:24.062411 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Oct 9 07:53:24.062429 kernel: NET: Registered PF_INET protocol family Oct 9 07:53:24.062448 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Oct 9 07:53:24.062467 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Oct 9 07:53:24.062486 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Oct 9 07:53:24.062504 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Oct 9 07:53:24.062522 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Oct 9 07:53:24.062540 kernel: TCP: Hash tables configured (established 16384 bind 16384) Oct 9 07:53:24.062555 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 9 07:53:24.062580 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 9 07:53:24.062593 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Oct 9 07:53:24.062607 kernel: NET: Registered PF_XDP protocol family Oct 9 07:53:24.064917 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Oct 9 07:53:24.065124 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Oct 9 07:53:24.065262 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Oct 9 07:53:24.065399 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Oct 9 07:53:24.065540 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Oct 9 07:53:24.065865 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Oct 9 07:53:24.066157 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Oct 9 07:53:24.066188 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Oct 9 07:53:24.066364 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x7a0 took 28469 usecs Oct 9 07:53:24.066387 kernel: PCI: CLS 0 bytes, default 64 Oct 9 07:53:24.066403 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Oct 9 07:53:24.066420 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Oct 9 07:53:24.066435 kernel: Initialise system trusted keyrings Oct 9 07:53:24.066465 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Oct 9 07:53:24.066480 kernel: Key type asymmetric registered Oct 9 07:53:24.066496 kernel: Asymmetric key parser 'x509' registered Oct 9 07:53:24.066510 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Oct 9 07:53:24.066523 kernel: io scheduler mq-deadline registered Oct 9 07:53:24.066538 kernel: io scheduler kyber registered Oct 9 07:53:24.066552 kernel: io scheduler bfq registered Oct 9 07:53:24.066566 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Oct 9 07:53:24.066581 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Oct 9 07:53:24.066596 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Oct 9 07:53:24.066617 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Oct 9 07:53:24.066633 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Oct 9 07:53:24.066648 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Oct 9 07:53:24.066663 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Oct 9 07:53:24.066677 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Oct 9 07:53:24.066690 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Oct 9 07:53:24.069070 kernel: rtc_cmos 00:03: RTC can wake from S4 Oct 9 07:53:24.069129 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Oct 9 07:53:24.069301 kernel: rtc_cmos 00:03: registered as rtc0 Oct 9 07:53:24.069496 kernel: rtc_cmos 00:03: setting system clock to 2024-10-09T07:53:23 UTC (1728460403) Oct 9 07:53:24.069644 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Oct 9 07:53:24.069664 kernel: intel_pstate: CPU model not supported Oct 9 07:53:24.069679 kernel: NET: Registered PF_INET6 protocol family Oct 9 07:53:24.069695 kernel: Segment Routing with IPv6 Oct 9 07:53:24.069795 kernel: In-situ OAM (IOAM) with IPv6 Oct 9 07:53:24.069810 kernel: NET: Registered PF_PACKET protocol family Oct 9 07:53:24.069835 kernel: Key type dns_resolver registered Oct 9 07:53:24.069848 kernel: IPI shorthand broadcast: enabled Oct 9 07:53:24.069862 kernel: sched_clock: Marking stable (936004997, 90373646)->(1130854806, -104476163) Oct 9 07:53:24.069875 kernel: registered taskstats version 1 Oct 9 07:53:24.069888 kernel: Loading compiled-in X.509 certificates Oct 9 07:53:24.069901 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.54-flatcar: 14ce23fc5070d0471461f1dd6e298a5588e7ba8f' Oct 9 07:53:24.069916 kernel: Key type .fscrypt registered Oct 9 07:53:24.069932 kernel: Key type fscrypt-provisioning registered Oct 9 07:53:24.069950 kernel: ima: No TPM chip found, activating TPM-bypass! Oct 9 07:53:24.069970 kernel: ima: Allocated hash algorithm: sha1 Oct 9 07:53:24.070003 kernel: ima: No architecture policies found Oct 9 07:53:24.070016 kernel: clk: Disabling unused clocks Oct 9 07:53:24.070029 kernel: Freeing unused kernel image (initmem) memory: 42828K Oct 9 07:53:24.070044 kernel: Write protecting the kernel read-only data: 36864k Oct 9 07:53:24.070088 kernel: Freeing unused kernel image (rodata/data gap) memory: 1860K Oct 9 07:53:24.070107 kernel: Run /init as init process Oct 9 07:53:24.070121 kernel: with arguments: Oct 9 07:53:24.070137 kernel: /init Oct 9 07:53:24.070155 kernel: with environment: Oct 9 07:53:24.070174 kernel: HOME=/ Oct 9 07:53:24.070189 kernel: TERM=linux Oct 9 07:53:24.070203 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Oct 9 07:53:24.070225 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Oct 9 07:53:24.070244 systemd[1]: Detected virtualization kvm. Oct 9 07:53:24.070260 systemd[1]: Detected architecture x86-64. Oct 9 07:53:24.070275 systemd[1]: Running in initrd. Oct 9 07:53:24.070295 systemd[1]: No hostname configured, using default hostname. Oct 9 07:53:24.070310 systemd[1]: Hostname set to . Oct 9 07:53:24.070326 systemd[1]: Initializing machine ID from VM UUID. Oct 9 07:53:24.070340 systemd[1]: Queued start job for default target initrd.target. Oct 9 07:53:24.070355 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 9 07:53:24.070369 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 9 07:53:24.070387 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Oct 9 07:53:24.070402 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 9 07:53:24.070421 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Oct 9 07:53:24.070436 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Oct 9 07:53:24.070453 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Oct 9 07:53:24.070486 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Oct 9 07:53:24.070503 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 9 07:53:24.070518 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 9 07:53:24.070540 systemd[1]: Reached target paths.target - Path Units. Oct 9 07:53:24.070559 systemd[1]: Reached target slices.target - Slice Units. Oct 9 07:53:24.070578 systemd[1]: Reached target swap.target - Swaps. Oct 9 07:53:24.070603 systemd[1]: Reached target timers.target - Timer Units. Oct 9 07:53:24.070624 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Oct 9 07:53:24.070641 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 9 07:53:24.070661 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 9 07:53:24.070675 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Oct 9 07:53:24.070691 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 9 07:53:24.073728 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 9 07:53:24.073809 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 9 07:53:24.073828 systemd[1]: Reached target sockets.target - Socket Units. Oct 9 07:53:24.073845 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Oct 9 07:53:24.073861 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 9 07:53:24.073890 systemd[1]: Finished network-cleanup.service - Network Cleanup. Oct 9 07:53:24.073908 systemd[1]: Starting systemd-fsck-usr.service... Oct 9 07:53:24.073925 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 9 07:53:24.073944 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 9 07:53:24.073961 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 9 07:53:24.073978 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Oct 9 07:53:24.073992 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 9 07:53:24.074070 systemd-journald[181]: Collecting audit messages is disabled. Oct 9 07:53:24.074121 systemd[1]: Finished systemd-fsck-usr.service. Oct 9 07:53:24.074139 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 9 07:53:24.074192 systemd-journald[181]: Journal started Oct 9 07:53:24.074232 systemd-journald[181]: Runtime Journal (/run/log/journal/d386489df0b640e8ad024550d651dee9) is 4.9M, max 39.3M, 34.4M free. Oct 9 07:53:24.039613 systemd-modules-load[183]: Inserted module 'overlay' Oct 9 07:53:24.082755 systemd[1]: Started systemd-journald.service - Journal Service. Oct 9 07:53:24.093837 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Oct 9 07:53:24.096991 systemd-modules-load[183]: Inserted module 'br_netfilter' Oct 9 07:53:24.097730 kernel: Bridge firewalling registered Oct 9 07:53:24.106390 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 9 07:53:24.135300 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 9 07:53:24.141298 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 9 07:53:24.142816 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 9 07:53:24.151971 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 9 07:53:24.154993 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 9 07:53:24.159981 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 9 07:53:24.163064 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 9 07:53:24.191621 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 9 07:53:24.193958 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 9 07:53:24.204153 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 9 07:53:24.205150 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 9 07:53:24.214235 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Oct 9 07:53:24.233658 dracut-cmdline[218]: dracut-dracut-053 Oct 9 07:53:24.239689 dracut-cmdline[218]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ed527eaf992abc270af9987554566193214d123941456fd3066b47855e5178a5 Oct 9 07:53:24.254820 systemd-resolved[217]: Positive Trust Anchors: Oct 9 07:53:24.254838 systemd-resolved[217]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 9 07:53:24.254875 systemd-resolved[217]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 9 07:53:24.257823 systemd-resolved[217]: Defaulting to hostname 'linux'. Oct 9 07:53:24.259997 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 9 07:53:24.260663 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 9 07:53:24.368777 kernel: SCSI subsystem initialized Oct 9 07:53:24.382751 kernel: Loading iSCSI transport class v2.0-870. Oct 9 07:53:24.397786 kernel: iscsi: registered transport (tcp) Oct 9 07:53:24.427085 kernel: iscsi: registered transport (qla4xxx) Oct 9 07:53:24.427234 kernel: QLogic iSCSI HBA Driver Oct 9 07:53:24.496849 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Oct 9 07:53:24.501058 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Oct 9 07:53:24.537487 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Oct 9 07:53:24.537607 kernel: device-mapper: uevent: version 1.0.3 Oct 9 07:53:24.538862 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Oct 9 07:53:24.588760 kernel: raid6: avx2x4 gen() 22442 MB/s Oct 9 07:53:24.605808 kernel: raid6: avx2x2 gen() 22261 MB/s Oct 9 07:53:24.623123 kernel: raid6: avx2x1 gen() 19902 MB/s Oct 9 07:53:24.623251 kernel: raid6: using algorithm avx2x4 gen() 22442 MB/s Oct 9 07:53:24.640869 kernel: raid6: .... xor() 4603 MB/s, rmw enabled Oct 9 07:53:24.641003 kernel: raid6: using avx2x2 recovery algorithm Oct 9 07:53:24.668784 kernel: xor: automatically using best checksumming function avx Oct 9 07:53:24.871779 kernel: Btrfs loaded, zoned=no, fsverity=no Oct 9 07:53:24.892847 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Oct 9 07:53:24.903172 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 9 07:53:24.935696 systemd-udevd[402]: Using default interface naming scheme 'v255'. Oct 9 07:53:24.943307 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 9 07:53:24.953035 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Oct 9 07:53:24.980046 dracut-pre-trigger[409]: rd.md=0: removing MD RAID activation Oct 9 07:53:25.040613 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Oct 9 07:53:25.047161 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 9 07:53:25.127805 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 9 07:53:25.136344 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Oct 9 07:53:25.173188 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Oct 9 07:53:25.178481 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Oct 9 07:53:25.179128 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 9 07:53:25.181997 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 9 07:53:25.193776 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Oct 9 07:53:25.228258 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Oct 9 07:53:25.276755 kernel: scsi host0: Virtio SCSI HBA Oct 9 07:53:25.278735 kernel: virtio_blk virtio4: 1/0/0 default/read/poll queues Oct 9 07:53:25.282880 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Oct 9 07:53:25.302059 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Oct 9 07:53:25.302154 kernel: GPT:9289727 != 125829119 Oct 9 07:53:25.302168 kernel: GPT:Alternate GPT header not at the end of the disk. Oct 9 07:53:25.302181 kernel: GPT:9289727 != 125829119 Oct 9 07:53:25.302193 kernel: GPT: Use GNU Parted to correct GPT errors. Oct 9 07:53:25.302220 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 9 07:53:25.302233 kernel: cryptd: max_cpu_qlen set to 1000 Oct 9 07:53:25.306788 kernel: virtio_blk virtio5: 1/0/0 default/read/poll queues Oct 9 07:53:25.307069 kernel: virtio_blk virtio5: [vdb] 968 512-byte logical blocks (496 kB/484 KiB) Oct 9 07:53:25.342820 kernel: AVX2 version of gcm_enc/dec engaged. Oct 9 07:53:25.342982 kernel: AES CTR mode by8 optimization enabled Oct 9 07:53:25.345515 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 9 07:53:25.345678 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 9 07:53:25.347548 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 9 07:53:25.348105 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 9 07:53:25.348332 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 9 07:53:25.351059 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 9 07:53:25.358176 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 9 07:53:25.406780 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (448) Oct 9 07:53:25.416092 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Oct 9 07:53:25.428114 kernel: BTRFS: device fsid a8680da2-059a-4648-a8e8-f62925ab33ec devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (469) Oct 9 07:53:25.428764 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Oct 9 07:53:25.480025 kernel: libata version 3.00 loaded. Oct 9 07:53:25.480087 kernel: ata_piix 0000:00:01.1: version 2.13 Oct 9 07:53:25.484149 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 9 07:53:25.493184 kernel: ACPI: bus type USB registered Oct 9 07:53:25.493302 kernel: usbcore: registered new interface driver usbfs Oct 9 07:53:25.493330 kernel: scsi host1: ata_piix Oct 9 07:53:25.493398 kernel: usbcore: registered new interface driver hub Oct 9 07:53:25.497769 kernel: usbcore: registered new device driver usb Oct 9 07:53:25.508512 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 9 07:53:25.513759 kernel: scsi host2: ata_piix Oct 9 07:53:25.515964 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Oct 9 07:53:25.516045 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Oct 9 07:53:25.519001 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Oct 9 07:53:25.520449 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Oct 9 07:53:25.529120 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Oct 9 07:53:25.531782 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 9 07:53:25.541989 disk-uuid[541]: Primary Header is updated. Oct 9 07:53:25.541989 disk-uuid[541]: Secondary Entries is updated. Oct 9 07:53:25.541989 disk-uuid[541]: Secondary Header is updated. Oct 9 07:53:25.550206 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Oct 9 07:53:25.550763 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Oct 9 07:53:25.551021 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Oct 9 07:53:25.551245 kernel: uhci_hcd 0000:00:01.2: irq 11, io port 0x0000c180 Oct 9 07:53:25.552780 kernel: hub 1-0:1.0: USB hub found Oct 9 07:53:25.553211 kernel: hub 1-0:1.0: 2 ports detected Oct 9 07:53:25.562781 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 9 07:53:25.569744 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 9 07:53:25.590835 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 9 07:53:26.580835 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Oct 9 07:53:26.582774 disk-uuid[542]: The operation has completed successfully. Oct 9 07:53:26.649044 systemd[1]: disk-uuid.service: Deactivated successfully. Oct 9 07:53:26.649210 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Oct 9 07:53:26.663120 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Oct 9 07:53:26.680493 sh[563]: Success Oct 9 07:53:26.700764 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Oct 9 07:53:26.779123 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Oct 9 07:53:26.797100 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Oct 9 07:53:26.805701 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Oct 9 07:53:26.820881 kernel: BTRFS info (device dm-0): first mount of filesystem a8680da2-059a-4648-a8e8-f62925ab33ec Oct 9 07:53:26.820972 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Oct 9 07:53:26.822103 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Oct 9 07:53:26.823995 kernel: BTRFS info (device dm-0): disabling log replay at mount time Oct 9 07:53:26.824106 kernel: BTRFS info (device dm-0): using free space tree Oct 9 07:53:26.833460 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Oct 9 07:53:26.835300 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Oct 9 07:53:26.845090 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Oct 9 07:53:26.849112 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Oct 9 07:53:26.865060 kernel: BTRFS info (device vda6): first mount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 9 07:53:26.865183 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 9 07:53:26.865200 kernel: BTRFS info (device vda6): using free space tree Oct 9 07:53:26.869889 kernel: BTRFS info (device vda6): auto enabling async discard Oct 9 07:53:26.887778 kernel: BTRFS info (device vda6): last unmount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 9 07:53:26.887360 systemd[1]: mnt-oem.mount: Deactivated successfully. Oct 9 07:53:26.900874 systemd[1]: Finished ignition-setup.service - Ignition (setup). Oct 9 07:53:26.908056 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Oct 9 07:53:27.033470 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 9 07:53:27.042147 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 9 07:53:27.073925 systemd-networkd[748]: lo: Link UP Oct 9 07:53:27.073936 systemd-networkd[748]: lo: Gained carrier Oct 9 07:53:27.078603 systemd-networkd[748]: Enumeration completed Oct 9 07:53:27.079134 systemd-networkd[748]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Oct 9 07:53:27.079139 systemd-networkd[748]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Oct 9 07:53:27.080487 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 9 07:53:27.081075 systemd[1]: Reached target network.target - Network. Oct 9 07:53:27.083999 systemd-networkd[748]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 9 07:53:27.084004 systemd-networkd[748]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 9 07:53:27.085403 systemd-networkd[748]: eth0: Link UP Oct 9 07:53:27.085408 systemd-networkd[748]: eth0: Gained carrier Oct 9 07:53:27.085422 systemd-networkd[748]: eth0: found matching network '/usr/lib/systemd/network/yy-digitalocean.network', based on potentially unpredictable interface name. Oct 9 07:53:27.089181 systemd-networkd[748]: eth1: Link UP Oct 9 07:53:27.089186 systemd-networkd[748]: eth1: Gained carrier Oct 9 07:53:27.089203 systemd-networkd[748]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 9 07:53:27.091995 ignition[654]: Ignition 2.19.0 Oct 9 07:53:27.092009 ignition[654]: Stage: fetch-offline Oct 9 07:53:27.092077 ignition[654]: no configs at "/usr/lib/ignition/base.d" Oct 9 07:53:27.092087 ignition[654]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 9 07:53:27.092261 ignition[654]: parsed url from cmdline: "" Oct 9 07:53:27.092266 ignition[654]: no config URL provided Oct 9 07:53:27.092272 ignition[654]: reading system config file "/usr/lib/ignition/user.ign" Oct 9 07:53:27.094795 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Oct 9 07:53:27.092288 ignition[654]: no config at "/usr/lib/ignition/user.ign" Oct 9 07:53:27.092296 ignition[654]: failed to fetch config: resource requires networking Oct 9 07:53:27.092588 ignition[654]: Ignition finished successfully Oct 9 07:53:27.100934 systemd-networkd[748]: eth0: DHCPv4 address 143.198.138.163/20, gateway 143.198.128.1 acquired from 169.254.169.253 Oct 9 07:53:27.101081 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Oct 9 07:53:27.115873 systemd-networkd[748]: eth1: DHCPv4 address 10.124.0.17/20 acquired from 169.254.169.253 Oct 9 07:53:27.125138 ignition[755]: Ignition 2.19.0 Oct 9 07:53:27.125153 ignition[755]: Stage: fetch Oct 9 07:53:27.125459 ignition[755]: no configs at "/usr/lib/ignition/base.d" Oct 9 07:53:27.125472 ignition[755]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 9 07:53:27.125597 ignition[755]: parsed url from cmdline: "" Oct 9 07:53:27.125601 ignition[755]: no config URL provided Oct 9 07:53:27.125607 ignition[755]: reading system config file "/usr/lib/ignition/user.ign" Oct 9 07:53:27.125615 ignition[755]: no config at "/usr/lib/ignition/user.ign" Oct 9 07:53:27.125638 ignition[755]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Oct 9 07:53:27.159559 ignition[755]: GET result: OK Oct 9 07:53:27.160363 ignition[755]: parsing config with SHA512: 41ca091d5a820f03037c677f522d2d8f3b5bbebb87fcc2a2f7f3d30d57ad6a0f84f9e3e1fa63e8e3da37488210a881dbe38976d23ad5ec24afdff4f4b31152d8 Oct 9 07:53:27.167938 unknown[755]: fetched base config from "system" Oct 9 07:53:27.167950 unknown[755]: fetched base config from "system" Oct 9 07:53:27.168904 ignition[755]: fetch: fetch complete Oct 9 07:53:27.167957 unknown[755]: fetched user config from "digitalocean" Oct 9 07:53:27.168915 ignition[755]: fetch: fetch passed Oct 9 07:53:27.171217 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Oct 9 07:53:27.169043 ignition[755]: Ignition finished successfully Oct 9 07:53:27.178180 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Oct 9 07:53:27.209181 ignition[763]: Ignition 2.19.0 Oct 9 07:53:27.209800 ignition[763]: Stage: kargs Oct 9 07:53:27.210170 ignition[763]: no configs at "/usr/lib/ignition/base.d" Oct 9 07:53:27.210191 ignition[763]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 9 07:53:27.215172 ignition[763]: kargs: kargs passed Oct 9 07:53:27.215320 ignition[763]: Ignition finished successfully Oct 9 07:53:27.218086 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Oct 9 07:53:27.224536 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Oct 9 07:53:27.260731 ignition[769]: Ignition 2.19.0 Oct 9 07:53:27.260747 ignition[769]: Stage: disks Oct 9 07:53:27.261025 ignition[769]: no configs at "/usr/lib/ignition/base.d" Oct 9 07:53:27.261043 ignition[769]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 9 07:53:27.263283 ignition[769]: disks: disks passed Oct 9 07:53:27.263387 ignition[769]: Ignition finished successfully Oct 9 07:53:27.264598 systemd[1]: Finished ignition-disks.service - Ignition (disks). Oct 9 07:53:27.269598 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Oct 9 07:53:27.270421 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 9 07:53:27.271282 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 9 07:53:27.272147 systemd[1]: Reached target sysinit.target - System Initialization. Oct 9 07:53:27.272878 systemd[1]: Reached target basic.target - Basic System. Oct 9 07:53:27.284135 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Oct 9 07:53:27.307692 systemd-fsck[777]: ROOT: clean, 14/553520 files, 52654/553472 blocks Oct 9 07:53:27.311858 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Oct 9 07:53:27.321107 systemd[1]: Mounting sysroot.mount - /sysroot... Oct 9 07:53:27.436781 kernel: EXT4-fs (vda9): mounted filesystem 1df90f14-3ad0-4280-9b7d-a34f65d70e4d r/w with ordered data mode. Quota mode: none. Oct 9 07:53:27.437747 systemd[1]: Mounted sysroot.mount - /sysroot. Oct 9 07:53:27.439509 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Oct 9 07:53:27.446973 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 9 07:53:27.450002 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Oct 9 07:53:27.459071 systemd[1]: Starting flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent... Oct 9 07:53:27.463784 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (785) Oct 9 07:53:27.464534 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Oct 9 07:53:27.467759 kernel: BTRFS info (device vda6): first mount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 9 07:53:27.469502 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 9 07:53:27.469573 kernel: BTRFS info (device vda6): using free space tree Oct 9 07:53:27.473365 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Oct 9 07:53:27.473451 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Oct 9 07:53:27.486439 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Oct 9 07:53:27.490389 kernel: BTRFS info (device vda6): auto enabling async discard Oct 9 07:53:27.499090 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Oct 9 07:53:27.506597 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 9 07:53:27.564658 coreos-metadata[788]: Oct 09 07:53:27.564 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 9 07:53:27.578446 coreos-metadata[788]: Oct 09 07:53:27.578 INFO Fetch successful Oct 9 07:53:27.580746 coreos-metadata[787]: Oct 09 07:53:27.580 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 9 07:53:27.591776 coreos-metadata[788]: Oct 09 07:53:27.589 INFO wrote hostname ci-4081.1.0-c-f399a429fa to /sysroot/etc/hostname Oct 9 07:53:27.590472 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Oct 9 07:53:27.598565 coreos-metadata[787]: Oct 09 07:53:27.596 INFO Fetch successful Oct 9 07:53:27.600417 initrd-setup-root[817]: cut: /sysroot/etc/passwd: No such file or directory Oct 9 07:53:27.607334 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. Oct 9 07:53:27.608430 systemd[1]: Finished flatcar-digitalocean-network.service - Flatcar DigitalOcean Network Agent. Oct 9 07:53:27.611428 initrd-setup-root[825]: cut: /sysroot/etc/group: No such file or directory Oct 9 07:53:27.617371 initrd-setup-root[832]: cut: /sysroot/etc/shadow: No such file or directory Oct 9 07:53:27.623726 initrd-setup-root[839]: cut: /sysroot/etc/gshadow: No such file or directory Oct 9 07:53:27.759364 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Oct 9 07:53:27.764899 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Oct 9 07:53:27.766972 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Oct 9 07:53:27.795753 kernel: BTRFS info (device vda6): last unmount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 9 07:53:27.818981 systemd[1]: sysroot-oem.mount: Deactivated successfully. Oct 9 07:53:27.824334 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Oct 9 07:53:27.834922 ignition[908]: INFO : Ignition 2.19.0 Oct 9 07:53:27.834922 ignition[908]: INFO : Stage: mount Oct 9 07:53:27.836640 ignition[908]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 9 07:53:27.836640 ignition[908]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 9 07:53:27.836640 ignition[908]: INFO : mount: mount passed Oct 9 07:53:27.836640 ignition[908]: INFO : Ignition finished successfully Oct 9 07:53:27.838057 systemd[1]: Finished ignition-mount.service - Ignition (mount). Oct 9 07:53:27.849025 systemd[1]: Starting ignition-files.service - Ignition (files)... Oct 9 07:53:27.877063 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 9 07:53:27.889802 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (918) Oct 9 07:53:27.892089 kernel: BTRFS info (device vda6): first mount of filesystem bfaca09e-98f3-46e8-bdd8-6fce748bf2b6 Oct 9 07:53:27.892170 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Oct 9 07:53:27.893993 kernel: BTRFS info (device vda6): using free space tree Oct 9 07:53:27.901807 kernel: BTRFS info (device vda6): auto enabling async discard Oct 9 07:53:27.904061 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 9 07:53:27.942240 ignition[935]: INFO : Ignition 2.19.0 Oct 9 07:53:27.942240 ignition[935]: INFO : Stage: files Oct 9 07:53:27.943411 ignition[935]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 9 07:53:27.943411 ignition[935]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 9 07:53:27.944318 ignition[935]: DEBUG : files: compiled without relabeling support, skipping Oct 9 07:53:27.946026 ignition[935]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Oct 9 07:53:27.946026 ignition[935]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Oct 9 07:53:27.949952 ignition[935]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Oct 9 07:53:27.950564 ignition[935]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Oct 9 07:53:27.951233 ignition[935]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Oct 9 07:53:27.950882 unknown[935]: wrote ssh authorized keys file for user: core Oct 9 07:53:27.952793 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Oct 9 07:53:27.953755 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Oct 9 07:53:27.953755 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Oct 9 07:53:27.953755 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Oct 9 07:53:27.990203 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Oct 9 07:53:28.150744 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Oct 9 07:53:28.150744 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Oct 9 07:53:28.152410 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Oct 9 07:53:28.610738 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Oct 9 07:53:28.677069 systemd-networkd[748]: eth1: Gained IPv6LL Oct 9 07:53:28.721454 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Oct 9 07:53:28.723631 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" Oct 9 07:53:28.723631 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" Oct 9 07:53:28.723631 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" Oct 9 07:53:28.726471 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Oct 9 07:53:28.728800 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-x86-64.raw: attempt #1 Oct 9 07:53:29.026193 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK Oct 9 07:53:29.061042 systemd-networkd[748]: eth0: Gained IPv6LL Oct 9 07:53:29.306535 ignition[935]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Oct 9 07:53:29.306535 ignition[935]: INFO : files: op(d): [started] processing unit "containerd.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(d): [finished] processing unit "containerd.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(f): [started] processing unit "prepare-helm.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Oct 9 07:53:29.308298 ignition[935]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Oct 9 07:53:29.308298 ignition[935]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Oct 9 07:53:29.308298 ignition[935]: INFO : files: files passed Oct 9 07:53:29.316638 ignition[935]: INFO : Ignition finished successfully Oct 9 07:53:29.310122 systemd[1]: Finished ignition-files.service - Ignition (files). Oct 9 07:53:29.318078 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Oct 9 07:53:29.321737 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Oct 9 07:53:29.329876 systemd[1]: ignition-quench.service: Deactivated successfully. Oct 9 07:53:29.331285 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Oct 9 07:53:29.341524 initrd-setup-root-after-ignition[964]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 9 07:53:29.341524 initrd-setup-root-after-ignition[964]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Oct 9 07:53:29.344485 initrd-setup-root-after-ignition[968]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 9 07:53:29.346929 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 9 07:53:29.348253 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Oct 9 07:53:29.352962 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Oct 9 07:53:29.389202 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Oct 9 07:53:29.389382 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Oct 9 07:53:29.390699 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Oct 9 07:53:29.391236 systemd[1]: Reached target initrd.target - Initrd Default Target. Oct 9 07:53:29.392213 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Oct 9 07:53:29.398195 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Oct 9 07:53:29.421239 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 9 07:53:29.430138 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Oct 9 07:53:29.443529 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Oct 9 07:53:29.444979 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 9 07:53:29.446178 systemd[1]: Stopped target timers.target - Timer Units. Oct 9 07:53:29.446955 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Oct 9 07:53:29.447136 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 9 07:53:29.448727 systemd[1]: Stopped target initrd.target - Initrd Default Target. Oct 9 07:53:29.449309 systemd[1]: Stopped target basic.target - Basic System. Oct 9 07:53:29.450214 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Oct 9 07:53:29.450854 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Oct 9 07:53:29.451704 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Oct 9 07:53:29.452541 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Oct 9 07:53:29.453479 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Oct 9 07:53:29.454389 systemd[1]: Stopped target sysinit.target - System Initialization. Oct 9 07:53:29.455126 systemd[1]: Stopped target local-fs.target - Local File Systems. Oct 9 07:53:29.455781 systemd[1]: Stopped target swap.target - Swaps. Oct 9 07:53:29.456442 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Oct 9 07:53:29.456610 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Oct 9 07:53:29.457524 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Oct 9 07:53:29.458501 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 9 07:53:29.459312 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Oct 9 07:53:29.459502 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 9 07:53:29.460096 systemd[1]: dracut-initqueue.service: Deactivated successfully. Oct 9 07:53:29.460244 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Oct 9 07:53:29.461299 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Oct 9 07:53:29.461458 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 9 07:53:29.462466 systemd[1]: ignition-files.service: Deactivated successfully. Oct 9 07:53:29.462581 systemd[1]: Stopped ignition-files.service - Ignition (files). Oct 9 07:53:29.463166 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Oct 9 07:53:29.463306 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Oct 9 07:53:29.470184 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Oct 9 07:53:29.470647 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Oct 9 07:53:29.470868 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Oct 9 07:53:29.474154 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Oct 9 07:53:29.475080 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Oct 9 07:53:29.475264 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Oct 9 07:53:29.488231 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Oct 9 07:53:29.490084 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Oct 9 07:53:29.497695 systemd[1]: initrd-cleanup.service: Deactivated successfully. Oct 9 07:53:29.499331 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Oct 9 07:53:29.509852 ignition[988]: INFO : Ignition 2.19.0 Oct 9 07:53:29.509852 ignition[988]: INFO : Stage: umount Oct 9 07:53:29.509852 ignition[988]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 9 07:53:29.509852 ignition[988]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Oct 9 07:53:29.516632 ignition[988]: INFO : umount: umount passed Oct 9 07:53:29.516632 ignition[988]: INFO : Ignition finished successfully Oct 9 07:53:29.512388 systemd[1]: ignition-mount.service: Deactivated successfully. Oct 9 07:53:29.512558 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Oct 9 07:53:29.515599 systemd[1]: ignition-disks.service: Deactivated successfully. Oct 9 07:53:29.517654 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Oct 9 07:53:29.518981 systemd[1]: ignition-kargs.service: Deactivated successfully. Oct 9 07:53:29.519072 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Oct 9 07:53:29.520348 systemd[1]: ignition-fetch.service: Deactivated successfully. Oct 9 07:53:29.520429 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Oct 9 07:53:29.520893 systemd[1]: Stopped target network.target - Network. Oct 9 07:53:29.521258 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Oct 9 07:53:29.521312 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Oct 9 07:53:29.529204 systemd[1]: Stopped target paths.target - Path Units. Oct 9 07:53:29.529564 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Oct 9 07:53:29.532835 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 9 07:53:29.533640 systemd[1]: Stopped target slices.target - Slice Units. Oct 9 07:53:29.535011 systemd[1]: Stopped target sockets.target - Socket Units. Oct 9 07:53:29.543781 systemd[1]: iscsid.socket: Deactivated successfully. Oct 9 07:53:29.544227 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Oct 9 07:53:29.544946 systemd[1]: iscsiuio.socket: Deactivated successfully. Oct 9 07:53:29.545016 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 9 07:53:29.555830 systemd[1]: ignition-setup.service: Deactivated successfully. Oct 9 07:53:29.555989 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Oct 9 07:53:29.556550 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Oct 9 07:53:29.556624 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Oct 9 07:53:29.560801 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Oct 9 07:53:29.561505 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Oct 9 07:53:29.564786 systemd[1]: sysroot-boot.mount: Deactivated successfully. Oct 9 07:53:29.566058 systemd[1]: sysroot-boot.service: Deactivated successfully. Oct 9 07:53:29.566227 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Oct 9 07:53:29.566825 systemd-networkd[748]: eth1: DHCPv6 lease lost Oct 9 07:53:29.568182 systemd[1]: systemd-resolved.service: Deactivated successfully. Oct 9 07:53:29.568369 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Oct 9 07:53:29.571854 systemd-networkd[748]: eth0: DHCPv6 lease lost Oct 9 07:53:29.573225 systemd[1]: initrd-setup-root.service: Deactivated successfully. Oct 9 07:53:29.573406 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Oct 9 07:53:29.575091 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Oct 9 07:53:29.575183 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 9 07:53:29.576450 systemd[1]: systemd-networkd.service: Deactivated successfully. Oct 9 07:53:29.576646 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Oct 9 07:53:29.579247 systemd[1]: systemd-networkd.socket: Deactivated successfully. Oct 9 07:53:29.579337 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Oct 9 07:53:29.591051 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Oct 9 07:53:29.592337 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Oct 9 07:53:29.592424 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 9 07:53:29.593386 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 9 07:53:29.593456 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 9 07:53:29.595010 systemd[1]: systemd-modules-load.service: Deactivated successfully. Oct 9 07:53:29.595126 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Oct 9 07:53:29.596283 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 9 07:53:29.609882 systemd[1]: network-cleanup.service: Deactivated successfully. Oct 9 07:53:29.610641 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Oct 9 07:53:29.614311 systemd[1]: systemd-udevd.service: Deactivated successfully. Oct 9 07:53:29.615063 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 9 07:53:29.617081 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Oct 9 07:53:29.617160 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Oct 9 07:53:29.618606 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Oct 9 07:53:29.618650 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Oct 9 07:53:29.619947 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Oct 9 07:53:29.620039 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Oct 9 07:53:29.621365 systemd[1]: dracut-cmdline.service: Deactivated successfully. Oct 9 07:53:29.621459 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Oct 9 07:53:29.622689 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 9 07:53:29.622778 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 9 07:53:29.631154 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Oct 9 07:53:29.631840 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Oct 9 07:53:29.631983 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 9 07:53:29.632655 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 9 07:53:29.632817 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 9 07:53:29.645511 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Oct 9 07:53:29.645730 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Oct 9 07:53:29.647672 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Oct 9 07:53:29.652084 systemd[1]: Starting initrd-switch-root.service - Switch Root... Oct 9 07:53:29.680463 systemd[1]: Switching root. Oct 9 07:53:29.720163 systemd-journald[181]: Journal stopped Oct 9 07:53:30.979002 systemd-journald[181]: Received SIGTERM from PID 1 (systemd). Oct 9 07:53:30.979143 kernel: SELinux: policy capability network_peer_controls=1 Oct 9 07:53:30.979169 kernel: SELinux: policy capability open_perms=1 Oct 9 07:53:30.979197 kernel: SELinux: policy capability extended_socket_class=1 Oct 9 07:53:30.979211 kernel: SELinux: policy capability always_check_network=0 Oct 9 07:53:30.979223 kernel: SELinux: policy capability cgroup_seclabel=1 Oct 9 07:53:30.979236 kernel: SELinux: policy capability nnp_nosuid_transition=1 Oct 9 07:53:30.979248 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Oct 9 07:53:30.979261 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Oct 9 07:53:30.979281 kernel: audit: type=1403 audit(1728460409.952:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Oct 9 07:53:30.979295 systemd[1]: Successfully loaded SELinux policy in 53.608ms. Oct 9 07:53:30.979337 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 18.869ms. Oct 9 07:53:30.979353 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Oct 9 07:53:30.979372 systemd[1]: Detected virtualization kvm. Oct 9 07:53:30.979387 systemd[1]: Detected architecture x86-64. Oct 9 07:53:30.979400 systemd[1]: Detected first boot. Oct 9 07:53:30.979417 systemd[1]: Hostname set to . Oct 9 07:53:30.979430 systemd[1]: Initializing machine ID from VM UUID. Oct 9 07:53:30.979443 zram_generator::config[1047]: No configuration found. Oct 9 07:53:30.979462 systemd[1]: Populated /etc with preset unit settings. Oct 9 07:53:30.979476 systemd[1]: Queued start job for default target multi-user.target. Oct 9 07:53:30.979508 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Oct 9 07:53:30.979524 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Oct 9 07:53:30.979546 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Oct 9 07:53:30.979558 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Oct 9 07:53:30.979572 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Oct 9 07:53:30.979585 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Oct 9 07:53:30.979598 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Oct 9 07:53:30.979614 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Oct 9 07:53:30.979627 systemd[1]: Created slice user.slice - User and Session Slice. Oct 9 07:53:30.979640 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 9 07:53:30.979653 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 9 07:53:30.979666 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Oct 9 07:53:30.979679 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Oct 9 07:53:30.979693 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Oct 9 07:53:30.984782 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 9 07:53:30.984844 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Oct 9 07:53:30.984872 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 9 07:53:30.984893 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Oct 9 07:53:30.984915 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 9 07:53:30.984943 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 9 07:53:30.984964 systemd[1]: Reached target slices.target - Slice Units. Oct 9 07:53:30.984986 systemd[1]: Reached target swap.target - Swaps. Oct 9 07:53:30.985012 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Oct 9 07:53:30.985035 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Oct 9 07:53:30.985057 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 9 07:53:30.985078 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Oct 9 07:53:30.985099 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 9 07:53:30.985114 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 9 07:53:30.985128 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 9 07:53:30.985143 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Oct 9 07:53:30.985163 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Oct 9 07:53:30.985182 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Oct 9 07:53:30.985206 systemd[1]: Mounting media.mount - External Media Directory... Oct 9 07:53:30.985224 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:30.985243 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Oct 9 07:53:30.985262 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Oct 9 07:53:30.985281 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Oct 9 07:53:30.985300 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Oct 9 07:53:30.985317 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 9 07:53:30.985335 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 9 07:53:30.985354 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Oct 9 07:53:30.985373 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 9 07:53:30.985392 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 9 07:53:30.985411 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 9 07:53:30.985434 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Oct 9 07:53:30.985450 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 9 07:53:30.985465 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 9 07:53:30.985479 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Oct 9 07:53:30.985499 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Oct 9 07:53:30.985512 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 9 07:53:30.985526 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 9 07:53:30.985539 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 9 07:53:30.985553 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Oct 9 07:53:30.985566 kernel: ACPI: bus type drm_connector registered Oct 9 07:53:30.985635 systemd-journald[1137]: Collecting audit messages is disabled. Oct 9 07:53:30.985674 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 9 07:53:30.985691 systemd-journald[1137]: Journal started Oct 9 07:53:30.985744 systemd-journald[1137]: Runtime Journal (/run/log/journal/d386489df0b640e8ad024550d651dee9) is 4.9M, max 39.3M, 34.4M free. Oct 9 07:53:30.992856 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:30.994751 kernel: fuse: init (API version 7.39) Oct 9 07:53:31.009740 kernel: loop: module loaded Oct 9 07:53:31.009848 systemd[1]: Started systemd-journald.service - Journal Service. Oct 9 07:53:31.010833 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Oct 9 07:53:31.015053 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Oct 9 07:53:31.015796 systemd[1]: Mounted media.mount - External Media Directory. Oct 9 07:53:31.016380 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Oct 9 07:53:31.016987 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Oct 9 07:53:31.017574 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Oct 9 07:53:31.018471 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 9 07:53:31.019240 systemd[1]: modprobe@configfs.service: Deactivated successfully. Oct 9 07:53:31.019436 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Oct 9 07:53:31.020699 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 9 07:53:31.020976 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 9 07:53:31.022191 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 9 07:53:31.022435 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 9 07:53:31.023086 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 9 07:53:31.023330 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 9 07:53:31.024530 systemd[1]: modprobe@fuse.service: Deactivated successfully. Oct 9 07:53:31.024743 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Oct 9 07:53:31.026073 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 9 07:53:31.026280 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 9 07:53:31.027003 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 9 07:53:31.028437 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 9 07:53:31.029203 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Oct 9 07:53:31.045345 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 9 07:53:31.054124 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Oct 9 07:53:31.068578 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Oct 9 07:53:31.069116 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 9 07:53:31.079157 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Oct 9 07:53:31.088194 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Oct 9 07:53:31.088979 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 9 07:53:31.099565 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Oct 9 07:53:31.103167 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 9 07:53:31.118054 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 9 07:53:31.126932 systemd-journald[1137]: Time spent on flushing to /var/log/journal/d386489df0b640e8ad024550d651dee9 is 61.564ms for 974 entries. Oct 9 07:53:31.126932 systemd-journald[1137]: System Journal (/var/log/journal/d386489df0b640e8ad024550d651dee9) is 8.0M, max 195.6M, 187.6M free. Oct 9 07:53:31.206994 systemd-journald[1137]: Received client request to flush runtime journal. Oct 9 07:53:31.128967 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 9 07:53:31.134017 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Oct 9 07:53:31.137188 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Oct 9 07:53:31.142921 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Oct 9 07:53:31.153065 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Oct 9 07:53:31.162107 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Oct 9 07:53:31.214446 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Oct 9 07:53:31.239613 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 9 07:53:31.250245 systemd-tmpfiles[1190]: ACLs are not supported, ignoring. Oct 9 07:53:31.250314 systemd-tmpfiles[1190]: ACLs are not supported, ignoring. Oct 9 07:53:31.265944 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 9 07:53:31.278196 systemd[1]: Starting systemd-sysusers.service - Create System Users... Oct 9 07:53:31.282221 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 9 07:53:31.296041 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Oct 9 07:53:31.333178 systemd[1]: Finished systemd-sysusers.service - Create System Users. Oct 9 07:53:31.343153 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 9 07:53:31.348356 udevadm[1209]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Oct 9 07:53:31.375535 systemd-tmpfiles[1215]: ACLs are not supported, ignoring. Oct 9 07:53:31.375560 systemd-tmpfiles[1215]: ACLs are not supported, ignoring. Oct 9 07:53:31.384958 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 9 07:53:32.017719 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Oct 9 07:53:32.026135 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 9 07:53:32.069718 systemd-udevd[1221]: Using default interface naming scheme 'v255'. Oct 9 07:53:32.092200 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 9 07:53:32.104008 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 9 07:53:32.137000 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Oct 9 07:53:32.194184 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1234) Oct 9 07:53:32.193473 systemd[1]: Started systemd-userdbd.service - User Database Manager. Oct 9 07:53:32.210729 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1234) Oct 9 07:53:32.237276 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:32.237468 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 9 07:53:32.243954 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 9 07:53:32.255946 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 9 07:53:32.266954 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 9 07:53:32.267687 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 9 07:53:32.267803 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 9 07:53:32.267873 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:32.272172 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 9 07:53:32.272396 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 9 07:53:32.282733 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Oct 9 07:53:32.286323 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 9 07:53:32.287171 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 9 07:53:32.292492 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 9 07:53:32.296123 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 9 07:53:32.303438 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 9 07:53:32.304915 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 9 07:53:32.307747 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1235) Oct 9 07:53:32.355436 systemd-networkd[1226]: lo: Link UP Oct 9 07:53:32.355450 systemd-networkd[1226]: lo: Gained carrier Oct 9 07:53:32.359295 systemd-networkd[1226]: Enumeration completed Oct 9 07:53:32.359894 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 9 07:53:32.359918 systemd-networkd[1226]: eth0: Configuring with /run/systemd/network/10-a2:f4:9a:c6:0f:b7.network. Oct 9 07:53:32.361166 systemd-networkd[1226]: eth1: Configuring with /run/systemd/network/10-5a:8f:ff:f8:50:e3.network. Oct 9 07:53:32.362119 systemd-networkd[1226]: eth0: Link UP Oct 9 07:53:32.362133 systemd-networkd[1226]: eth0: Gained carrier Oct 9 07:53:32.366343 systemd-networkd[1226]: eth1: Link UP Oct 9 07:53:32.366359 systemd-networkd[1226]: eth1: Gained carrier Oct 9 07:53:32.369008 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Oct 9 07:53:32.429660 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Oct 9 07:53:32.460401 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Oct 9 07:53:32.460934 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Oct 9 07:53:32.472874 kernel: ACPI: button: Power Button [PWRF] Oct 9 07:53:32.546891 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Oct 9 07:53:32.564003 kernel: mousedev: PS/2 mouse device common for all mice Oct 9 07:53:32.577363 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 9 07:53:32.600742 kernel: [drm] pci: virtio-vga detected at 0000:00:02.0 Oct 9 07:53:32.600873 kernel: virtio-pci 0000:00:02.0: vgaarb: deactivate vga console Oct 9 07:53:32.616751 kernel: Console: switching to colour dummy device 80x25 Oct 9 07:53:32.616878 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Oct 9 07:53:32.616905 kernel: [drm] features: -context_init Oct 9 07:53:32.616926 kernel: [drm] number of scanouts: 1 Oct 9 07:53:32.616995 kernel: [drm] number of cap sets: 0 Oct 9 07:53:32.617021 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:02.0 on minor 0 Oct 9 07:53:32.640748 kernel: fbcon: virtio_gpudrmfb (fb0) is primary device Oct 9 07:53:32.644736 kernel: Console: switching to colour frame buffer device 128x48 Oct 9 07:53:32.652740 kernel: virtio-pci 0000:00:02.0: [drm] fb0: virtio_gpudrmfb frame buffer device Oct 9 07:53:32.666309 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 9 07:53:32.666747 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 9 07:53:32.715673 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 9 07:53:32.804137 kernel: EDAC MC: Ver: 3.0.0 Oct 9 07:53:32.810657 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 9 07:53:32.842390 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Oct 9 07:53:32.855027 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Oct 9 07:53:32.875696 lvm[1284]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Oct 9 07:53:32.919849 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Oct 9 07:53:32.920858 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 9 07:53:32.927220 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Oct 9 07:53:32.952059 lvm[1287]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Oct 9 07:53:32.986899 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Oct 9 07:53:32.987638 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 9 07:53:33.006051 systemd[1]: Mounting media-configdrive.mount - /media/configdrive... Oct 9 07:53:33.006284 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Oct 9 07:53:33.006341 systemd[1]: Reached target machines.target - Containers. Oct 9 07:53:33.008977 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Oct 9 07:53:33.025769 kernel: ISO 9660 Extensions: RRIP_1991A Oct 9 07:53:33.028432 systemd[1]: Mounted media-configdrive.mount - /media/configdrive. Oct 9 07:53:33.033591 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 9 07:53:33.036149 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Oct 9 07:53:33.044154 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Oct 9 07:53:33.049079 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Oct 9 07:53:33.049500 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 9 07:53:33.059079 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Oct 9 07:53:33.066963 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Oct 9 07:53:33.068384 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Oct 9 07:53:33.074629 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Oct 9 07:53:33.108574 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Oct 9 07:53:33.116697 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Oct 9 07:53:33.120796 kernel: loop0: detected capacity change from 0 to 142488 Oct 9 07:53:33.166919 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Oct 9 07:53:33.185925 kernel: loop1: detected capacity change from 0 to 8 Oct 9 07:53:33.211807 kernel: loop2: detected capacity change from 0 to 140768 Oct 9 07:53:33.250556 kernel: loop3: detected capacity change from 0 to 211296 Oct 9 07:53:33.306756 kernel: loop4: detected capacity change from 0 to 142488 Oct 9 07:53:33.329633 kernel: loop5: detected capacity change from 0 to 8 Oct 9 07:53:33.332742 kernel: loop6: detected capacity change from 0 to 140768 Oct 9 07:53:33.359234 kernel: loop7: detected capacity change from 0 to 211296 Oct 9 07:53:33.374496 (sd-merge)[1316]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-digitalocean'. Oct 9 07:53:33.375409 (sd-merge)[1316]: Merged extensions into '/usr'. Oct 9 07:53:33.383820 systemd[1]: Reloading requested from client PID 1304 ('systemd-sysext') (unit systemd-sysext.service)... Oct 9 07:53:33.383849 systemd[1]: Reloading... Oct 9 07:53:33.477152 systemd-networkd[1226]: eth0: Gained IPv6LL Oct 9 07:53:33.494964 zram_generator::config[1345]: No configuration found. Oct 9 07:53:33.645826 ldconfig[1301]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Oct 9 07:53:33.736830 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 9 07:53:33.815527 systemd[1]: Reloading finished in 430 ms. Oct 9 07:53:33.836890 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Oct 9 07:53:33.839075 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Oct 9 07:53:33.840840 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Oct 9 07:53:33.853132 systemd[1]: Starting ensure-sysext.service... Oct 9 07:53:33.857510 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 9 07:53:33.880076 systemd[1]: Reloading requested from client PID 1397 ('systemctl') (unit ensure-sysext.service)... Oct 9 07:53:33.880124 systemd[1]: Reloading... Oct 9 07:53:33.920466 systemd-tmpfiles[1398]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Oct 9 07:53:33.923028 systemd-tmpfiles[1398]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Oct 9 07:53:33.924935 systemd-tmpfiles[1398]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Oct 9 07:53:33.925660 systemd-tmpfiles[1398]: ACLs are not supported, ignoring. Oct 9 07:53:33.926018 systemd-tmpfiles[1398]: ACLs are not supported, ignoring. Oct 9 07:53:33.931539 systemd-tmpfiles[1398]: Detected autofs mount point /boot during canonicalization of boot. Oct 9 07:53:33.931985 systemd-tmpfiles[1398]: Skipping /boot Oct 9 07:53:33.954507 systemd-tmpfiles[1398]: Detected autofs mount point /boot during canonicalization of boot. Oct 9 07:53:33.954767 systemd-tmpfiles[1398]: Skipping /boot Oct 9 07:53:33.988761 zram_generator::config[1424]: No configuration found. Oct 9 07:53:34.192465 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 9 07:53:34.269549 systemd[1]: Reloading finished in 388 ms. Oct 9 07:53:34.288822 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 9 07:53:34.314050 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Oct 9 07:53:34.328219 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Oct 9 07:53:34.333866 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Oct 9 07:53:34.350519 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 9 07:53:34.369119 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Oct 9 07:53:34.392145 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:34.392799 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 9 07:53:34.401143 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 9 07:53:34.421362 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 9 07:53:34.438024 systemd-networkd[1226]: eth1: Gained IPv6LL Oct 9 07:53:34.439732 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 9 07:53:34.448674 augenrules[1501]: No rules Oct 9 07:53:34.445595 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 9 07:53:34.447096 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:34.462009 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Oct 9 07:53:34.472537 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Oct 9 07:53:34.478329 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 9 07:53:34.478550 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 9 07:53:34.479649 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 9 07:53:34.482399 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 9 07:53:34.486394 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 9 07:53:34.488657 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 9 07:53:34.521997 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Oct 9 07:53:34.530362 systemd[1]: Finished ensure-sysext.service. Oct 9 07:53:34.535039 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Oct 9 07:53:34.542075 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:34.542426 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 9 07:53:34.550191 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 9 07:53:34.563773 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 9 07:53:34.572222 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 9 07:53:34.584104 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 9 07:53:34.589000 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 9 07:53:34.600007 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Oct 9 07:53:34.624030 systemd[1]: Starting systemd-update-done.service - Update is Completed... Oct 9 07:53:34.624758 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 9 07:53:34.624817 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 9 07:53:34.627569 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 9 07:53:34.627876 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 9 07:53:34.630444 systemd-resolved[1487]: Positive Trust Anchors: Oct 9 07:53:34.630465 systemd-resolved[1487]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 9 07:53:34.630524 systemd-resolved[1487]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 9 07:53:34.637168 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 9 07:53:34.637419 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 9 07:53:34.640453 systemd-resolved[1487]: Using system hostname 'ci-4081.1.0-c-f399a429fa'. Oct 9 07:53:34.642514 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 9 07:53:34.643083 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 9 07:53:34.649826 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 9 07:53:34.651377 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 9 07:53:34.653003 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 9 07:53:34.665691 systemd[1]: Reached target network.target - Network. Oct 9 07:53:34.667834 systemd[1]: Reached target network-online.target - Network is Online. Oct 9 07:53:34.668251 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 9 07:53:34.668662 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 9 07:53:34.670256 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 9 07:53:34.671664 systemd[1]: Finished systemd-update-done.service - Update is Completed. Oct 9 07:53:34.742110 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Oct 9 07:53:34.743278 systemd[1]: Reached target sysinit.target - System Initialization. Oct 9 07:53:34.744899 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Oct 9 07:53:34.745651 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Oct 9 07:53:34.746591 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Oct 9 07:53:34.747414 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Oct 9 07:53:34.747464 systemd[1]: Reached target paths.target - Path Units. Oct 9 07:53:34.748555 systemd[1]: Reached target time-set.target - System Time Set. Oct 9 07:53:34.749394 systemd[1]: Started logrotate.timer - Daily rotation of log files. Oct 9 07:53:34.750177 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Oct 9 07:53:34.750774 systemd[1]: Reached target timers.target - Timer Units. Oct 9 07:53:34.752156 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Oct 9 07:53:34.755067 systemd[1]: Starting docker.socket - Docker Socket for the API... Oct 9 07:53:34.760371 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Oct 9 07:53:34.764879 systemd[1]: Listening on docker.socket - Docker Socket for the API. Oct 9 07:53:34.766020 systemd[1]: Reached target sockets.target - Socket Units. Oct 9 07:53:34.766477 systemd[1]: Reached target basic.target - Basic System. Oct 9 07:53:34.769039 systemd[1]: System is tainted: cgroupsv1 Oct 9 07:53:34.769129 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Oct 9 07:53:34.769162 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Oct 9 07:53:34.777049 systemd[1]: Starting containerd.service - containerd container runtime... Oct 9 07:53:34.782978 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Oct 9 07:53:34.788992 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Oct 9 07:53:34.795920 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Oct 9 07:53:34.807273 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Oct 9 07:53:34.809255 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Oct 9 07:53:34.820976 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 9 07:53:34.823064 systemd-timesyncd[1529]: Contacted time server 72.30.35.88:123 (0.flatcar.pool.ntp.org). Oct 9 07:53:34.823142 systemd-timesyncd[1529]: Initial clock synchronization to Wed 2024-10-09 07:53:35.186308 UTC. Oct 9 07:53:34.835074 jq[1547]: false Oct 9 07:53:34.835156 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Oct 9 07:53:34.857083 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Oct 9 07:53:34.864252 coreos-metadata[1545]: Oct 09 07:53:34.863 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 9 07:53:34.868495 extend-filesystems[1550]: Found loop4 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found loop5 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found loop6 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found loop7 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda1 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda2 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda3 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found usr Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda4 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda6 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda7 Oct 9 07:53:34.879670 extend-filesystems[1550]: Found vda9 Oct 9 07:53:34.879670 extend-filesystems[1550]: Checking size of /dev/vda9 Oct 9 07:53:34.874915 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Oct 9 07:53:34.915028 coreos-metadata[1545]: Oct 09 07:53:34.876 INFO Fetch successful Oct 9 07:53:34.910086 dbus-daemon[1546]: [system] SELinux support is enabled Oct 9 07:53:34.894958 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Oct 9 07:53:34.924147 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Oct 9 07:53:34.950983 systemd[1]: Starting systemd-logind.service - User Login Management... Oct 9 07:53:34.964006 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Oct 9 07:53:34.973330 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1224) Oct 9 07:53:34.983638 systemd[1]: Starting update-engine.service - Update Engine... Oct 9 07:53:34.988374 extend-filesystems[1550]: Resized partition /dev/vda9 Oct 9 07:53:35.006289 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Oct 9 07:53:35.013239 extend-filesystems[1582]: resize2fs 1.47.1 (20-May-2024) Oct 9 07:53:35.024108 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Oct 9 07:53:35.030036 systemd[1]: Started dbus.service - D-Bus System Message Bus. Oct 9 07:53:35.045870 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Oct 9 07:53:35.046232 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Oct 9 07:53:35.050048 systemd[1]: motdgen.service: Deactivated successfully. Oct 9 07:53:35.050374 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Oct 9 07:53:35.060313 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Oct 9 07:53:35.077853 jq[1581]: true Oct 9 07:53:35.080378 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Oct 9 07:53:35.088206 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Oct 9 07:53:35.156162 update_engine[1577]: I20241009 07:53:35.155126 1577 main.cc:92] Flatcar Update Engine starting Oct 9 07:53:35.182917 update_engine[1577]: I20241009 07:53:35.177722 1577 update_check_scheduler.cc:74] Next update check in 4m53s Oct 9 07:53:35.186493 (ntainerd)[1597]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Oct 9 07:53:35.202002 jq[1595]: true Oct 9 07:53:35.247698 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Oct 9 07:53:35.272371 tar[1591]: linux-amd64/helm Oct 9 07:53:35.289059 systemd[1]: Started update-engine.service - Update Engine. Oct 9 07:53:35.293934 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Oct 9 07:53:35.294148 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Oct 9 07:53:35.294218 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Oct 9 07:53:35.295185 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Oct 9 07:53:35.295330 systemd[1]: user-configdrive.service - Load cloud-config from /media/configdrive was skipped because of an unmet condition check (ConditionKernelCommandLine=!flatcar.oem.id=digitalocean). Oct 9 07:53:35.295373 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Oct 9 07:53:35.299724 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Oct 9 07:53:35.307149 systemd[1]: Started locksmithd.service - Cluster reboot manager. Oct 9 07:53:35.339199 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Oct 9 07:53:35.357772 extend-filesystems[1582]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Oct 9 07:53:35.357772 extend-filesystems[1582]: old_desc_blocks = 1, new_desc_blocks = 8 Oct 9 07:53:35.357772 extend-filesystems[1582]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Oct 9 07:53:35.365508 extend-filesystems[1550]: Resized filesystem in /dev/vda9 Oct 9 07:53:35.365508 extend-filesystems[1550]: Found vdb Oct 9 07:53:35.362463 systemd[1]: extend-filesystems.service: Deactivated successfully. Oct 9 07:53:35.365559 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Oct 9 07:53:35.395440 systemd-logind[1572]: New seat seat0. Oct 9 07:53:35.415915 systemd-logind[1572]: Watching system buttons on /dev/input/event1 (Power Button) Oct 9 07:53:35.415952 systemd-logind[1572]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Oct 9 07:53:35.416380 systemd[1]: Started systemd-logind.service - User Login Management. Oct 9 07:53:35.480705 bash[1636]: Updated "/home/core/.ssh/authorized_keys" Oct 9 07:53:35.470527 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Oct 9 07:53:35.505490 systemd[1]: Starting sshkeys.service... Oct 9 07:53:35.566151 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Oct 9 07:53:35.584336 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Oct 9 07:53:35.640164 coreos-metadata[1656]: Oct 09 07:53:35.639 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Oct 9 07:53:35.645364 locksmithd[1617]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Oct 9 07:53:35.655713 coreos-metadata[1656]: Oct 09 07:53:35.654 INFO Fetch successful Oct 9 07:53:35.669593 unknown[1656]: wrote ssh authorized keys file for user: core Oct 9 07:53:35.729140 update-ssh-keys[1663]: Updated "/home/core/.ssh/authorized_keys" Oct 9 07:53:35.733352 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Oct 9 07:53:35.747166 systemd[1]: Finished sshkeys.service. Oct 9 07:53:35.930911 containerd[1597]: time="2024-10-09T07:53:35.930305483Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Oct 9 07:53:35.999072 containerd[1597]: time="2024-10-09T07:53:35.998736044Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009095 containerd[1597]: time="2024-10-09T07:53:36.008599293Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.54-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009095 containerd[1597]: time="2024-10-09T07:53:36.008650383Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Oct 9 07:53:36.009095 containerd[1597]: time="2024-10-09T07:53:36.008695205Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Oct 9 07:53:36.009095 containerd[1597]: time="2024-10-09T07:53:36.008908776Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Oct 9 07:53:36.009095 containerd[1597]: time="2024-10-09T07:53:36.008933138Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009095 containerd[1597]: time="2024-10-09T07:53:36.008999737Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009095 containerd[1597]: time="2024-10-09T07:53:36.009013399Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009674 containerd[1597]: time="2024-10-09T07:53:36.009645383Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009782 containerd[1597]: time="2024-10-09T07:53:36.009731814Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009843 containerd[1597]: time="2024-10-09T07:53:36.009829171Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Oct 9 07:53:36.009906 containerd[1597]: time="2024-10-09T07:53:36.009893321Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Oct 9 07:53:36.010372 containerd[1597]: time="2024-10-09T07:53:36.010056965Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Oct 9 07:53:36.010372 containerd[1597]: time="2024-10-09T07:53:36.010333150Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Oct 9 07:53:36.010656 containerd[1597]: time="2024-10-09T07:53:36.010637245Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Oct 9 07:53:36.010823 containerd[1597]: time="2024-10-09T07:53:36.010807130Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Oct 9 07:53:36.010972 containerd[1597]: time="2024-10-09T07:53:36.010956654Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Oct 9 07:53:36.011203 containerd[1597]: time="2024-10-09T07:53:36.011185675Z" level=info msg="metadata content store policy set" policy=shared Oct 9 07:53:36.014582 containerd[1597]: time="2024-10-09T07:53:36.014536749Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.014756004Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.014789162Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.014846000Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.014862678Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015078369Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015456935Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015614646Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015632990Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015648675Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015663107Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015676612Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015689434Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015706204Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.016783 containerd[1597]: time="2024-10-09T07:53:36.015721814Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015736025Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015776370Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015828925Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015853482Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015868892Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015883454Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015908724Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015922967Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015937075Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015952209Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015965690Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015982954Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.015998021Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017186 containerd[1597]: time="2024-10-09T07:53:36.016011331Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016024409Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016042913Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016064356Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016088439Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016120338Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016142284Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016194214Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016212994Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016225213Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016236611Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016246847Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016259898Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016271711Z" level=info msg="NRI interface is disabled by configuration." Oct 9 07:53:36.017499 containerd[1597]: time="2024-10-09T07:53:36.016283295Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Oct 9 07:53:36.017809 containerd[1597]: time="2024-10-09T07:53:36.016616955Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Oct 9 07:53:36.017809 containerd[1597]: time="2024-10-09T07:53:36.016676273Z" level=info msg="Connect containerd service" Oct 9 07:53:36.017809 containerd[1597]: time="2024-10-09T07:53:36.016726036Z" level=info msg="using legacy CRI server" Oct 9 07:53:36.017809 containerd[1597]: time="2024-10-09T07:53:36.016734777Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Oct 9 07:53:36.018422 containerd[1597]: time="2024-10-09T07:53:36.018396286Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Oct 9 07:53:36.019287 containerd[1597]: time="2024-10-09T07:53:36.019258006Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 9 07:53:36.019843 containerd[1597]: time="2024-10-09T07:53:36.019820742Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Oct 9 07:53:36.019964 containerd[1597]: time="2024-10-09T07:53:36.019951624Z" level=info msg=serving... address=/run/containerd/containerd.sock Oct 9 07:53:36.020106 containerd[1597]: time="2024-10-09T07:53:36.020075906Z" level=info msg="Start subscribing containerd event" Oct 9 07:53:36.020171 containerd[1597]: time="2024-10-09T07:53:36.020162233Z" level=info msg="Start recovering state" Oct 9 07:53:36.020403 containerd[1597]: time="2024-10-09T07:53:36.020386753Z" level=info msg="Start event monitor" Oct 9 07:53:36.020602 containerd[1597]: time="2024-10-09T07:53:36.020586442Z" level=info msg="Start snapshots syncer" Oct 9 07:53:36.020655 containerd[1597]: time="2024-10-09T07:53:36.020646214Z" level=info msg="Start cni network conf syncer for default" Oct 9 07:53:36.020804 containerd[1597]: time="2024-10-09T07:53:36.020790939Z" level=info msg="Start streaming server" Oct 9 07:53:36.021913 containerd[1597]: time="2024-10-09T07:53:36.021017083Z" level=info msg="containerd successfully booted in 0.095645s" Oct 9 07:53:36.021220 systemd[1]: Started containerd.service - containerd container runtime. Oct 9 07:53:36.225519 sshd_keygen[1589]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Oct 9 07:53:36.263549 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Oct 9 07:53:36.277239 systemd[1]: Starting issuegen.service - Generate /run/issue... Oct 9 07:53:36.314449 systemd[1]: issuegen.service: Deactivated successfully. Oct 9 07:53:36.315045 systemd[1]: Finished issuegen.service - Generate /run/issue. Oct 9 07:53:36.327220 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Oct 9 07:53:36.365820 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Oct 9 07:53:36.381304 systemd[1]: Started getty@tty1.service - Getty on tty1. Oct 9 07:53:36.391312 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Oct 9 07:53:36.393312 systemd[1]: Reached target getty.target - Login Prompts. Oct 9 07:53:36.495160 tar[1591]: linux-amd64/LICENSE Oct 9 07:53:36.495624 tar[1591]: linux-amd64/README.md Oct 9 07:53:36.512716 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Oct 9 07:53:36.854382 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 9 07:53:36.858862 (kubelet)[1707]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 9 07:53:36.859972 systemd[1]: Reached target multi-user.target - Multi-User System. Oct 9 07:53:36.863726 systemd[1]: Startup finished in 7.307s (kernel) + 6.963s (userspace) = 14.270s. Oct 9 07:53:37.030415 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Oct 9 07:53:37.040092 systemd[1]: Started sshd@0-143.198.138.163:22-139.178.89.65:40394.service - OpenSSH per-connection server daemon (139.178.89.65:40394). Oct 9 07:53:37.130421 sshd[1712]: Accepted publickey for core from 139.178.89.65 port 40394 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:53:37.133586 sshd[1712]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:53:37.152645 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Oct 9 07:53:37.152776 systemd-logind[1572]: New session 1 of user core. Oct 9 07:53:37.162304 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Oct 9 07:53:37.195114 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Oct 9 07:53:37.211412 systemd[1]: Starting user@500.service - User Manager for UID 500... Oct 9 07:53:37.235118 (systemd)[1722]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Oct 9 07:53:37.360095 systemd[1722]: Queued start job for default target default.target. Oct 9 07:53:37.360591 systemd[1722]: Created slice app.slice - User Application Slice. Oct 9 07:53:37.360623 systemd[1722]: Reached target paths.target - Paths. Oct 9 07:53:37.360643 systemd[1722]: Reached target timers.target - Timers. Oct 9 07:53:37.368986 systemd[1722]: Starting dbus.socket - D-Bus User Message Bus Socket... Oct 9 07:53:37.380974 systemd[1722]: Listening on dbus.socket - D-Bus User Message Bus Socket. Oct 9 07:53:37.384451 systemd[1722]: Reached target sockets.target - Sockets. Oct 9 07:53:37.384495 systemd[1722]: Reached target basic.target - Basic System. Oct 9 07:53:37.384582 systemd[1722]: Reached target default.target - Main User Target. Oct 9 07:53:37.384617 systemd[1722]: Startup finished in 139ms. Oct 9 07:53:37.386152 systemd[1]: Started user@500.service - User Manager for UID 500. Oct 9 07:53:37.392236 systemd[1]: Started session-1.scope - Session 1 of User core. Oct 9 07:53:37.468221 systemd[1]: Started sshd@1-143.198.138.163:22-139.178.89.65:40398.service - OpenSSH per-connection server daemon (139.178.89.65:40398). Oct 9 07:53:37.558167 sshd[1734]: Accepted publickey for core from 139.178.89.65 port 40398 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:53:37.561113 sshd[1734]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:53:37.569113 systemd-logind[1572]: New session 2 of user core. Oct 9 07:53:37.580297 systemd[1]: Started session-2.scope - Session 2 of User core. Oct 9 07:53:37.665364 sshd[1734]: pam_unix(sshd:session): session closed for user core Oct 9 07:53:37.680391 systemd[1]: Started sshd@2-143.198.138.163:22-139.178.89.65:40412.service - OpenSSH per-connection server daemon (139.178.89.65:40412). Oct 9 07:53:37.681119 systemd[1]: sshd@1-143.198.138.163:22-139.178.89.65:40398.service: Deactivated successfully. Oct 9 07:53:37.693397 systemd-logind[1572]: Session 2 logged out. Waiting for processes to exit. Oct 9 07:53:37.694975 systemd[1]: session-2.scope: Deactivated successfully. Oct 9 07:53:37.698283 systemd-logind[1572]: Removed session 2. Oct 9 07:53:37.766133 sshd[1741]: Accepted publickey for core from 139.178.89.65 port 40412 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:53:37.769261 sshd[1741]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:53:37.779052 systemd-logind[1572]: New session 3 of user core. Oct 9 07:53:37.785427 systemd[1]: Started session-3.scope - Session 3 of User core. Oct 9 07:53:37.836030 kubelet[1707]: E1009 07:53:37.835854 1707 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 9 07:53:37.841321 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 9 07:53:37.841660 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 9 07:53:37.853990 sshd[1741]: pam_unix(sshd:session): session closed for user core Oct 9 07:53:37.863581 systemd[1]: Started sshd@3-143.198.138.163:22-139.178.89.65:40424.service - OpenSSH per-connection server daemon (139.178.89.65:40424). Oct 9 07:53:37.864334 systemd[1]: sshd@2-143.198.138.163:22-139.178.89.65:40412.service: Deactivated successfully. Oct 9 07:53:37.870158 systemd[1]: session-3.scope: Deactivated successfully. Oct 9 07:53:37.870513 systemd-logind[1572]: Session 3 logged out. Waiting for processes to exit. Oct 9 07:53:37.877924 systemd-logind[1572]: Removed session 3. Oct 9 07:53:37.927408 sshd[1751]: Accepted publickey for core from 139.178.89.65 port 40424 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:53:37.929529 sshd[1751]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:53:37.936083 systemd-logind[1572]: New session 4 of user core. Oct 9 07:53:37.943393 systemd[1]: Started session-4.scope - Session 4 of User core. Oct 9 07:53:38.012057 sshd[1751]: pam_unix(sshd:session): session closed for user core Oct 9 07:53:38.021595 systemd[1]: Started sshd@4-143.198.138.163:22-139.178.89.65:40426.service - OpenSSH per-connection server daemon (139.178.89.65:40426). Oct 9 07:53:38.022459 systemd[1]: sshd@3-143.198.138.163:22-139.178.89.65:40424.service: Deactivated successfully. Oct 9 07:53:38.025684 systemd[1]: session-4.scope: Deactivated successfully. Oct 9 07:53:38.037015 systemd-logind[1572]: Session 4 logged out. Waiting for processes to exit. Oct 9 07:53:38.040044 systemd-logind[1572]: Removed session 4. Oct 9 07:53:38.087586 sshd[1759]: Accepted publickey for core from 139.178.89.65 port 40426 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:53:38.090515 sshd[1759]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:53:38.099867 systemd-logind[1572]: New session 5 of user core. Oct 9 07:53:38.102297 systemd[1]: Started session-5.scope - Session 5 of User core. Oct 9 07:53:38.180533 sudo[1766]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Oct 9 07:53:38.181005 sudo[1766]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 9 07:53:38.199089 sudo[1766]: pam_unix(sudo:session): session closed for user root Oct 9 07:53:38.203857 sshd[1759]: pam_unix(sshd:session): session closed for user core Oct 9 07:53:38.220418 systemd[1]: Started sshd@5-143.198.138.163:22-139.178.89.65:40432.service - OpenSSH per-connection server daemon (139.178.89.65:40432). Oct 9 07:53:38.221295 systemd[1]: sshd@4-143.198.138.163:22-139.178.89.65:40426.service: Deactivated successfully. Oct 9 07:53:38.225689 systemd[1]: session-5.scope: Deactivated successfully. Oct 9 07:53:38.229802 systemd-logind[1572]: Session 5 logged out. Waiting for processes to exit. Oct 9 07:53:38.231595 systemd-logind[1572]: Removed session 5. Oct 9 07:53:38.275585 sshd[1769]: Accepted publickey for core from 139.178.89.65 port 40432 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:53:38.278874 sshd[1769]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:53:38.286989 systemd-logind[1572]: New session 6 of user core. Oct 9 07:53:38.292572 systemd[1]: Started session-6.scope - Session 6 of User core. Oct 9 07:53:38.359637 sudo[1776]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Oct 9 07:53:38.360778 sudo[1776]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 9 07:53:38.366762 sudo[1776]: pam_unix(sudo:session): session closed for user root Oct 9 07:53:38.374842 sudo[1775]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Oct 9 07:53:38.375343 sudo[1775]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 9 07:53:38.393239 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Oct 9 07:53:38.409714 auditctl[1779]: No rules Oct 9 07:53:38.410283 systemd[1]: audit-rules.service: Deactivated successfully. Oct 9 07:53:38.410722 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Oct 9 07:53:38.427818 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Oct 9 07:53:38.469189 augenrules[1798]: No rules Oct 9 07:53:38.472603 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Oct 9 07:53:38.477179 sudo[1775]: pam_unix(sudo:session): session closed for user root Oct 9 07:53:38.482073 sshd[1769]: pam_unix(sshd:session): session closed for user core Oct 9 07:53:38.493430 systemd[1]: Started sshd@6-143.198.138.163:22-139.178.89.65:40434.service - OpenSSH per-connection server daemon (139.178.89.65:40434). Oct 9 07:53:38.494271 systemd[1]: sshd@5-143.198.138.163:22-139.178.89.65:40432.service: Deactivated successfully. Oct 9 07:53:38.502214 systemd[1]: session-6.scope: Deactivated successfully. Oct 9 07:53:38.505577 systemd-logind[1572]: Session 6 logged out. Waiting for processes to exit. Oct 9 07:53:38.507604 systemd-logind[1572]: Removed session 6. Oct 9 07:53:38.550820 sshd[1804]: Accepted publickey for core from 139.178.89.65 port 40434 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:53:38.553175 sshd[1804]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:53:38.563238 systemd-logind[1572]: New session 7 of user core. Oct 9 07:53:38.570446 systemd[1]: Started session-7.scope - Session 7 of User core. Oct 9 07:53:38.641464 sudo[1811]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Oct 9 07:53:38.641971 sudo[1811]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 9 07:53:39.157551 systemd[1]: Starting docker.service - Docker Application Container Engine... Oct 9 07:53:39.157767 (dockerd)[1827]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Oct 9 07:53:39.654442 dockerd[1827]: time="2024-10-09T07:53:39.654281866Z" level=info msg="Starting up" Oct 9 07:53:39.837572 systemd[1]: var-lib-docker-metacopy\x2dcheck2696664957-merged.mount: Deactivated successfully. Oct 9 07:53:39.867592 dockerd[1827]: time="2024-10-09T07:53:39.867430464Z" level=info msg="Loading containers: start." Oct 9 07:53:40.019869 kernel: Initializing XFRM netlink socket Oct 9 07:53:40.127858 systemd-networkd[1226]: docker0: Link UP Oct 9 07:53:40.149864 dockerd[1827]: time="2024-10-09T07:53:40.149812080Z" level=info msg="Loading containers: done." Oct 9 07:53:40.173505 dockerd[1827]: time="2024-10-09T07:53:40.173428539Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Oct 9 07:53:40.173812 dockerd[1827]: time="2024-10-09T07:53:40.173596335Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Oct 9 07:53:40.173812 dockerd[1827]: time="2024-10-09T07:53:40.173725990Z" level=info msg="Daemon has completed initialization" Oct 9 07:53:40.221144 dockerd[1827]: time="2024-10-09T07:53:40.221018610Z" level=info msg="API listen on /run/docker.sock" Oct 9 07:53:40.222046 systemd[1]: Started docker.service - Docker Application Container Engine. Oct 9 07:53:41.213432 containerd[1597]: time="2024-10-09T07:53:41.213240230Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.9\"" Oct 9 07:53:41.818393 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1398933447.mount: Deactivated successfully. Oct 9 07:53:43.380772 containerd[1597]: time="2024-10-09T07:53:43.379598859Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.29.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:43.382425 containerd[1597]: time="2024-10-09T07:53:43.382371717Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.29.9: active requests=0, bytes read=35213841" Oct 9 07:53:43.383626 containerd[1597]: time="2024-10-09T07:53:43.383578542Z" level=info msg="ImageCreate event name:\"sha256:bc1ec5c2b6c60a3b18e7f54a99f0452c038400ecaaa2576931fd5342a0586abb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:43.387060 containerd[1597]: time="2024-10-09T07:53:43.387006932Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:b88538e7fdf73583c8670540eec5b3620af75c9ec200434a5815ee7fba5021f3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:43.392819 containerd[1597]: time="2024-10-09T07:53:43.392680719Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.29.9\" with image id \"sha256:bc1ec5c2b6c60a3b18e7f54a99f0452c038400ecaaa2576931fd5342a0586abb\", repo tag \"registry.k8s.io/kube-apiserver:v1.29.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:b88538e7fdf73583c8670540eec5b3620af75c9ec200434a5815ee7fba5021f3\", size \"35210641\" in 2.17935963s" Oct 9 07:53:43.393028 containerd[1597]: time="2024-10-09T07:53:43.393011373Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.9\" returns image reference \"sha256:bc1ec5c2b6c60a3b18e7f54a99f0452c038400ecaaa2576931fd5342a0586abb\"" Oct 9 07:53:43.426644 containerd[1597]: time="2024-10-09T07:53:43.426585794Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.9\"" Oct 9 07:53:45.291459 containerd[1597]: time="2024-10-09T07:53:45.291374769Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.29.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:45.292808 containerd[1597]: time="2024-10-09T07:53:45.292751072Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.29.9: active requests=0, bytes read=32208673" Oct 9 07:53:45.293615 containerd[1597]: time="2024-10-09T07:53:45.293549475Z" level=info msg="ImageCreate event name:\"sha256:5abda0d0a9153cd1f90fd828be379f7a16a6c814e6efbbbf31e247e13c3843e5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:45.297077 containerd[1597]: time="2024-10-09T07:53:45.296933452Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:f2f18973ccb6996687d10ba5bd1b8f303e3dd2fed80f831a44d2ac8191e5bb9b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:45.298485 containerd[1597]: time="2024-10-09T07:53:45.298291573Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.29.9\" with image id \"sha256:5abda0d0a9153cd1f90fd828be379f7a16a6c814e6efbbbf31e247e13c3843e5\", repo tag \"registry.k8s.io/kube-controller-manager:v1.29.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:f2f18973ccb6996687d10ba5bd1b8f303e3dd2fed80f831a44d2ac8191e5bb9b\", size \"33739229\" in 1.871662678s" Oct 9 07:53:45.298485 containerd[1597]: time="2024-10-09T07:53:45.298336734Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.9\" returns image reference \"sha256:5abda0d0a9153cd1f90fd828be379f7a16a6c814e6efbbbf31e247e13c3843e5\"" Oct 9 07:53:45.338295 containerd[1597]: time="2024-10-09T07:53:45.338244351Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.9\"" Oct 9 07:53:46.542763 containerd[1597]: time="2024-10-09T07:53:46.542268164Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.29.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:46.544381 containerd[1597]: time="2024-10-09T07:53:46.544051840Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.29.9: active requests=0, bytes read=17320456" Oct 9 07:53:46.545185 containerd[1597]: time="2024-10-09T07:53:46.544988142Z" level=info msg="ImageCreate event name:\"sha256:059957505b3370d4c57d793e79cc70f9063d7ab75767f7040f5cc85572fe7e8d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:46.548770 containerd[1597]: time="2024-10-09T07:53:46.548217882Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:9c164076eebaefdaebad46a5ccd550e9f38c63588c02d35163c6a09e164ab8a8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:46.550104 containerd[1597]: time="2024-10-09T07:53:46.549640672Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.29.9\" with image id \"sha256:059957505b3370d4c57d793e79cc70f9063d7ab75767f7040f5cc85572fe7e8d\", repo tag \"registry.k8s.io/kube-scheduler:v1.29.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:9c164076eebaefdaebad46a5ccd550e9f38c63588c02d35163c6a09e164ab8a8\", size \"18851030\" in 1.211106334s" Oct 9 07:53:46.550104 containerd[1597]: time="2024-10-09T07:53:46.549686893Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.9\" returns image reference \"sha256:059957505b3370d4c57d793e79cc70f9063d7ab75767f7040f5cc85572fe7e8d\"" Oct 9 07:53:46.582592 containerd[1597]: time="2024-10-09T07:53:46.582306022Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.9\"" Oct 9 07:53:47.750527 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1659709330.mount: Deactivated successfully. Oct 9 07:53:48.092816 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Oct 9 07:53:48.101027 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 9 07:53:48.299450 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 9 07:53:48.311436 (kubelet)[2071]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 9 07:53:48.350500 containerd[1597]: time="2024-10-09T07:53:48.350327959Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.29.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:48.352775 containerd[1597]: time="2024-10-09T07:53:48.352695444Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.29.9: active requests=0, bytes read=28601750" Oct 9 07:53:48.355344 containerd[1597]: time="2024-10-09T07:53:48.354853460Z" level=info msg="ImageCreate event name:\"sha256:dd650d127e51776919ec1622a4469a8b141b2dfee5a33fbc5cb9729372e0dcfa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:48.359402 containerd[1597]: time="2024-10-09T07:53:48.359344768Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:124040dbe6b5294352355f5d34c692ecbc940cdc57a8fd06d0f38f76b6138906\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:48.361367 containerd[1597]: time="2024-10-09T07:53:48.360786234Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.29.9\" with image id \"sha256:dd650d127e51776919ec1622a4469a8b141b2dfee5a33fbc5cb9729372e0dcfa\", repo tag \"registry.k8s.io/kube-proxy:v1.29.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:124040dbe6b5294352355f5d34c692ecbc940cdc57a8fd06d0f38f76b6138906\", size \"28600769\" in 1.778428852s" Oct 9 07:53:48.361367 containerd[1597]: time="2024-10-09T07:53:48.360866895Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.9\" returns image reference \"sha256:dd650d127e51776919ec1622a4469a8b141b2dfee5a33fbc5cb9729372e0dcfa\"" Oct 9 07:53:48.400270 containerd[1597]: time="2024-10-09T07:53:48.400169415Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Oct 9 07:53:48.402201 systemd-resolved[1487]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.3. Oct 9 07:53:48.407822 kubelet[2071]: E1009 07:53:48.407659 2071 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 9 07:53:48.413582 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 9 07:53:48.415373 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 9 07:53:48.933137 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3591456671.mount: Deactivated successfully. Oct 9 07:53:49.843733 containerd[1597]: time="2024-10-09T07:53:49.843633912Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:49.845340 containerd[1597]: time="2024-10-09T07:53:49.844946274Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Oct 9 07:53:49.846795 containerd[1597]: time="2024-10-09T07:53:49.846104966Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:49.853414 containerd[1597]: time="2024-10-09T07:53:49.853335905Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:49.858140 containerd[1597]: time="2024-10-09T07:53:49.858063399Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.457572836s" Oct 9 07:53:49.858408 containerd[1597]: time="2024-10-09T07:53:49.858380647Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Oct 9 07:53:49.895656 containerd[1597]: time="2024-10-09T07:53:49.895603872Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Oct 9 07:53:50.416315 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3044198465.mount: Deactivated successfully. Oct 9 07:53:50.422741 containerd[1597]: time="2024-10-09T07:53:50.421328763Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:50.422741 containerd[1597]: time="2024-10-09T07:53:50.422237186Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322290" Oct 9 07:53:50.422741 containerd[1597]: time="2024-10-09T07:53:50.422243398Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:50.424645 containerd[1597]: time="2024-10-09T07:53:50.424582873Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:50.426215 containerd[1597]: time="2024-10-09T07:53:50.425629054Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 529.691727ms" Oct 9 07:53:50.426215 containerd[1597]: time="2024-10-09T07:53:50.425669156Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Oct 9 07:53:50.455958 containerd[1597]: time="2024-10-09T07:53:50.455918441Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\"" Oct 9 07:53:51.012364 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1877108096.mount: Deactivated successfully. Oct 9 07:53:51.461993 systemd-resolved[1487]: Using degraded feature set UDP instead of UDP+EDNS0 for DNS server 67.207.67.2. Oct 9 07:53:52.768228 containerd[1597]: time="2024-10-09T07:53:52.768124533Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.10-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:52.769828 containerd[1597]: time="2024-10-09T07:53:52.769755010Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.10-0: active requests=0, bytes read=56651625" Oct 9 07:53:52.772809 containerd[1597]: time="2024-10-09T07:53:52.772749438Z" level=info msg="ImageCreate event name:\"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:52.775702 containerd[1597]: time="2024-10-09T07:53:52.775638180Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:53:52.778204 containerd[1597]: time="2024-10-09T07:53:52.778153623Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.10-0\" with image id \"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\", repo tag \"registry.k8s.io/etcd:3.5.10-0\", repo digest \"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\", size \"56649232\" in 2.322193241s" Oct 9 07:53:52.778204 containerd[1597]: time="2024-10-09T07:53:52.778211972Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\" returns image reference \"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\"" Oct 9 07:53:55.367898 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 9 07:53:55.378077 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 9 07:53:55.412821 systemd[1]: Reloading requested from client PID 2256 ('systemctl') (unit session-7.scope)... Oct 9 07:53:55.412853 systemd[1]: Reloading... Oct 9 07:53:55.524008 zram_generator::config[2296]: No configuration found. Oct 9 07:53:55.693433 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 9 07:53:55.778492 systemd[1]: Reloading finished in 365 ms. Oct 9 07:53:55.838450 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Oct 9 07:53:55.838591 systemd[1]: kubelet.service: Failed with result 'signal'. Oct 9 07:53:55.839669 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 9 07:53:55.859352 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 9 07:53:56.003143 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 9 07:53:56.016276 (kubelet)[2361]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 9 07:53:56.089269 kubelet[2361]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 9 07:53:56.090755 kubelet[2361]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 9 07:53:56.090755 kubelet[2361]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 9 07:53:56.090755 kubelet[2361]: I1009 07:53:56.089890 2361 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 9 07:53:56.385977 kubelet[2361]: I1009 07:53:56.385850 2361 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Oct 9 07:53:56.385977 kubelet[2361]: I1009 07:53:56.385894 2361 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 9 07:53:56.386196 kubelet[2361]: I1009 07:53:56.386152 2361 server.go:919] "Client rotation is on, will bootstrap in background" Oct 9 07:53:56.413392 kubelet[2361]: I1009 07:53:56.413156 2361 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 9 07:53:56.416100 kubelet[2361]: E1009 07:53:56.416027 2361 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://143.198.138.163:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.434861 kubelet[2361]: I1009 07:53:56.434326 2361 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 9 07:53:56.435458 kubelet[2361]: I1009 07:53:56.435425 2361 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 9 07:53:56.437925 kubelet[2361]: I1009 07:53:56.437852 2361 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Oct 9 07:53:56.439205 kubelet[2361]: I1009 07:53:56.438269 2361 topology_manager.go:138] "Creating topology manager with none policy" Oct 9 07:53:56.439205 kubelet[2361]: I1009 07:53:56.438303 2361 container_manager_linux.go:301] "Creating device plugin manager" Oct 9 07:53:56.439205 kubelet[2361]: I1009 07:53:56.438507 2361 state_mem.go:36] "Initialized new in-memory state store" Oct 9 07:53:56.439205 kubelet[2361]: I1009 07:53:56.438693 2361 kubelet.go:396] "Attempting to sync node with API server" Oct 9 07:53:56.439205 kubelet[2361]: I1009 07:53:56.438770 2361 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 9 07:53:56.439205 kubelet[2361]: I1009 07:53:56.438820 2361 kubelet.go:312] "Adding apiserver pod source" Oct 9 07:53:56.439205 kubelet[2361]: I1009 07:53:56.438845 2361 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 9 07:53:56.441093 kubelet[2361]: W1009 07:53:56.441029 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://143.198.138.163:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.1.0-c-f399a429fa&limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.441327 kubelet[2361]: E1009 07:53:56.441305 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://143.198.138.163:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.1.0-c-f399a429fa&limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.442055 kubelet[2361]: I1009 07:53:56.442029 2361 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Oct 9 07:53:56.447172 kubelet[2361]: W1009 07:53:56.447084 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://143.198.138.163:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.447172 kubelet[2361]: E1009 07:53:56.447185 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://143.198.138.163:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.448228 kubelet[2361]: I1009 07:53:56.447518 2361 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 9 07:53:56.448228 kubelet[2361]: W1009 07:53:56.447612 2361 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Oct 9 07:53:56.448731 kubelet[2361]: I1009 07:53:56.448682 2361 server.go:1256] "Started kubelet" Oct 9 07:53:56.451269 kubelet[2361]: I1009 07:53:56.451235 2361 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 9 07:53:56.454985 kubelet[2361]: E1009 07:53:56.454557 2361 event.go:355] "Unable to write event (may retry after sleeping)" err="Post \"https://143.198.138.163:6443/api/v1/namespaces/default/events\": dial tcp 143.198.138.163:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.1.0-c-f399a429fa.17fcb99bf2d67160 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.1.0-c-f399a429fa,UID:ci-4081.1.0-c-f399a429fa,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.1.0-c-f399a429fa,},FirstTimestamp:2024-10-09 07:53:56.44863728 +0000 UTC m=+0.425180879,LastTimestamp:2024-10-09 07:53:56.44863728 +0000 UTC m=+0.425180879,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.1.0-c-f399a429fa,}" Oct 9 07:53:56.460348 kubelet[2361]: I1009 07:53:56.458473 2361 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Oct 9 07:53:56.460348 kubelet[2361]: I1009 07:53:56.459666 2361 server.go:461] "Adding debug handlers to kubelet server" Oct 9 07:53:56.461088 kubelet[2361]: I1009 07:53:56.461061 2361 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 9 07:53:56.463054 kubelet[2361]: I1009 07:53:56.461332 2361 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 9 07:53:56.464600 kubelet[2361]: I1009 07:53:56.464569 2361 volume_manager.go:291] "Starting Kubelet Volume Manager" Oct 9 07:53:56.464750 kubelet[2361]: I1009 07:53:56.464667 2361 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Oct 9 07:53:56.464817 kubelet[2361]: I1009 07:53:56.464754 2361 reconciler_new.go:29] "Reconciler: start to sync state" Oct 9 07:53:56.465267 kubelet[2361]: W1009 07:53:56.465219 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://143.198.138.163:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.465267 kubelet[2361]: E1009 07:53:56.465269 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://143.198.138.163:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.466406 kubelet[2361]: E1009 07:53:56.466378 2361 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.138.163:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.1.0-c-f399a429fa?timeout=10s\": dial tcp 143.198.138.163:6443: connect: connection refused" interval="200ms" Oct 9 07:53:56.466488 kubelet[2361]: E1009 07:53:56.466474 2361 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 9 07:53:56.467292 kubelet[2361]: I1009 07:53:56.467274 2361 factory.go:221] Registration of the systemd container factory successfully Oct 9 07:53:56.467369 kubelet[2361]: I1009 07:53:56.467355 2361 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 9 07:53:56.469399 kubelet[2361]: I1009 07:53:56.469377 2361 factory.go:221] Registration of the containerd container factory successfully Oct 9 07:53:56.482887 kubelet[2361]: I1009 07:53:56.482856 2361 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 9 07:53:56.484475 kubelet[2361]: I1009 07:53:56.484451 2361 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 9 07:53:56.484596 kubelet[2361]: I1009 07:53:56.484589 2361 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 9 07:53:56.484696 kubelet[2361]: I1009 07:53:56.484687 2361 kubelet.go:2329] "Starting kubelet main sync loop" Oct 9 07:53:56.484879 kubelet[2361]: E1009 07:53:56.484866 2361 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 9 07:53:56.503055 kubelet[2361]: W1009 07:53:56.500507 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.RuntimeClass: Get "https://143.198.138.163:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.504159 kubelet[2361]: E1009 07:53:56.503454 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://143.198.138.163:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:56.504159 kubelet[2361]: I1009 07:53:56.503596 2361 cpu_manager.go:214] "Starting CPU manager" policy="none" Oct 9 07:53:56.504159 kubelet[2361]: I1009 07:53:56.503609 2361 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Oct 9 07:53:56.504159 kubelet[2361]: I1009 07:53:56.503649 2361 state_mem.go:36] "Initialized new in-memory state store" Oct 9 07:53:56.505732 kubelet[2361]: I1009 07:53:56.505681 2361 policy_none.go:49] "None policy: Start" Oct 9 07:53:56.506725 kubelet[2361]: I1009 07:53:56.506676 2361 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 9 07:53:56.506885 kubelet[2361]: I1009 07:53:56.506874 2361 state_mem.go:35] "Initializing new in-memory state store" Oct 9 07:53:56.514017 kubelet[2361]: I1009 07:53:56.513972 2361 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 9 07:53:56.514749 kubelet[2361]: I1009 07:53:56.514626 2361 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 9 07:53:56.521121 kubelet[2361]: E1009 07:53:56.521043 2361 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.1.0-c-f399a429fa\" not found" Oct 9 07:53:56.565987 kubelet[2361]: I1009 07:53:56.565954 2361 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.566490 kubelet[2361]: E1009 07:53:56.566440 2361 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://143.198.138.163:6443/api/v1/nodes\": dial tcp 143.198.138.163:6443: connect: connection refused" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.585963 kubelet[2361]: I1009 07:53:56.585883 2361 topology_manager.go:215] "Topology Admit Handler" podUID="aa4e66df739a44d7ff32a0ba783aca88" podNamespace="kube-system" podName="kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.587890 kubelet[2361]: I1009 07:53:56.587203 2361 topology_manager.go:215] "Topology Admit Handler" podUID="229d0f29c4ee5fb8707a9d2238b7ed72" podNamespace="kube-system" podName="kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.588391 kubelet[2361]: I1009 07:53:56.588365 2361 topology_manager.go:215] "Topology Admit Handler" podUID="35d3e9a003275f83fee7f476c309ac14" podNamespace="kube-system" podName="kube-scheduler-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.667329 kubelet[2361]: E1009 07:53:56.667190 2361 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.138.163:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.1.0-c-f399a429fa?timeout=10s\": dial tcp 143.198.138.163:6443: connect: connection refused" interval="400ms" Oct 9 07:53:56.765930 kubelet[2361]: I1009 07:53:56.765660 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.765930 kubelet[2361]: I1009 07:53:56.765920 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aa4e66df739a44d7ff32a0ba783aca88-ca-certs\") pod \"kube-apiserver-ci-4081.1.0-c-f399a429fa\" (UID: \"aa4e66df739a44d7ff32a0ba783aca88\") " pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.766137 kubelet[2361]: I1009 07:53:56.765978 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aa4e66df739a44d7ff32a0ba783aca88-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.1.0-c-f399a429fa\" (UID: \"aa4e66df739a44d7ff32a0ba783aca88\") " pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.766137 kubelet[2361]: I1009 07:53:56.766005 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-ca-certs\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.766137 kubelet[2361]: I1009 07:53:56.766024 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.766137 kubelet[2361]: I1009 07:53:56.766045 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/35d3e9a003275f83fee7f476c309ac14-kubeconfig\") pod \"kube-scheduler-ci-4081.1.0-c-f399a429fa\" (UID: \"35d3e9a003275f83fee7f476c309ac14\") " pod="kube-system/kube-scheduler-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.766137 kubelet[2361]: I1009 07:53:56.766062 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aa4e66df739a44d7ff32a0ba783aca88-k8s-certs\") pod \"kube-apiserver-ci-4081.1.0-c-f399a429fa\" (UID: \"aa4e66df739a44d7ff32a0ba783aca88\") " pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.766291 kubelet[2361]: I1009 07:53:56.766079 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-k8s-certs\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.766291 kubelet[2361]: I1009 07:53:56.766103 2361 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-kubeconfig\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.767852 kubelet[2361]: I1009 07:53:56.767404 2361 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.767852 kubelet[2361]: E1009 07:53:56.767786 2361 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://143.198.138.163:6443/api/v1/nodes\": dial tcp 143.198.138.163:6443: connect: connection refused" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:56.892068 kubelet[2361]: E1009 07:53:56.892018 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:56.892835 containerd[1597]: time="2024-10-09T07:53:56.892789482Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.1.0-c-f399a429fa,Uid:aa4e66df739a44d7ff32a0ba783aca88,Namespace:kube-system,Attempt:0,}" Oct 9 07:53:56.894110 kubelet[2361]: E1009 07:53:56.893971 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:56.894636 kubelet[2361]: E1009 07:53:56.894613 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:56.897405 systemd-resolved[1487]: Using degraded feature set TCP instead of UDP for DNS server 67.207.67.2. Oct 9 07:53:56.898563 containerd[1597]: time="2024-10-09T07:53:56.898483603Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.1.0-c-f399a429fa,Uid:229d0f29c4ee5fb8707a9d2238b7ed72,Namespace:kube-system,Attempt:0,}" Oct 9 07:53:56.899134 containerd[1597]: time="2024-10-09T07:53:56.899045450Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.1.0-c-f399a429fa,Uid:35d3e9a003275f83fee7f476c309ac14,Namespace:kube-system,Attempt:0,}" Oct 9 07:53:57.068772 kubelet[2361]: E1009 07:53:57.068612 2361 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.138.163:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.1.0-c-f399a429fa?timeout=10s\": dial tcp 143.198.138.163:6443: connect: connection refused" interval="800ms" Oct 9 07:53:57.169221 kubelet[2361]: I1009 07:53:57.169184 2361 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:57.169697 kubelet[2361]: E1009 07:53:57.169508 2361 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://143.198.138.163:6443/api/v1/nodes\": dial tcp 143.198.138.163:6443: connect: connection refused" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:57.493468 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3245433924.mount: Deactivated successfully. Oct 9 07:53:57.497330 containerd[1597]: time="2024-10-09T07:53:57.497272394Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 9 07:53:57.498181 containerd[1597]: time="2024-10-09T07:53:57.498130070Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Oct 9 07:53:57.498671 kubelet[2361]: W1009 07:53:57.498607 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.RuntimeClass: Get "https://143.198.138.163:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:57.498808 kubelet[2361]: E1009 07:53:57.498681 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://143.198.138.163:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:57.502337 containerd[1597]: time="2024-10-09T07:53:57.502273375Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 9 07:53:57.506098 containerd[1597]: time="2024-10-09T07:53:57.505168428Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Oct 9 07:53:57.506098 containerd[1597]: time="2024-10-09T07:53:57.505299559Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 9 07:53:57.507213 containerd[1597]: time="2024-10-09T07:53:57.507160618Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Oct 9 07:53:57.510737 containerd[1597]: time="2024-10-09T07:53:57.508823930Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 609.123793ms" Oct 9 07:53:57.512284 containerd[1597]: time="2024-10-09T07:53:57.512238458Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 9 07:53:57.514937 containerd[1597]: time="2024-10-09T07:53:57.514888828Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 9 07:53:57.517500 containerd[1597]: time="2024-10-09T07:53:57.517444780Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 624.562304ms" Oct 9 07:53:57.523112 containerd[1597]: time="2024-10-09T07:53:57.523045194Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 624.249577ms" Oct 9 07:53:57.611044 kubelet[2361]: W1009 07:53:57.610878 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://143.198.138.163:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:57.611044 kubelet[2361]: E1009 07:53:57.611011 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://143.198.138.163:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:57.721381 containerd[1597]: time="2024-10-09T07:53:57.720809542Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:53:57.723006 containerd[1597]: time="2024-10-09T07:53:57.722786005Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:53:57.723006 containerd[1597]: time="2024-10-09T07:53:57.722821127Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:53:57.724125 containerd[1597]: time="2024-10-09T07:53:57.723109721Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:53:57.724750 containerd[1597]: time="2024-10-09T07:53:57.724059221Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:53:57.724750 containerd[1597]: time="2024-10-09T07:53:57.724432692Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:53:57.724750 containerd[1597]: time="2024-10-09T07:53:57.724472838Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:53:57.724750 containerd[1597]: time="2024-10-09T07:53:57.724621758Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:53:57.731844 containerd[1597]: time="2024-10-09T07:53:57.731403783Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:53:57.731844 containerd[1597]: time="2024-10-09T07:53:57.731491664Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:53:57.731844 containerd[1597]: time="2024-10-09T07:53:57.731525398Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:53:57.731844 containerd[1597]: time="2024-10-09T07:53:57.731682723Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:53:57.828864 kubelet[2361]: W1009 07:53:57.828647 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://143.198.138.163:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.1.0-c-f399a429fa&limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:57.828864 kubelet[2361]: E1009 07:53:57.828741 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://143.198.138.163:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.1.0-c-f399a429fa&limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:57.870017 kubelet[2361]: E1009 07:53:57.869754 2361 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://143.198.138.163:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.1.0-c-f399a429fa?timeout=10s\": dial tcp 143.198.138.163:6443: connect: connection refused" interval="1.6s" Oct 9 07:53:57.887048 containerd[1597]: time="2024-10-09T07:53:57.886622093Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.1.0-c-f399a429fa,Uid:aa4e66df739a44d7ff32a0ba783aca88,Namespace:kube-system,Attempt:0,} returns sandbox id \"adb994573ac60dab7605c3633bd7d897aeeddb4cb3f46b5c84902541b0ecf95b\"" Oct 9 07:53:57.889271 containerd[1597]: time="2024-10-09T07:53:57.889144040Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.1.0-c-f399a429fa,Uid:229d0f29c4ee5fb8707a9d2238b7ed72,Namespace:kube-system,Attempt:0,} returns sandbox id \"e2411d782a22962d9d8701c9d92d6bfacc6dea71392cd36853197957b0638929\"" Oct 9 07:53:57.890209 kubelet[2361]: E1009 07:53:57.890180 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:57.890445 kubelet[2361]: E1009 07:53:57.890333 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:57.894357 containerd[1597]: time="2024-10-09T07:53:57.894311789Z" level=info msg="CreateContainer within sandbox \"adb994573ac60dab7605c3633bd7d897aeeddb4cb3f46b5c84902541b0ecf95b\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Oct 9 07:53:57.895904 containerd[1597]: time="2024-10-09T07:53:57.895869495Z" level=info msg="CreateContainer within sandbox \"e2411d782a22962d9d8701c9d92d6bfacc6dea71392cd36853197957b0638929\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Oct 9 07:53:57.902827 containerd[1597]: time="2024-10-09T07:53:57.902781379Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.1.0-c-f399a429fa,Uid:35d3e9a003275f83fee7f476c309ac14,Namespace:kube-system,Attempt:0,} returns sandbox id \"747126c0552bc9ab4616554a26c0b901acfefadf8a687eff47ceb37976dd1b97\"" Oct 9 07:53:57.904112 kubelet[2361]: E1009 07:53:57.903977 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:57.907389 containerd[1597]: time="2024-10-09T07:53:57.907332775Z" level=info msg="CreateContainer within sandbox \"747126c0552bc9ab4616554a26c0b901acfefadf8a687eff47ceb37976dd1b97\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Oct 9 07:53:57.914131 containerd[1597]: time="2024-10-09T07:53:57.913954624Z" level=info msg="CreateContainer within sandbox \"e2411d782a22962d9d8701c9d92d6bfacc6dea71392cd36853197957b0638929\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"901e47334be5a4ef9eb79e8946b41fc7a7d20e9f45a39e5f09207323238d1cf3\"" Oct 9 07:53:57.915488 containerd[1597]: time="2024-10-09T07:53:57.915449756Z" level=info msg="StartContainer for \"901e47334be5a4ef9eb79e8946b41fc7a7d20e9f45a39e5f09207323238d1cf3\"" Oct 9 07:53:57.917803 containerd[1597]: time="2024-10-09T07:53:57.917621872Z" level=info msg="CreateContainer within sandbox \"adb994573ac60dab7605c3633bd7d897aeeddb4cb3f46b5c84902541b0ecf95b\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"fcec1f67fa279d87b5a6a74663ed78e4113c9b95d27fd84b4abe215870c299ba\"" Oct 9 07:53:57.918444 containerd[1597]: time="2024-10-09T07:53:57.918394317Z" level=info msg="StartContainer for \"fcec1f67fa279d87b5a6a74663ed78e4113c9b95d27fd84b4abe215870c299ba\"" Oct 9 07:53:57.927045 containerd[1597]: time="2024-10-09T07:53:57.926950485Z" level=info msg="CreateContainer within sandbox \"747126c0552bc9ab4616554a26c0b901acfefadf8a687eff47ceb37976dd1b97\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"7d283b08b323dabb1d4125b58b8cdf63aa93bf08e5fadc99c02c04d2059838d8\"" Oct 9 07:53:57.928748 containerd[1597]: time="2024-10-09T07:53:57.928169156Z" level=info msg="StartContainer for \"7d283b08b323dabb1d4125b58b8cdf63aa93bf08e5fadc99c02c04d2059838d8\"" Oct 9 07:53:57.972659 kubelet[2361]: I1009 07:53:57.972614 2361 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:57.974118 kubelet[2361]: E1009 07:53:57.974070 2361 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://143.198.138.163:6443/api/v1/nodes\": dial tcp 143.198.138.163:6443: connect: connection refused" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:53:57.988765 kubelet[2361]: W1009 07:53:57.987582 2361 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://143.198.138.163:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:57.989667 kubelet[2361]: E1009 07:53:57.989639 2361 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://143.198.138.163:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 143.198.138.163:6443: connect: connection refused Oct 9 07:53:58.073522 containerd[1597]: time="2024-10-09T07:53:58.073460831Z" level=info msg="StartContainer for \"fcec1f67fa279d87b5a6a74663ed78e4113c9b95d27fd84b4abe215870c299ba\" returns successfully" Oct 9 07:53:58.088287 containerd[1597]: time="2024-10-09T07:53:58.087778476Z" level=info msg="StartContainer for \"7d283b08b323dabb1d4125b58b8cdf63aa93bf08e5fadc99c02c04d2059838d8\" returns successfully" Oct 9 07:53:58.098672 containerd[1597]: time="2024-10-09T07:53:58.098606279Z" level=info msg="StartContainer for \"901e47334be5a4ef9eb79e8946b41fc7a7d20e9f45a39e5f09207323238d1cf3\" returns successfully" Oct 9 07:53:58.521064 kubelet[2361]: E1009 07:53:58.520324 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:58.538225 kubelet[2361]: E1009 07:53:58.538178 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:58.544460 kubelet[2361]: E1009 07:53:58.544419 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:59.546256 kubelet[2361]: E1009 07:53:59.543811 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:53:59.576588 kubelet[2361]: I1009 07:53:59.576541 2361 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:54:00.285799 kubelet[2361]: I1009 07:54:00.285745 2361 kubelet_node_status.go:76] "Successfully registered node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:54:00.384333 kubelet[2361]: E1009 07:54:00.384281 2361 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-node-lease\" not found" interval="3.2s" Oct 9 07:54:00.442801 kubelet[2361]: I1009 07:54:00.442116 2361 apiserver.go:52] "Watching apiserver" Oct 9 07:54:00.465027 kubelet[2361]: I1009 07:54:00.464954 2361 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Oct 9 07:54:02.875905 kubelet[2361]: W1009 07:54:02.875844 2361 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 9 07:54:02.876534 kubelet[2361]: E1009 07:54:02.876312 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:03.316880 kubelet[2361]: W1009 07:54:03.316663 2361 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 9 07:54:03.317179 kubelet[2361]: E1009 07:54:03.317160 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:03.475894 kubelet[2361]: W1009 07:54:03.475851 2361 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 9 07:54:03.476902 kubelet[2361]: E1009 07:54:03.476833 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:03.550434 kubelet[2361]: E1009 07:54:03.549638 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:03.550434 kubelet[2361]: E1009 07:54:03.549638 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:03.550434 kubelet[2361]: E1009 07:54:03.550331 2361 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:03.701424 systemd[1]: Reloading requested from client PID 2634 ('systemctl') (unit session-7.scope)... Oct 9 07:54:03.702215 systemd[1]: Reloading... Oct 9 07:54:03.834747 zram_generator::config[2679]: No configuration found. Oct 9 07:54:04.049008 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Oct 9 07:54:04.138054 systemd[1]: Reloading finished in 435 ms. Oct 9 07:54:04.184593 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Oct 9 07:54:04.195291 systemd[1]: kubelet.service: Deactivated successfully. Oct 9 07:54:04.195699 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 9 07:54:04.201414 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 9 07:54:04.356970 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 9 07:54:04.371318 (kubelet)[2734]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 9 07:54:04.455607 kubelet[2734]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 9 07:54:04.455607 kubelet[2734]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 9 07:54:04.455607 kubelet[2734]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 9 07:54:04.456081 kubelet[2734]: I1009 07:54:04.455683 2734 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 9 07:54:04.465748 kubelet[2734]: I1009 07:54:04.464774 2734 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Oct 9 07:54:04.465748 kubelet[2734]: I1009 07:54:04.464827 2734 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 9 07:54:04.465748 kubelet[2734]: I1009 07:54:04.465536 2734 server.go:919] "Client rotation is on, will bootstrap in background" Oct 9 07:54:04.477812 kubelet[2734]: I1009 07:54:04.477442 2734 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 9 07:54:04.481035 kubelet[2734]: I1009 07:54:04.480994 2734 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 9 07:54:04.493783 kubelet[2734]: I1009 07:54:04.492416 2734 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 9 07:54:04.495623 kubelet[2734]: I1009 07:54:04.495012 2734 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 9 07:54:04.495623 kubelet[2734]: I1009 07:54:04.495317 2734 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Oct 9 07:54:04.495623 kubelet[2734]: I1009 07:54:04.495360 2734 topology_manager.go:138] "Creating topology manager with none policy" Oct 9 07:54:04.495623 kubelet[2734]: I1009 07:54:04.495376 2734 container_manager_linux.go:301] "Creating device plugin manager" Oct 9 07:54:04.495623 kubelet[2734]: I1009 07:54:04.495480 2734 state_mem.go:36] "Initialized new in-memory state store" Oct 9 07:54:04.498897 kubelet[2734]: I1009 07:54:04.498785 2734 kubelet.go:396] "Attempting to sync node with API server" Oct 9 07:54:04.498897 kubelet[2734]: I1009 07:54:04.498829 2734 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 9 07:54:04.498897 kubelet[2734]: I1009 07:54:04.498871 2734 kubelet.go:312] "Adding apiserver pod source" Oct 9 07:54:04.498897 kubelet[2734]: I1009 07:54:04.498891 2734 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 9 07:54:04.501587 sudo[2749]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Oct 9 07:54:04.503138 sudo[2749]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Oct 9 07:54:04.519002 kubelet[2734]: I1009 07:54:04.518923 2734 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Oct 9 07:54:04.519461 kubelet[2734]: I1009 07:54:04.519401 2734 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 9 07:54:04.520147 kubelet[2734]: I1009 07:54:04.520128 2734 server.go:1256] "Started kubelet" Oct 9 07:54:04.522937 kubelet[2734]: I1009 07:54:04.522915 2734 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 9 07:54:04.531741 kubelet[2734]: I1009 07:54:04.531042 2734 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Oct 9 07:54:04.533678 kubelet[2734]: I1009 07:54:04.533154 2734 volume_manager.go:291] "Starting Kubelet Volume Manager" Oct 9 07:54:04.533678 kubelet[2734]: I1009 07:54:04.533480 2734 server.go:461] "Adding debug handlers to kubelet server" Oct 9 07:54:04.537295 kubelet[2734]: I1009 07:54:04.537256 2734 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 9 07:54:04.543771 kubelet[2734]: I1009 07:54:04.543745 2734 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 9 07:54:04.546863 kubelet[2734]: I1009 07:54:04.537762 2734 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Oct 9 07:54:04.547122 kubelet[2734]: I1009 07:54:04.538026 2734 reconciler_new.go:29] "Reconciler: start to sync state" Oct 9 07:54:04.547542 kubelet[2734]: I1009 07:54:04.547436 2734 factory.go:221] Registration of the systemd container factory successfully Oct 9 07:54:04.548927 kubelet[2734]: I1009 07:54:04.547698 2734 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 9 07:54:04.558074 kubelet[2734]: E1009 07:54:04.558043 2734 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 9 07:54:04.560914 kubelet[2734]: I1009 07:54:04.560321 2734 factory.go:221] Registration of the containerd container factory successfully Oct 9 07:54:04.571784 kubelet[2734]: I1009 07:54:04.571447 2734 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 9 07:54:04.579172 kubelet[2734]: I1009 07:54:04.579139 2734 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 9 07:54:04.579442 kubelet[2734]: I1009 07:54:04.579428 2734 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 9 07:54:04.579519 kubelet[2734]: I1009 07:54:04.579511 2734 kubelet.go:2329] "Starting kubelet main sync loop" Oct 9 07:54:04.579635 kubelet[2734]: E1009 07:54:04.579624 2734 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 9 07:54:04.639923 kubelet[2734]: I1009 07:54:04.639664 2734 kubelet_node_status.go:73] "Attempting to register node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.669970 kubelet[2734]: I1009 07:54:04.669930 2734 kubelet_node_status.go:112] "Node was previously registered" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.670123 kubelet[2734]: I1009 07:54:04.670038 2734 kubelet_node_status.go:76] "Successfully registered node" node="ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.680919 kubelet[2734]: E1009 07:54:04.680306 2734 kubelet.go:2353] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 9 07:54:04.710803 kubelet[2734]: I1009 07:54:04.710767 2734 cpu_manager.go:214] "Starting CPU manager" policy="none" Oct 9 07:54:04.710803 kubelet[2734]: I1009 07:54:04.710815 2734 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Oct 9 07:54:04.711027 kubelet[2734]: I1009 07:54:04.710855 2734 state_mem.go:36] "Initialized new in-memory state store" Oct 9 07:54:04.711746 kubelet[2734]: I1009 07:54:04.711266 2734 state_mem.go:88] "Updated default CPUSet" cpuSet="" Oct 9 07:54:04.711746 kubelet[2734]: I1009 07:54:04.711319 2734 state_mem.go:96] "Updated CPUSet assignments" assignments={} Oct 9 07:54:04.711746 kubelet[2734]: I1009 07:54:04.711336 2734 policy_none.go:49] "None policy: Start" Oct 9 07:54:04.714202 kubelet[2734]: I1009 07:54:04.714173 2734 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 9 07:54:04.714202 kubelet[2734]: I1009 07:54:04.714213 2734 state_mem.go:35] "Initializing new in-memory state store" Oct 9 07:54:04.714787 kubelet[2734]: I1009 07:54:04.714375 2734 state_mem.go:75] "Updated machine memory state" Oct 9 07:54:04.721233 kubelet[2734]: I1009 07:54:04.721064 2734 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 9 07:54:04.731735 kubelet[2734]: I1009 07:54:04.731507 2734 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 9 07:54:04.881294 kubelet[2734]: I1009 07:54:04.881250 2734 topology_manager.go:215] "Topology Admit Handler" podUID="aa4e66df739a44d7ff32a0ba783aca88" podNamespace="kube-system" podName="kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.881430 kubelet[2734]: I1009 07:54:04.881357 2734 topology_manager.go:215] "Topology Admit Handler" podUID="229d0f29c4ee5fb8707a9d2238b7ed72" podNamespace="kube-system" podName="kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.881430 kubelet[2734]: I1009 07:54:04.881393 2734 topology_manager.go:215] "Topology Admit Handler" podUID="35d3e9a003275f83fee7f476c309ac14" podNamespace="kube-system" podName="kube-scheduler-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.896068 kubelet[2734]: W1009 07:54:04.895355 2734 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 9 07:54:04.896194 kubelet[2734]: W1009 07:54:04.896069 2734 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 9 07:54:04.896194 kubelet[2734]: E1009 07:54:04.896153 2734 kubelet.go:1921] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4081.1.0-c-f399a429fa\" already exists" pod="kube-system/kube-scheduler-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.896287 kubelet[2734]: W1009 07:54:04.896232 2734 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Oct 9 07:54:04.896287 kubelet[2734]: E1009 07:54:04.896263 2734 kubelet.go:1921] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4081.1.0-c-f399a429fa\" already exists" pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.896340 kubelet[2734]: E1009 07:54:04.896316 2734 kubelet.go:1921] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" already exists" pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.951291 kubelet[2734]: I1009 07:54:04.951235 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-ca-certs\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.951444 kubelet[2734]: I1009 07:54:04.951362 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.951444 kubelet[2734]: I1009 07:54:04.951424 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-k8s-certs\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.951526 kubelet[2734]: I1009 07:54:04.951486 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.951555 kubelet[2734]: I1009 07:54:04.951526 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/229d0f29c4ee5fb8707a9d2238b7ed72-kubeconfig\") pod \"kube-controller-manager-ci-4081.1.0-c-f399a429fa\" (UID: \"229d0f29c4ee5fb8707a9d2238b7ed72\") " pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.952435 kubelet[2734]: I1009 07:54:04.951579 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/35d3e9a003275f83fee7f476c309ac14-kubeconfig\") pod \"kube-scheduler-ci-4081.1.0-c-f399a429fa\" (UID: \"35d3e9a003275f83fee7f476c309ac14\") " pod="kube-system/kube-scheduler-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.952435 kubelet[2734]: I1009 07:54:04.951633 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aa4e66df739a44d7ff32a0ba783aca88-ca-certs\") pod \"kube-apiserver-ci-4081.1.0-c-f399a429fa\" (UID: \"aa4e66df739a44d7ff32a0ba783aca88\") " pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.952435 kubelet[2734]: I1009 07:54:04.951661 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aa4e66df739a44d7ff32a0ba783aca88-k8s-certs\") pod \"kube-apiserver-ci-4081.1.0-c-f399a429fa\" (UID: \"aa4e66df739a44d7ff32a0ba783aca88\") " pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:04.952435 kubelet[2734]: I1009 07:54:04.951746 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aa4e66df739a44d7ff32a0ba783aca88-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.1.0-c-f399a429fa\" (UID: \"aa4e66df739a44d7ff32a0ba783aca88\") " pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" Oct 9 07:54:05.199829 kubelet[2734]: E1009 07:54:05.199668 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:05.200522 kubelet[2734]: E1009 07:54:05.200491 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:05.200908 kubelet[2734]: E1009 07:54:05.200885 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:05.292461 sudo[2749]: pam_unix(sudo:session): session closed for user root Oct 9 07:54:05.523865 kubelet[2734]: I1009 07:54:05.522086 2734 apiserver.go:52] "Watching apiserver" Oct 9 07:54:05.547480 kubelet[2734]: I1009 07:54:05.547369 2734 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Oct 9 07:54:05.633730 kubelet[2734]: E1009 07:54:05.633539 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:05.634408 kubelet[2734]: E1009 07:54:05.634280 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:05.636881 kubelet[2734]: E1009 07:54:05.636040 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:05.700285 kubelet[2734]: I1009 07:54:05.700061 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.1.0-c-f399a429fa" podStartSLOduration=2.6999874090000002 podStartE2EDuration="2.699987409s" podCreationTimestamp="2024-10-09 07:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-09 07:54:05.676071686 +0000 UTC m=+1.296606107" watchObservedRunningTime="2024-10-09 07:54:05.699987409 +0000 UTC m=+1.320521831" Oct 9 07:54:05.727355 kubelet[2734]: I1009 07:54:05.726107 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.1.0-c-f399a429fa" podStartSLOduration=2.726063711 podStartE2EDuration="2.726063711s" podCreationTimestamp="2024-10-09 07:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-09 07:54:05.70395034 +0000 UTC m=+1.324484765" watchObservedRunningTime="2024-10-09 07:54:05.726063711 +0000 UTC m=+1.346598112" Oct 9 07:54:05.763742 kubelet[2734]: I1009 07:54:05.761526 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.1.0-c-f399a429fa" podStartSLOduration=3.76144486 podStartE2EDuration="3.76144486s" podCreationTimestamp="2024-10-09 07:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-09 07:54:05.726391703 +0000 UTC m=+1.346926124" watchObservedRunningTime="2024-10-09 07:54:05.76144486 +0000 UTC m=+1.381979285" Oct 9 07:54:06.635912 kubelet[2734]: E1009 07:54:06.635797 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:07.374961 kubelet[2734]: E1009 07:54:07.374452 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:07.444142 sudo[1811]: pam_unix(sudo:session): session closed for user root Oct 9 07:54:07.448508 sshd[1804]: pam_unix(sshd:session): session closed for user core Oct 9 07:54:07.452809 systemd[1]: sshd@6-143.198.138.163:22-139.178.89.65:40434.service: Deactivated successfully. Oct 9 07:54:07.457049 systemd-logind[1572]: Session 7 logged out. Waiting for processes to exit. Oct 9 07:54:07.458241 systemd[1]: session-7.scope: Deactivated successfully. Oct 9 07:54:07.459816 systemd-logind[1572]: Removed session 7. Oct 9 07:54:10.286602 kubelet[2734]: E1009 07:54:10.285331 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:10.643704 kubelet[2734]: E1009 07:54:10.643555 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:11.646087 kubelet[2734]: E1009 07:54:11.645931 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:13.207223 kubelet[2734]: E1009 07:54:13.205989 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:13.650246 kubelet[2734]: E1009 07:54:13.650159 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:16.360544 kubelet[2734]: I1009 07:54:16.360506 2734 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Oct 9 07:54:16.361115 containerd[1597]: time="2024-10-09T07:54:16.361062561Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Oct 9 07:54:16.362306 kubelet[2734]: I1009 07:54:16.361376 2734 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Oct 9 07:54:16.559370 kubelet[2734]: I1009 07:54:16.559323 2734 topology_manager.go:215] "Topology Admit Handler" podUID="78b390d5-9be5-49e9-8aff-024bd409365c" podNamespace="kube-system" podName="kube-proxy-jhkc9" Oct 9 07:54:16.571823 kubelet[2734]: I1009 07:54:16.571773 2734 topology_manager.go:215] "Topology Admit Handler" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" podNamespace="kube-system" podName="cilium-vzlv4" Oct 9 07:54:16.626680 kubelet[2734]: I1009 07:54:16.626416 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-run\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.626680 kubelet[2734]: I1009 07:54:16.626486 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/78b390d5-9be5-49e9-8aff-024bd409365c-xtables-lock\") pod \"kube-proxy-jhkc9\" (UID: \"78b390d5-9be5-49e9-8aff-024bd409365c\") " pod="kube-system/kube-proxy-jhkc9" Oct 9 07:54:16.627728 kubelet[2734]: I1009 07:54:16.627158 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hubble-tls\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.627728 kubelet[2734]: I1009 07:54:16.627229 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-bpf-maps\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.627728 kubelet[2734]: I1009 07:54:16.627259 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-net\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.627728 kubelet[2734]: I1009 07:54:16.627290 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4ggv\" (UniqueName: \"kubernetes.io/projected/78b390d5-9be5-49e9-8aff-024bd409365c-kube-api-access-d4ggv\") pod \"kube-proxy-jhkc9\" (UID: \"78b390d5-9be5-49e9-8aff-024bd409365c\") " pod="kube-system/kube-proxy-jhkc9" Oct 9 07:54:16.627728 kubelet[2734]: I1009 07:54:16.627312 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-etc-cni-netd\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.627728 kubelet[2734]: I1009 07:54:16.627334 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/78b390d5-9be5-49e9-8aff-024bd409365c-kube-proxy\") pod \"kube-proxy-jhkc9\" (UID: \"78b390d5-9be5-49e9-8aff-024bd409365c\") " pod="kube-system/kube-proxy-jhkc9" Oct 9 07:54:16.629559 kubelet[2734]: I1009 07:54:16.627360 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/78b390d5-9be5-49e9-8aff-024bd409365c-lib-modules\") pod \"kube-proxy-jhkc9\" (UID: \"78b390d5-9be5-49e9-8aff-024bd409365c\") " pod="kube-system/kube-proxy-jhkc9" Oct 9 07:54:16.629559 kubelet[2734]: I1009 07:54:16.627379 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hostproc\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.629559 kubelet[2734]: I1009 07:54:16.627398 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-xtables-lock\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.629559 kubelet[2734]: I1009 07:54:16.627420 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-kernel\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.629559 kubelet[2734]: I1009 07:54:16.627438 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnn67\" (UniqueName: \"kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-kube-api-access-jnn67\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.629559 kubelet[2734]: I1009 07:54:16.627456 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-cgroup\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.630802 kubelet[2734]: I1009 07:54:16.627476 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-lib-modules\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.630802 kubelet[2734]: I1009 07:54:16.627493 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-config-path\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.630802 kubelet[2734]: I1009 07:54:16.627515 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cni-path\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.630802 kubelet[2734]: I1009 07:54:16.630056 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3481fcc3-e066-4e00-a996-ec6227b2a6a2-clustermesh-secrets\") pod \"cilium-vzlv4\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " pod="kube-system/cilium-vzlv4" Oct 9 07:54:16.879543 kubelet[2734]: E1009 07:54:16.879405 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:16.882272 containerd[1597]: time="2024-10-09T07:54:16.882166052Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jhkc9,Uid:78b390d5-9be5-49e9-8aff-024bd409365c,Namespace:kube-system,Attempt:0,}" Oct 9 07:54:16.883785 kubelet[2734]: E1009 07:54:16.882659 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:16.885040 containerd[1597]: time="2024-10-09T07:54:16.884797094Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vzlv4,Uid:3481fcc3-e066-4e00-a996-ec6227b2a6a2,Namespace:kube-system,Attempt:0,}" Oct 9 07:54:16.943572 containerd[1597]: time="2024-10-09T07:54:16.943275657Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:54:16.943572 containerd[1597]: time="2024-10-09T07:54:16.943374655Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:54:16.943572 containerd[1597]: time="2024-10-09T07:54:16.943390719Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:16.943572 containerd[1597]: time="2024-10-09T07:54:16.943512057Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:16.947767 containerd[1597]: time="2024-10-09T07:54:16.947522705Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:54:16.947767 containerd[1597]: time="2024-10-09T07:54:16.947589505Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:54:16.947767 containerd[1597]: time="2024-10-09T07:54:16.947613601Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:16.948282 containerd[1597]: time="2024-10-09T07:54:16.947761803Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:17.060753 containerd[1597]: time="2024-10-09T07:54:17.060194025Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vzlv4,Uid:3481fcc3-e066-4e00-a996-ec6227b2a6a2,Namespace:kube-system,Attempt:0,} returns sandbox id \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\"" Oct 9 07:54:17.067805 kubelet[2734]: E1009 07:54:17.063638 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:17.067998 containerd[1597]: time="2024-10-09T07:54:17.066243577Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jhkc9,Uid:78b390d5-9be5-49e9-8aff-024bd409365c,Namespace:kube-system,Attempt:0,} returns sandbox id \"a26cfef7e24cc05a39262db2fcae9a329a90d25305dc5408718c725304206aa1\"" Oct 9 07:54:17.074748 kubelet[2734]: E1009 07:54:17.074048 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:17.075431 containerd[1597]: time="2024-10-09T07:54:17.075386818Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Oct 9 07:54:17.089070 containerd[1597]: time="2024-10-09T07:54:17.086911086Z" level=info msg="CreateContainer within sandbox \"a26cfef7e24cc05a39262db2fcae9a329a90d25305dc5408718c725304206aa1\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Oct 9 07:54:17.105742 kubelet[2734]: I1009 07:54:17.105132 2734 topology_manager.go:215] "Topology Admit Handler" podUID="515d30f1-a4f9-465b-95eb-b1a89e8c318f" podNamespace="kube-system" podName="cilium-operator-5cc964979-4q8km" Oct 9 07:54:17.121974 containerd[1597]: time="2024-10-09T07:54:17.121926366Z" level=info msg="CreateContainer within sandbox \"a26cfef7e24cc05a39262db2fcae9a329a90d25305dc5408718c725304206aa1\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"6d9513274f464ce8384f57c271d36d9c7d3cb0dd16d98f8c6dd586cced541504\"" Oct 9 07:54:17.136567 containerd[1597]: time="2024-10-09T07:54:17.128924387Z" level=info msg="StartContainer for \"6d9513274f464ce8384f57c271d36d9c7d3cb0dd16d98f8c6dd586cced541504\"" Oct 9 07:54:17.136755 kubelet[2734]: I1009 07:54:17.133093 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkbdq\" (UniqueName: \"kubernetes.io/projected/515d30f1-a4f9-465b-95eb-b1a89e8c318f-kube-api-access-qkbdq\") pod \"cilium-operator-5cc964979-4q8km\" (UID: \"515d30f1-a4f9-465b-95eb-b1a89e8c318f\") " pod="kube-system/cilium-operator-5cc964979-4q8km" Oct 9 07:54:17.136755 kubelet[2734]: I1009 07:54:17.133686 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/515d30f1-a4f9-465b-95eb-b1a89e8c318f-cilium-config-path\") pod \"cilium-operator-5cc964979-4q8km\" (UID: \"515d30f1-a4f9-465b-95eb-b1a89e8c318f\") " pod="kube-system/cilium-operator-5cc964979-4q8km" Oct 9 07:54:17.221926 containerd[1597]: time="2024-10-09T07:54:17.221878947Z" level=info msg="StartContainer for \"6d9513274f464ce8384f57c271d36d9c7d3cb0dd16d98f8c6dd586cced541504\" returns successfully" Oct 9 07:54:17.381945 kubelet[2734]: E1009 07:54:17.381855 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:17.421213 kubelet[2734]: E1009 07:54:17.421010 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:17.424128 containerd[1597]: time="2024-10-09T07:54:17.423441602Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-4q8km,Uid:515d30f1-a4f9-465b-95eb-b1a89e8c318f,Namespace:kube-system,Attempt:0,}" Oct 9 07:54:17.460045 containerd[1597]: time="2024-10-09T07:54:17.459003448Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:54:17.460045 containerd[1597]: time="2024-10-09T07:54:17.459816095Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:54:17.460045 containerd[1597]: time="2024-10-09T07:54:17.459834375Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:17.460045 containerd[1597]: time="2024-10-09T07:54:17.459943926Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:17.545589 containerd[1597]: time="2024-10-09T07:54:17.545501956Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-4q8km,Uid:515d30f1-a4f9-465b-95eb-b1a89e8c318f,Namespace:kube-system,Attempt:0,} returns sandbox id \"4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a\"" Oct 9 07:54:17.547663 kubelet[2734]: E1009 07:54:17.547630 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:17.666793 kubelet[2734]: E1009 07:54:17.665491 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:17.699052 kubelet[2734]: I1009 07:54:17.698376 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-jhkc9" podStartSLOduration=1.698298724 podStartE2EDuration="1.698298724s" podCreationTimestamp="2024-10-09 07:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-09 07:54:17.69716144 +0000 UTC m=+13.317695861" watchObservedRunningTime="2024-10-09 07:54:17.698298724 +0000 UTC m=+13.318833147" Oct 9 07:54:20.605888 update_engine[1577]: I20241009 07:54:20.605785 1577 update_attempter.cc:509] Updating boot flags... Oct 9 07:54:20.666763 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (3100) Oct 9 07:54:20.755103 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (3100) Oct 9 07:54:22.256010 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3153490002.mount: Deactivated successfully. Oct 9 07:54:24.506966 containerd[1597]: time="2024-10-09T07:54:24.506749304Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166735343" Oct 9 07:54:24.506966 containerd[1597]: time="2024-10-09T07:54:24.506911115Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:54:24.509255 containerd[1597]: time="2024-10-09T07:54:24.508976663Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 7.43274495s" Oct 9 07:54:24.509255 containerd[1597]: time="2024-10-09T07:54:24.509021886Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Oct 9 07:54:24.512744 containerd[1597]: time="2024-10-09T07:54:24.512418659Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:54:24.515931 containerd[1597]: time="2024-10-09T07:54:24.515896549Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Oct 9 07:54:24.521138 containerd[1597]: time="2024-10-09T07:54:24.520846820Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Oct 9 07:54:24.636891 containerd[1597]: time="2024-10-09T07:54:24.636704670Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\"" Oct 9 07:54:24.640752 containerd[1597]: time="2024-10-09T07:54:24.640051747Z" level=info msg="StartContainer for \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\"" Oct 9 07:54:24.744740 containerd[1597]: time="2024-10-09T07:54:24.744684650Z" level=info msg="StartContainer for \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\" returns successfully" Oct 9 07:54:24.816072 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a-rootfs.mount: Deactivated successfully. Oct 9 07:54:24.845900 containerd[1597]: time="2024-10-09T07:54:24.833096040Z" level=info msg="shim disconnected" id=a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a namespace=k8s.io Oct 9 07:54:24.845900 containerd[1597]: time="2024-10-09T07:54:24.845659063Z" level=warning msg="cleaning up after shim disconnected" id=a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a namespace=k8s.io Oct 9 07:54:24.845900 containerd[1597]: time="2024-10-09T07:54:24.845677449Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:54:24.864763 containerd[1597]: time="2024-10-09T07:54:24.864503050Z" level=warning msg="cleanup warnings time=\"2024-10-09T07:54:24Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Oct 9 07:54:25.698233 kubelet[2734]: E1009 07:54:25.698198 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:25.702973 containerd[1597]: time="2024-10-09T07:54:25.702770780Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Oct 9 07:54:25.739552 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3541352572.mount: Deactivated successfully. Oct 9 07:54:25.748889 containerd[1597]: time="2024-10-09T07:54:25.748766502Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\"" Oct 9 07:54:25.750751 containerd[1597]: time="2024-10-09T07:54:25.750036338Z" level=info msg="StartContainer for \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\"" Oct 9 07:54:25.853893 containerd[1597]: time="2024-10-09T07:54:25.853764591Z" level=info msg="StartContainer for \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\" returns successfully" Oct 9 07:54:25.871146 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 9 07:54:25.871474 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 9 07:54:25.871561 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Oct 9 07:54:25.881221 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 9 07:54:25.917973 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 9 07:54:25.946807 containerd[1597]: time="2024-10-09T07:54:25.946625170Z" level=info msg="shim disconnected" id=cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470 namespace=k8s.io Oct 9 07:54:25.946807 containerd[1597]: time="2024-10-09T07:54:25.946697922Z" level=warning msg="cleaning up after shim disconnected" id=cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470 namespace=k8s.io Oct 9 07:54:25.947357 containerd[1597]: time="2024-10-09T07:54:25.947179106Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:54:26.350106 containerd[1597]: time="2024-10-09T07:54:26.349840616Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:54:26.350999 containerd[1597]: time="2024-10-09T07:54:26.350606188Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18907197" Oct 9 07:54:26.350999 containerd[1597]: time="2024-10-09T07:54:26.350943557Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 9 07:54:26.352496 containerd[1597]: time="2024-10-09T07:54:26.352432481Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 1.831541005s" Oct 9 07:54:26.352496 containerd[1597]: time="2024-10-09T07:54:26.352473252Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Oct 9 07:54:26.356511 containerd[1597]: time="2024-10-09T07:54:26.356460069Z" level=info msg="CreateContainer within sandbox \"4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Oct 9 07:54:26.375630 containerd[1597]: time="2024-10-09T07:54:26.375381730Z" level=info msg="CreateContainer within sandbox \"4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\"" Oct 9 07:54:26.376731 containerd[1597]: time="2024-10-09T07:54:26.376594313Z" level=info msg="StartContainer for \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\"" Oct 9 07:54:26.442545 containerd[1597]: time="2024-10-09T07:54:26.442035585Z" level=info msg="StartContainer for \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\" returns successfully" Oct 9 07:54:26.612684 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470-rootfs.mount: Deactivated successfully. Oct 9 07:54:26.705287 kubelet[2734]: E1009 07:54:26.704945 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:26.712650 kubelet[2734]: E1009 07:54:26.712127 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:26.716751 containerd[1597]: time="2024-10-09T07:54:26.716502763Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Oct 9 07:54:26.746394 containerd[1597]: time="2024-10-09T07:54:26.746288534Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\"" Oct 9 07:54:26.748576 containerd[1597]: time="2024-10-09T07:54:26.747243748Z" level=info msg="StartContainer for \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\"" Oct 9 07:54:26.850175 kubelet[2734]: I1009 07:54:26.850127 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-5cc964979-4q8km" podStartSLOduration=2.046899451 podStartE2EDuration="10.850079731s" podCreationTimestamp="2024-10-09 07:54:16 +0000 UTC" firstStartedPulling="2024-10-09 07:54:17.54955302 +0000 UTC m=+13.170087431" lastFinishedPulling="2024-10-09 07:54:26.35273331 +0000 UTC m=+21.973267711" observedRunningTime="2024-10-09 07:54:26.744533899 +0000 UTC m=+22.365068314" watchObservedRunningTime="2024-10-09 07:54:26.850079731 +0000 UTC m=+22.470614152" Oct 9 07:54:26.942782 containerd[1597]: time="2024-10-09T07:54:26.942639281Z" level=info msg="StartContainer for \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\" returns successfully" Oct 9 07:54:27.010321 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46-rootfs.mount: Deactivated successfully. Oct 9 07:54:27.017012 containerd[1597]: time="2024-10-09T07:54:27.016913138Z" level=info msg="shim disconnected" id=8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46 namespace=k8s.io Oct 9 07:54:27.017328 containerd[1597]: time="2024-10-09T07:54:27.017082895Z" level=warning msg="cleaning up after shim disconnected" id=8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46 namespace=k8s.io Oct 9 07:54:27.017328 containerd[1597]: time="2024-10-09T07:54:27.017099903Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:54:27.716896 kubelet[2734]: E1009 07:54:27.716411 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:27.719423 kubelet[2734]: E1009 07:54:27.719362 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:27.729202 containerd[1597]: time="2024-10-09T07:54:27.729152348Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Oct 9 07:54:27.757608 containerd[1597]: time="2024-10-09T07:54:27.757032393Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\"" Oct 9 07:54:27.759743 containerd[1597]: time="2024-10-09T07:54:27.758983561Z" level=info msg="StartContainer for \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\"" Oct 9 07:54:27.837343 containerd[1597]: time="2024-10-09T07:54:27.837181069Z" level=info msg="StartContainer for \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\" returns successfully" Oct 9 07:54:27.863301 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65-rootfs.mount: Deactivated successfully. Oct 9 07:54:27.864033 containerd[1597]: time="2024-10-09T07:54:27.863788971Z" level=info msg="shim disconnected" id=89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65 namespace=k8s.io Oct 9 07:54:27.864033 containerd[1597]: time="2024-10-09T07:54:27.863859448Z" level=warning msg="cleaning up after shim disconnected" id=89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65 namespace=k8s.io Oct 9 07:54:27.864033 containerd[1597]: time="2024-10-09T07:54:27.863872438Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:54:28.724662 kubelet[2734]: E1009 07:54:28.722199 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:28.726288 containerd[1597]: time="2024-10-09T07:54:28.725482322Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Oct 9 07:54:28.755495 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount899242734.mount: Deactivated successfully. Oct 9 07:54:28.765449 containerd[1597]: time="2024-10-09T07:54:28.765388298Z" level=info msg="CreateContainer within sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\"" Oct 9 07:54:28.769956 containerd[1597]: time="2024-10-09T07:54:28.769878269Z" level=info msg="StartContainer for \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\"" Oct 9 07:54:28.852970 containerd[1597]: time="2024-10-09T07:54:28.852809662Z" level=info msg="StartContainer for \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\" returns successfully" Oct 9 07:54:29.006914 kubelet[2734]: I1009 07:54:29.006880 2734 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Oct 9 07:54:29.045682 kubelet[2734]: I1009 07:54:29.045628 2734 topology_manager.go:215] "Topology Admit Handler" podUID="90b51489-005a-408c-8757-54b552fddc3f" podNamespace="kube-system" podName="coredns-76f75df574-n62tj" Oct 9 07:54:29.051891 kubelet[2734]: I1009 07:54:29.047440 2734 topology_manager.go:215] "Topology Admit Handler" podUID="35e137ea-8f33-425b-a35d-8ab05164f05b" podNamespace="kube-system" podName="coredns-76f75df574-m9zpt" Oct 9 07:54:29.145880 kubelet[2734]: I1009 07:54:29.145833 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9rm6\" (UniqueName: \"kubernetes.io/projected/35e137ea-8f33-425b-a35d-8ab05164f05b-kube-api-access-n9rm6\") pod \"coredns-76f75df574-m9zpt\" (UID: \"35e137ea-8f33-425b-a35d-8ab05164f05b\") " pod="kube-system/coredns-76f75df574-m9zpt" Oct 9 07:54:29.146177 kubelet[2734]: I1009 07:54:29.146155 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgvm4\" (UniqueName: \"kubernetes.io/projected/90b51489-005a-408c-8757-54b552fddc3f-kube-api-access-hgvm4\") pod \"coredns-76f75df574-n62tj\" (UID: \"90b51489-005a-408c-8757-54b552fddc3f\") " pod="kube-system/coredns-76f75df574-n62tj" Oct 9 07:54:29.146408 kubelet[2734]: I1009 07:54:29.146391 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e137ea-8f33-425b-a35d-8ab05164f05b-config-volume\") pod \"coredns-76f75df574-m9zpt\" (UID: \"35e137ea-8f33-425b-a35d-8ab05164f05b\") " pod="kube-system/coredns-76f75df574-m9zpt" Oct 9 07:54:29.146623 kubelet[2734]: I1009 07:54:29.146554 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90b51489-005a-408c-8757-54b552fddc3f-config-volume\") pod \"coredns-76f75df574-n62tj\" (UID: \"90b51489-005a-408c-8757-54b552fddc3f\") " pod="kube-system/coredns-76f75df574-n62tj" Oct 9 07:54:29.362989 kubelet[2734]: E1009 07:54:29.362865 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:29.364075 kubelet[2734]: E1009 07:54:29.363925 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:29.367314 containerd[1597]: time="2024-10-09T07:54:29.367199156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-m9zpt,Uid:35e137ea-8f33-425b-a35d-8ab05164f05b,Namespace:kube-system,Attempt:0,}" Oct 9 07:54:29.368087 containerd[1597]: time="2024-10-09T07:54:29.367219004Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-n62tj,Uid:90b51489-005a-408c-8757-54b552fddc3f,Namespace:kube-system,Attempt:0,}" Oct 9 07:54:29.731853 kubelet[2734]: E1009 07:54:29.730303 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:30.735895 kubelet[2734]: E1009 07:54:30.735851 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:31.114340 systemd-networkd[1226]: cilium_host: Link UP Oct 9 07:54:31.115851 systemd-networkd[1226]: cilium_net: Link UP Oct 9 07:54:31.116697 systemd-networkd[1226]: cilium_net: Gained carrier Oct 9 07:54:31.118376 systemd-networkd[1226]: cilium_host: Gained carrier Oct 9 07:54:31.260980 systemd-networkd[1226]: cilium_vxlan: Link UP Oct 9 07:54:31.260989 systemd-networkd[1226]: cilium_vxlan: Gained carrier Oct 9 07:54:31.653095 systemd-networkd[1226]: cilium_net: Gained IPv6LL Oct 9 07:54:31.653458 systemd-networkd[1226]: cilium_host: Gained IPv6LL Oct 9 07:54:31.692769 kernel: NET: Registered PF_ALG protocol family Oct 9 07:54:31.738107 kubelet[2734]: E1009 07:54:31.737831 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:32.624092 systemd-networkd[1226]: lxc_health: Link UP Oct 9 07:54:32.632363 systemd-networkd[1226]: lxc_health: Gained carrier Oct 9 07:54:32.886659 kubelet[2734]: E1009 07:54:32.885981 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:32.911910 kubelet[2734]: I1009 07:54:32.911862 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-vzlv4" podStartSLOduration=9.472416444 podStartE2EDuration="16.911808895s" podCreationTimestamp="2024-10-09 07:54:16 +0000 UTC" firstStartedPulling="2024-10-09 07:54:17.069995801 +0000 UTC m=+12.690530202" lastFinishedPulling="2024-10-09 07:54:24.50938824 +0000 UTC m=+20.129922653" observedRunningTime="2024-10-09 07:54:29.759740279 +0000 UTC m=+25.380274700" watchObservedRunningTime="2024-10-09 07:54:32.911808895 +0000 UTC m=+28.532343331" Oct 9 07:54:32.997063 systemd-networkd[1226]: lxc974c2623f000: Link UP Oct 9 07:54:33.004844 kernel: eth0: renamed from tmpb4f3b Oct 9 07:54:33.017849 systemd-networkd[1226]: lxc974c2623f000: Gained carrier Oct 9 07:54:33.076191 systemd-networkd[1226]: lxc00fb2a6f58ff: Link UP Oct 9 07:54:33.086832 kernel: eth0: renamed from tmp30561 Oct 9 07:54:33.094400 systemd-networkd[1226]: lxc00fb2a6f58ff: Gained carrier Oct 9 07:54:33.130074 systemd-networkd[1226]: cilium_vxlan: Gained IPv6LL Oct 9 07:54:34.213094 systemd-networkd[1226]: lxc_health: Gained IPv6LL Oct 9 07:54:34.408524 systemd-networkd[1226]: lxc00fb2a6f58ff: Gained IPv6LL Oct 9 07:54:34.597112 systemd-networkd[1226]: lxc974c2623f000: Gained IPv6LL Oct 9 07:54:38.971181 containerd[1597]: time="2024-10-09T07:54:38.970561227Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:54:38.971181 containerd[1597]: time="2024-10-09T07:54:38.970636387Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:54:38.971181 containerd[1597]: time="2024-10-09T07:54:38.970674462Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:38.971181 containerd[1597]: time="2024-10-09T07:54:38.970881874Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:38.984060 containerd[1597]: time="2024-10-09T07:54:38.979125094Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:54:38.984060 containerd[1597]: time="2024-10-09T07:54:38.979182563Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:54:38.984060 containerd[1597]: time="2024-10-09T07:54:38.979215281Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:38.984060 containerd[1597]: time="2024-10-09T07:54:38.979374446Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:54:39.089191 containerd[1597]: time="2024-10-09T07:54:39.089108564Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-m9zpt,Uid:35e137ea-8f33-425b-a35d-8ab05164f05b,Namespace:kube-system,Attempt:0,} returns sandbox id \"305614022ff8f2ec536c0ebbb01f48e0f6e5eef69f6d81d49542e71b30c5c9f0\"" Oct 9 07:54:39.094902 kubelet[2734]: E1009 07:54:39.092594 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:39.105068 containerd[1597]: time="2024-10-09T07:54:39.102697552Z" level=info msg="CreateContainer within sandbox \"305614022ff8f2ec536c0ebbb01f48e0f6e5eef69f6d81d49542e71b30c5c9f0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 9 07:54:39.163130 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1891699853.mount: Deactivated successfully. Oct 9 07:54:39.176793 containerd[1597]: time="2024-10-09T07:54:39.174211242Z" level=info msg="CreateContainer within sandbox \"305614022ff8f2ec536c0ebbb01f48e0f6e5eef69f6d81d49542e71b30c5c9f0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a8da1fcd5012c9c302837983998f0a74f3612b64790c9d0c03b2b8168b7ea430\"" Oct 9 07:54:39.176793 containerd[1597]: time="2024-10-09T07:54:39.175317526Z" level=info msg="StartContainer for \"a8da1fcd5012c9c302837983998f0a74f3612b64790c9d0c03b2b8168b7ea430\"" Oct 9 07:54:39.195046 containerd[1597]: time="2024-10-09T07:54:39.194234143Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-n62tj,Uid:90b51489-005a-408c-8757-54b552fddc3f,Namespace:kube-system,Attempt:0,} returns sandbox id \"b4f3b459a173e4655299bb8c9198d9948d6bdcdc1bae4d6ad3c44ffe5d6a1ed4\"" Oct 9 07:54:39.200584 kubelet[2734]: E1009 07:54:39.200119 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:39.216303 containerd[1597]: time="2024-10-09T07:54:39.216247349Z" level=info msg="CreateContainer within sandbox \"b4f3b459a173e4655299bb8c9198d9948d6bdcdc1bae4d6ad3c44ffe5d6a1ed4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 9 07:54:39.235400 containerd[1597]: time="2024-10-09T07:54:39.235246354Z" level=info msg="CreateContainer within sandbox \"b4f3b459a173e4655299bb8c9198d9948d6bdcdc1bae4d6ad3c44ffe5d6a1ed4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"55cf7c6a310e9b7fb88bbc0967255a2a5a10b1fe3e39c3b6ca7eb2013dcd8432\"" Oct 9 07:54:39.239685 containerd[1597]: time="2024-10-09T07:54:39.238374555Z" level=info msg="StartContainer for \"55cf7c6a310e9b7fb88bbc0967255a2a5a10b1fe3e39c3b6ca7eb2013dcd8432\"" Oct 9 07:54:39.316123 containerd[1597]: time="2024-10-09T07:54:39.315786936Z" level=info msg="StartContainer for \"a8da1fcd5012c9c302837983998f0a74f3612b64790c9d0c03b2b8168b7ea430\" returns successfully" Oct 9 07:54:39.337243 containerd[1597]: time="2024-10-09T07:54:39.337179537Z" level=info msg="StartContainer for \"55cf7c6a310e9b7fb88bbc0967255a2a5a10b1fe3e39c3b6ca7eb2013dcd8432\" returns successfully" Oct 9 07:54:39.774770 kubelet[2734]: E1009 07:54:39.772921 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:39.777753 kubelet[2734]: E1009 07:54:39.777392 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:39.798331 kubelet[2734]: I1009 07:54:39.798145 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-m9zpt" podStartSLOduration=23.798070061 podStartE2EDuration="23.798070061s" podCreationTimestamp="2024-10-09 07:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-09 07:54:39.795438993 +0000 UTC m=+35.415973489" watchObservedRunningTime="2024-10-09 07:54:39.798070061 +0000 UTC m=+35.418604484" Oct 9 07:54:39.837733 kubelet[2734]: I1009 07:54:39.835969 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-n62tj" podStartSLOduration=23.835923835 podStartE2EDuration="23.835923835s" podCreationTimestamp="2024-10-09 07:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-09 07:54:39.81323133 +0000 UTC m=+35.433765749" watchObservedRunningTime="2024-10-09 07:54:39.835923835 +0000 UTC m=+35.456458250" Oct 9 07:54:40.782643 kubelet[2734]: E1009 07:54:40.780315 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:40.786908 kubelet[2734]: E1009 07:54:40.786860 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:41.781168 kubelet[2734]: E1009 07:54:41.781134 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:42.995077 systemd[1]: Started sshd@7-143.198.138.163:22-139.178.89.65:53236.service - OpenSSH per-connection server daemon (139.178.89.65:53236). Oct 9 07:54:43.077675 sshd[4118]: Accepted publickey for core from 139.178.89.65 port 53236 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:54:43.080172 sshd[4118]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:54:43.094767 systemd-logind[1572]: New session 8 of user core. Oct 9 07:54:43.099347 systemd[1]: Started session-8.scope - Session 8 of User core. Oct 9 07:54:43.676828 sshd[4118]: pam_unix(sshd:session): session closed for user core Oct 9 07:54:43.685140 systemd-logind[1572]: Session 8 logged out. Waiting for processes to exit. Oct 9 07:54:43.687144 systemd[1]: sshd@7-143.198.138.163:22-139.178.89.65:53236.service: Deactivated successfully. Oct 9 07:54:43.692218 systemd[1]: session-8.scope: Deactivated successfully. Oct 9 07:54:43.696094 systemd-logind[1572]: Removed session 8. Oct 9 07:54:45.638962 kubelet[2734]: I1009 07:54:45.638687 2734 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 9 07:54:45.640166 kubelet[2734]: E1009 07:54:45.639666 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:45.811825 kubelet[2734]: E1009 07:54:45.811578 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:54:48.688238 systemd[1]: Started sshd@8-143.198.138.163:22-139.178.89.65:33132.service - OpenSSH per-connection server daemon (139.178.89.65:33132). Oct 9 07:54:48.745899 sshd[4134]: Accepted publickey for core from 139.178.89.65 port 33132 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:54:48.748269 sshd[4134]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:54:48.755003 systemd-logind[1572]: New session 9 of user core. Oct 9 07:54:48.761620 systemd[1]: Started session-9.scope - Session 9 of User core. Oct 9 07:54:48.914725 sshd[4134]: pam_unix(sshd:session): session closed for user core Oct 9 07:54:48.921354 systemd[1]: sshd@8-143.198.138.163:22-139.178.89.65:33132.service: Deactivated successfully. Oct 9 07:54:48.926247 systemd[1]: session-9.scope: Deactivated successfully. Oct 9 07:54:48.927331 systemd-logind[1572]: Session 9 logged out. Waiting for processes to exit. Oct 9 07:54:48.928884 systemd-logind[1572]: Removed session 9. Oct 9 07:54:53.926180 systemd[1]: Started sshd@9-143.198.138.163:22-139.178.89.65:33146.service - OpenSSH per-connection server daemon (139.178.89.65:33146). Oct 9 07:54:53.996904 sshd[4149]: Accepted publickey for core from 139.178.89.65 port 33146 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:54:53.998452 sshd[4149]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:54:54.003868 systemd-logind[1572]: New session 10 of user core. Oct 9 07:54:54.010494 systemd[1]: Started session-10.scope - Session 10 of User core. Oct 9 07:54:54.159006 sshd[4149]: pam_unix(sshd:session): session closed for user core Oct 9 07:54:54.163507 systemd[1]: sshd@9-143.198.138.163:22-139.178.89.65:33146.service: Deactivated successfully. Oct 9 07:54:54.169008 systemd-logind[1572]: Session 10 logged out. Waiting for processes to exit. Oct 9 07:54:54.169102 systemd[1]: session-10.scope: Deactivated successfully. Oct 9 07:54:54.172905 systemd-logind[1572]: Removed session 10. Oct 9 07:54:59.172185 systemd[1]: Started sshd@10-143.198.138.163:22-139.178.89.65:60382.service - OpenSSH per-connection server daemon (139.178.89.65:60382). Oct 9 07:54:59.217143 sshd[4164]: Accepted publickey for core from 139.178.89.65 port 60382 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:54:59.219556 sshd[4164]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:54:59.225612 systemd-logind[1572]: New session 11 of user core. Oct 9 07:54:59.241369 systemd[1]: Started session-11.scope - Session 11 of User core. Oct 9 07:54:59.406021 sshd[4164]: pam_unix(sshd:session): session closed for user core Oct 9 07:54:59.416208 systemd[1]: Started sshd@11-143.198.138.163:22-139.178.89.65:60390.service - OpenSSH per-connection server daemon (139.178.89.65:60390). Oct 9 07:54:59.419367 systemd[1]: sshd@10-143.198.138.163:22-139.178.89.65:60382.service: Deactivated successfully. Oct 9 07:54:59.425133 systemd[1]: session-11.scope: Deactivated successfully. Oct 9 07:54:59.427119 systemd-logind[1572]: Session 11 logged out. Waiting for processes to exit. Oct 9 07:54:59.431870 systemd-logind[1572]: Removed session 11. Oct 9 07:54:59.478599 sshd[4176]: Accepted publickey for core from 139.178.89.65 port 60390 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:54:59.481322 sshd[4176]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:54:59.490052 systemd-logind[1572]: New session 12 of user core. Oct 9 07:54:59.495412 systemd[1]: Started session-12.scope - Session 12 of User core. Oct 9 07:54:59.730184 sshd[4176]: pam_unix(sshd:session): session closed for user core Oct 9 07:54:59.742215 systemd[1]: Started sshd@12-143.198.138.163:22-139.178.89.65:60406.service - OpenSSH per-connection server daemon (139.178.89.65:60406). Oct 9 07:54:59.743087 systemd[1]: sshd@11-143.198.138.163:22-139.178.89.65:60390.service: Deactivated successfully. Oct 9 07:54:59.747102 systemd[1]: session-12.scope: Deactivated successfully. Oct 9 07:54:59.750159 systemd-logind[1572]: Session 12 logged out. Waiting for processes to exit. Oct 9 07:54:59.759532 systemd-logind[1572]: Removed session 12. Oct 9 07:54:59.805804 sshd[4189]: Accepted publickey for core from 139.178.89.65 port 60406 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:54:59.808616 sshd[4189]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:54:59.821857 systemd-logind[1572]: New session 13 of user core. Oct 9 07:54:59.828170 systemd[1]: Started session-13.scope - Session 13 of User core. Oct 9 07:54:59.994736 sshd[4189]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:00.001194 systemd[1]: sshd@12-143.198.138.163:22-139.178.89.65:60406.service: Deactivated successfully. Oct 9 07:55:00.009039 systemd-logind[1572]: Session 13 logged out. Waiting for processes to exit. Oct 9 07:55:00.010227 systemd[1]: session-13.scope: Deactivated successfully. Oct 9 07:55:00.012492 systemd-logind[1572]: Removed session 13. Oct 9 07:55:05.006965 systemd[1]: Started sshd@13-143.198.138.163:22-139.178.89.65:35408.service - OpenSSH per-connection server daemon (139.178.89.65:35408). Oct 9 07:55:05.076389 sshd[4208]: Accepted publickey for core from 139.178.89.65 port 35408 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:05.079020 sshd[4208]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:05.089360 systemd-logind[1572]: New session 14 of user core. Oct 9 07:55:05.098660 systemd[1]: Started session-14.scope - Session 14 of User core. Oct 9 07:55:05.282027 sshd[4208]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:05.290350 systemd[1]: sshd@13-143.198.138.163:22-139.178.89.65:35408.service: Deactivated successfully. Oct 9 07:55:05.297220 systemd[1]: session-14.scope: Deactivated successfully. Oct 9 07:55:05.299332 systemd-logind[1572]: Session 14 logged out. Waiting for processes to exit. Oct 9 07:55:05.301445 systemd-logind[1572]: Removed session 14. Oct 9 07:55:10.296226 systemd[1]: Started sshd@14-143.198.138.163:22-139.178.89.65:35418.service - OpenSSH per-connection server daemon (139.178.89.65:35418). Oct 9 07:55:10.347839 sshd[4222]: Accepted publickey for core from 139.178.89.65 port 35418 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:10.350220 sshd[4222]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:10.358099 systemd-logind[1572]: New session 15 of user core. Oct 9 07:55:10.369260 systemd[1]: Started session-15.scope - Session 15 of User core. Oct 9 07:55:10.512666 sshd[4222]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:10.519231 systemd[1]: sshd@14-143.198.138.163:22-139.178.89.65:35418.service: Deactivated successfully. Oct 9 07:55:10.524182 systemd-logind[1572]: Session 15 logged out. Waiting for processes to exit. Oct 9 07:55:10.524856 systemd[1]: session-15.scope: Deactivated successfully. Oct 9 07:55:10.526690 systemd-logind[1572]: Removed session 15. Oct 9 07:55:15.521167 systemd[1]: Started sshd@15-143.198.138.163:22-139.178.89.65:45966.service - OpenSSH per-connection server daemon (139.178.89.65:45966). Oct 9 07:55:15.579756 sshd[4236]: Accepted publickey for core from 139.178.89.65 port 45966 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:15.580927 kubelet[2734]: E1009 07:55:15.580850 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:15.582478 sshd[4236]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:15.589509 systemd-logind[1572]: New session 16 of user core. Oct 9 07:55:15.599240 systemd[1]: Started session-16.scope - Session 16 of User core. Oct 9 07:55:15.748523 sshd[4236]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:15.760551 systemd[1]: Started sshd@16-143.198.138.163:22-139.178.89.65:45974.service - OpenSSH per-connection server daemon (139.178.89.65:45974). Oct 9 07:55:15.761645 systemd[1]: sshd@15-143.198.138.163:22-139.178.89.65:45966.service: Deactivated successfully. Oct 9 07:55:15.767950 systemd[1]: session-16.scope: Deactivated successfully. Oct 9 07:55:15.770354 systemd-logind[1572]: Session 16 logged out. Waiting for processes to exit. Oct 9 07:55:15.774424 systemd-logind[1572]: Removed session 16. Oct 9 07:55:15.819816 sshd[4248]: Accepted publickey for core from 139.178.89.65 port 45974 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:15.822107 sshd[4248]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:15.828643 systemd-logind[1572]: New session 17 of user core. Oct 9 07:55:15.835124 systemd[1]: Started session-17.scope - Session 17 of User core. Oct 9 07:55:16.155411 sshd[4248]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:16.169037 systemd[1]: Started sshd@17-143.198.138.163:22-139.178.89.65:45982.service - OpenSSH per-connection server daemon (139.178.89.65:45982). Oct 9 07:55:16.172044 systemd[1]: sshd@16-143.198.138.163:22-139.178.89.65:45974.service: Deactivated successfully. Oct 9 07:55:16.177567 systemd[1]: session-17.scope: Deactivated successfully. Oct 9 07:55:16.178887 systemd-logind[1572]: Session 17 logged out. Waiting for processes to exit. Oct 9 07:55:16.181882 systemd-logind[1572]: Removed session 17. Oct 9 07:55:16.223934 sshd[4259]: Accepted publickey for core from 139.178.89.65 port 45982 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:16.226072 sshd[4259]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:16.232395 systemd-logind[1572]: New session 18 of user core. Oct 9 07:55:16.237116 systemd[1]: Started session-18.scope - Session 18 of User core. Oct 9 07:55:18.010447 sshd[4259]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:18.033096 systemd[1]: Started sshd@18-143.198.138.163:22-139.178.89.65:45992.service - OpenSSH per-connection server daemon (139.178.89.65:45992). Oct 9 07:55:18.033659 systemd[1]: sshd@17-143.198.138.163:22-139.178.89.65:45982.service: Deactivated successfully. Oct 9 07:55:18.047087 systemd-logind[1572]: Session 18 logged out. Waiting for processes to exit. Oct 9 07:55:18.048683 systemd[1]: session-18.scope: Deactivated successfully. Oct 9 07:55:18.051345 systemd-logind[1572]: Removed session 18. Oct 9 07:55:18.146543 sshd[4279]: Accepted publickey for core from 139.178.89.65 port 45992 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:18.148527 sshd[4279]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:18.155808 systemd-logind[1572]: New session 19 of user core. Oct 9 07:55:18.164286 systemd[1]: Started session-19.scope - Session 19 of User core. Oct 9 07:55:18.540683 sshd[4279]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:18.563433 systemd[1]: Started sshd@19-143.198.138.163:22-139.178.89.65:46002.service - OpenSSH per-connection server daemon (139.178.89.65:46002). Oct 9 07:55:18.568301 systemd[1]: sshd@18-143.198.138.163:22-139.178.89.65:45992.service: Deactivated successfully. Oct 9 07:55:18.571490 systemd[1]: session-19.scope: Deactivated successfully. Oct 9 07:55:18.577090 systemd-logind[1572]: Session 19 logged out. Waiting for processes to exit. Oct 9 07:55:18.582106 systemd-logind[1572]: Removed session 19. Oct 9 07:55:18.618996 sshd[4292]: Accepted publickey for core from 139.178.89.65 port 46002 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:18.621131 sshd[4292]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:18.630848 systemd-logind[1572]: New session 20 of user core. Oct 9 07:55:18.639419 systemd[1]: Started session-20.scope - Session 20 of User core. Oct 9 07:55:18.790959 sshd[4292]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:18.795314 systemd[1]: sshd@19-143.198.138.163:22-139.178.89.65:46002.service: Deactivated successfully. Oct 9 07:55:18.802513 systemd[1]: session-20.scope: Deactivated successfully. Oct 9 07:55:18.803692 systemd-logind[1572]: Session 20 logged out. Waiting for processes to exit. Oct 9 07:55:18.804987 systemd-logind[1572]: Removed session 20. Oct 9 07:55:23.801180 systemd[1]: Started sshd@20-143.198.138.163:22-139.178.89.65:46006.service - OpenSSH per-connection server daemon (139.178.89.65:46006). Oct 9 07:55:23.861873 sshd[4308]: Accepted publickey for core from 139.178.89.65 port 46006 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:23.864315 sshd[4308]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:23.871786 systemd-logind[1572]: New session 21 of user core. Oct 9 07:55:23.880641 systemd[1]: Started session-21.scope - Session 21 of User core. Oct 9 07:55:24.028428 sshd[4308]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:24.032423 systemd[1]: sshd@20-143.198.138.163:22-139.178.89.65:46006.service: Deactivated successfully. Oct 9 07:55:24.039509 systemd[1]: session-21.scope: Deactivated successfully. Oct 9 07:55:24.042025 systemd-logind[1572]: Session 21 logged out. Waiting for processes to exit. Oct 9 07:55:24.043668 systemd-logind[1572]: Removed session 21. Oct 9 07:55:29.043670 systemd[1]: Started sshd@21-143.198.138.163:22-139.178.89.65:41678.service - OpenSSH per-connection server daemon (139.178.89.65:41678). Oct 9 07:55:29.099051 sshd[4326]: Accepted publickey for core from 139.178.89.65 port 41678 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:29.102520 sshd[4326]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:29.111082 systemd-logind[1572]: New session 22 of user core. Oct 9 07:55:29.118341 systemd[1]: Started session-22.scope - Session 22 of User core. Oct 9 07:55:29.268413 sshd[4326]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:29.276777 systemd[1]: sshd@21-143.198.138.163:22-139.178.89.65:41678.service: Deactivated successfully. Oct 9 07:55:29.276908 systemd-logind[1572]: Session 22 logged out. Waiting for processes to exit. Oct 9 07:55:29.285183 systemd[1]: session-22.scope: Deactivated successfully. Oct 9 07:55:29.288404 systemd-logind[1572]: Removed session 22. Oct 9 07:55:34.280201 systemd[1]: Started sshd@22-143.198.138.163:22-139.178.89.65:41692.service - OpenSSH per-connection server daemon (139.178.89.65:41692). Oct 9 07:55:34.331378 sshd[4340]: Accepted publickey for core from 139.178.89.65 port 41692 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:34.334235 sshd[4340]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:34.339771 systemd-logind[1572]: New session 23 of user core. Oct 9 07:55:34.345229 systemd[1]: Started session-23.scope - Session 23 of User core. Oct 9 07:55:34.490994 sshd[4340]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:34.498198 systemd-logind[1572]: Session 23 logged out. Waiting for processes to exit. Oct 9 07:55:34.498371 systemd[1]: sshd@22-143.198.138.163:22-139.178.89.65:41692.service: Deactivated successfully. Oct 9 07:55:34.505519 systemd[1]: session-23.scope: Deactivated successfully. Oct 9 07:55:34.507194 systemd-logind[1572]: Removed session 23. Oct 9 07:55:36.581819 kubelet[2734]: E1009 07:55:36.581309 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:36.582821 kubelet[2734]: E1009 07:55:36.582421 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:39.499301 systemd[1]: Started sshd@23-143.198.138.163:22-139.178.89.65:42158.service - OpenSSH per-connection server daemon (139.178.89.65:42158). Oct 9 07:55:39.544784 sshd[4354]: Accepted publickey for core from 139.178.89.65 port 42158 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:39.545902 sshd[4354]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:39.552846 systemd-logind[1572]: New session 24 of user core. Oct 9 07:55:39.556150 systemd[1]: Started session-24.scope - Session 24 of User core. Oct 9 07:55:39.691963 sshd[4354]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:39.696005 systemd[1]: sshd@23-143.198.138.163:22-139.178.89.65:42158.service: Deactivated successfully. Oct 9 07:55:39.702224 systemd[1]: session-24.scope: Deactivated successfully. Oct 9 07:55:39.703839 systemd-logind[1572]: Session 24 logged out. Waiting for processes to exit. Oct 9 07:55:39.709113 systemd[1]: Started sshd@24-143.198.138.163:22-139.178.89.65:42164.service - OpenSSH per-connection server daemon (139.178.89.65:42164). Oct 9 07:55:39.711409 systemd-logind[1572]: Removed session 24. Oct 9 07:55:39.766547 sshd[4368]: Accepted publickey for core from 139.178.89.65 port 42164 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:39.768917 sshd[4368]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:39.777026 systemd-logind[1572]: New session 25 of user core. Oct 9 07:55:39.782747 systemd[1]: Started session-25.scope - Session 25 of User core. Oct 9 07:55:41.362226 containerd[1597]: time="2024-10-09T07:55:41.362144413Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 9 07:55:41.380404 containerd[1597]: time="2024-10-09T07:55:41.380195399Z" level=info msg="StopContainer for \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\" with timeout 2 (s)" Oct 9 07:55:41.380404 containerd[1597]: time="2024-10-09T07:55:41.380264207Z" level=info msg="StopContainer for \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\" with timeout 30 (s)" Oct 9 07:55:41.382434 containerd[1597]: time="2024-10-09T07:55:41.382295473Z" level=info msg="Stop container \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\" with signal terminated" Oct 9 07:55:41.382434 containerd[1597]: time="2024-10-09T07:55:41.382346563Z" level=info msg="Stop container \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\" with signal terminated" Oct 9 07:55:41.399056 systemd-networkd[1226]: lxc_health: Link DOWN Oct 9 07:55:41.399067 systemd-networkd[1226]: lxc_health: Lost carrier Oct 9 07:55:41.461044 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f-rootfs.mount: Deactivated successfully. Oct 9 07:55:41.467878 containerd[1597]: time="2024-10-09T07:55:41.467646239Z" level=info msg="shim disconnected" id=679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f namespace=k8s.io Oct 9 07:55:41.467878 containerd[1597]: time="2024-10-09T07:55:41.467843709Z" level=warning msg="cleaning up after shim disconnected" id=679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f namespace=k8s.io Oct 9 07:55:41.468492 containerd[1597]: time="2024-10-09T07:55:41.467858378Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:41.473838 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834-rootfs.mount: Deactivated successfully. Oct 9 07:55:41.476662 containerd[1597]: time="2024-10-09T07:55:41.476590405Z" level=info msg="shim disconnected" id=f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834 namespace=k8s.io Oct 9 07:55:41.476928 containerd[1597]: time="2024-10-09T07:55:41.476746517Z" level=warning msg="cleaning up after shim disconnected" id=f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834 namespace=k8s.io Oct 9 07:55:41.476928 containerd[1597]: time="2024-10-09T07:55:41.476760650Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:41.500608 containerd[1597]: time="2024-10-09T07:55:41.500481300Z" level=info msg="StopContainer for \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\" returns successfully" Oct 9 07:55:41.509756 containerd[1597]: time="2024-10-09T07:55:41.508112151Z" level=warning msg="cleanup warnings time=\"2024-10-09T07:55:41Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Oct 9 07:55:41.515768 containerd[1597]: time="2024-10-09T07:55:41.515319203Z" level=info msg="StopContainer for \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\" returns successfully" Oct 9 07:55:41.515768 containerd[1597]: time="2024-10-09T07:55:41.515610529Z" level=info msg="StopPodSandbox for \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\"" Oct 9 07:55:41.515768 containerd[1597]: time="2024-10-09T07:55:41.515670955Z" level=info msg="Container to stop \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 9 07:55:41.515768 containerd[1597]: time="2024-10-09T07:55:41.515684001Z" level=info msg="Container to stop \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 9 07:55:41.515768 containerd[1597]: time="2024-10-09T07:55:41.515695351Z" level=info msg="Container to stop \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 9 07:55:41.515768 containerd[1597]: time="2024-10-09T07:55:41.515719484Z" level=info msg="Container to stop \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 9 07:55:41.515768 containerd[1597]: time="2024-10-09T07:55:41.515729527Z" level=info msg="Container to stop \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 9 07:55:41.520048 containerd[1597]: time="2024-10-09T07:55:41.517692224Z" level=info msg="StopPodSandbox for \"4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a\"" Oct 9 07:55:41.520790 containerd[1597]: time="2024-10-09T07:55:41.520742793Z" level=info msg="Container to stop \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Oct 9 07:55:41.521092 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434-shm.mount: Deactivated successfully. Oct 9 07:55:41.527640 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a-shm.mount: Deactivated successfully. Oct 9 07:55:41.597157 containerd[1597]: time="2024-10-09T07:55:41.596665838Z" level=info msg="shim disconnected" id=3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434 namespace=k8s.io Oct 9 07:55:41.597157 containerd[1597]: time="2024-10-09T07:55:41.596760172Z" level=warning msg="cleaning up after shim disconnected" id=3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434 namespace=k8s.io Oct 9 07:55:41.597157 containerd[1597]: time="2024-10-09T07:55:41.596774091Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:41.599329 containerd[1597]: time="2024-10-09T07:55:41.597333988Z" level=info msg="shim disconnected" id=4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a namespace=k8s.io Oct 9 07:55:41.599329 containerd[1597]: time="2024-10-09T07:55:41.599020676Z" level=warning msg="cleaning up after shim disconnected" id=4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a namespace=k8s.io Oct 9 07:55:41.599329 containerd[1597]: time="2024-10-09T07:55:41.599036138Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:41.625835 containerd[1597]: time="2024-10-09T07:55:41.625401499Z" level=warning msg="cleanup warnings time=\"2024-10-09T07:55:41Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Oct 9 07:55:41.628695 containerd[1597]: time="2024-10-09T07:55:41.628505880Z" level=info msg="TearDown network for sandbox \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" successfully" Oct 9 07:55:41.628695 containerd[1597]: time="2024-10-09T07:55:41.628547885Z" level=info msg="StopPodSandbox for \"3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434\" returns successfully" Oct 9 07:55:41.636648 containerd[1597]: time="2024-10-09T07:55:41.634915138Z" level=info msg="TearDown network for sandbox \"4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a\" successfully" Oct 9 07:55:41.636648 containerd[1597]: time="2024-10-09T07:55:41.634943637Z" level=info msg="StopPodSandbox for \"4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a\" returns successfully" Oct 9 07:55:41.761530 kubelet[2734]: I1009 07:55:41.761439 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hubble-tls\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.761530 kubelet[2734]: I1009 07:55:41.761515 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-bpf-maps\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762130 kubelet[2734]: I1009 07:55:41.761555 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3481fcc3-e066-4e00-a996-ec6227b2a6a2-clustermesh-secrets\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762130 kubelet[2734]: I1009 07:55:41.761577 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkbdq\" (UniqueName: \"kubernetes.io/projected/515d30f1-a4f9-465b-95eb-b1a89e8c318f-kube-api-access-qkbdq\") pod \"515d30f1-a4f9-465b-95eb-b1a89e8c318f\" (UID: \"515d30f1-a4f9-465b-95eb-b1a89e8c318f\") " Oct 9 07:55:41.762130 kubelet[2734]: I1009 07:55:41.761598 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnn67\" (UniqueName: \"kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-kube-api-access-jnn67\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762130 kubelet[2734]: I1009 07:55:41.761619 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-config-path\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762130 kubelet[2734]: I1009 07:55:41.761640 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hostproc\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762130 kubelet[2734]: I1009 07:55:41.761660 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-net\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762300 kubelet[2734]: I1009 07:55:41.761682 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-etc-cni-netd\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762300 kubelet[2734]: I1009 07:55:41.761737 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-xtables-lock\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762300 kubelet[2734]: I1009 07:55:41.761768 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-lib-modules\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.762300 kubelet[2734]: I1009 07:55:41.761793 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-kernel\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.763218 kubelet[2734]: I1009 07:55:41.762813 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/515d30f1-a4f9-465b-95eb-b1a89e8c318f-cilium-config-path\") pod \"515d30f1-a4f9-465b-95eb-b1a89e8c318f\" (UID: \"515d30f1-a4f9-465b-95eb-b1a89e8c318f\") " Oct 9 07:55:41.763218 kubelet[2734]: I1009 07:55:41.762908 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cni-path\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.763218 kubelet[2734]: I1009 07:55:41.762931 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-run\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.763218 kubelet[2734]: I1009 07:55:41.762951 2734 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-cgroup\") pod \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\" (UID: \"3481fcc3-e066-4e00-a996-ec6227b2a6a2\") " Oct 9 07:55:41.764320 kubelet[2734]: I1009 07:55:41.763105 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hostproc" (OuterVolumeSpecName: "hostproc") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.764985 kubelet[2734]: I1009 07:55:41.764524 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 9 07:55:41.764985 kubelet[2734]: I1009 07:55:41.764641 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.764985 kubelet[2734]: I1009 07:55:41.764664 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.764985 kubelet[2734]: I1009 07:55:41.764680 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.764985 kubelet[2734]: I1009 07:55:41.764695 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.765193 kubelet[2734]: I1009 07:55:41.764724 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.767804 kubelet[2734]: I1009 07:55:41.767758 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/515d30f1-a4f9-465b-95eb-b1a89e8c318f-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "515d30f1-a4f9-465b-95eb-b1a89e8c318f" (UID: "515d30f1-a4f9-465b-95eb-b1a89e8c318f"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 9 07:55:41.768394 kubelet[2734]: I1009 07:55:41.768009 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cni-path" (OuterVolumeSpecName: "cni-path") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.768394 kubelet[2734]: I1009 07:55:41.768039 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.768394 kubelet[2734]: I1009 07:55:41.768054 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.769059 kubelet[2734]: I1009 07:55:41.769023 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 9 07:55:41.771792 kubelet[2734]: I1009 07:55:41.771746 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3481fcc3-e066-4e00-a996-ec6227b2a6a2-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 9 07:55:41.771997 kubelet[2734]: I1009 07:55:41.771982 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 9 07:55:41.772600 kubelet[2734]: I1009 07:55:41.772557 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/515d30f1-a4f9-465b-95eb-b1a89e8c318f-kube-api-access-qkbdq" (OuterVolumeSpecName: "kube-api-access-qkbdq") pod "515d30f1-a4f9-465b-95eb-b1a89e8c318f" (UID: "515d30f1-a4f9-465b-95eb-b1a89e8c318f"). InnerVolumeSpecName "kube-api-access-qkbdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 9 07:55:41.775297 kubelet[2734]: I1009 07:55:41.775242 2734 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-kube-api-access-jnn67" (OuterVolumeSpecName: "kube-api-access-jnn67") pod "3481fcc3-e066-4e00-a996-ec6227b2a6a2" (UID: "3481fcc3-e066-4e00-a996-ec6227b2a6a2"). InnerVolumeSpecName "kube-api-access-jnn67". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 9 07:55:41.863746 kubelet[2734]: I1009 07:55:41.863616 2734 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hostproc\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.863746 kubelet[2734]: I1009 07:55:41.863688 2734 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-net\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.863746 kubelet[2734]: I1009 07:55:41.863699 2734 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-etc-cni-netd\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.863746 kubelet[2734]: I1009 07:55:41.863729 2734 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-xtables-lock\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.863746 kubelet[2734]: I1009 07:55:41.863742 2734 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-lib-modules\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.863746 kubelet[2734]: I1009 07:55:41.863753 2734 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-host-proc-sys-kernel\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.863746 kubelet[2734]: I1009 07:55:41.863763 2734 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/515d30f1-a4f9-465b-95eb-b1a89e8c318f-cilium-config-path\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863775 2734 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-run\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863787 2734 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-cgroup\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863796 2734 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cni-path\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863805 2734 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-hubble-tls\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863816 2734 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3481fcc3-e066-4e00-a996-ec6227b2a6a2-bpf-maps\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863828 2734 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3481fcc3-e066-4e00-a996-ec6227b2a6a2-clustermesh-secrets\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863838 2734 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-qkbdq\" (UniqueName: \"kubernetes.io/projected/515d30f1-a4f9-465b-95eb-b1a89e8c318f-kube-api-access-qkbdq\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864123 kubelet[2734]: I1009 07:55:41.863849 2734 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3481fcc3-e066-4e00-a996-ec6227b2a6a2-cilium-config-path\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.864318 kubelet[2734]: I1009 07:55:41.863864 2734 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-jnn67\" (UniqueName: \"kubernetes.io/projected/3481fcc3-e066-4e00-a996-ec6227b2a6a2-kube-api-access-jnn67\") on node \"ci-4081.1.0-c-f399a429fa\" DevicePath \"\"" Oct 9 07:55:41.974941 kubelet[2734]: I1009 07:55:41.974569 2734 scope.go:117] "RemoveContainer" containerID="f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834" Oct 9 07:55:41.980754 containerd[1597]: time="2024-10-09T07:55:41.980582817Z" level=info msg="RemoveContainer for \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\"" Oct 9 07:55:41.986562 containerd[1597]: time="2024-10-09T07:55:41.986372172Z" level=info msg="RemoveContainer for \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\" returns successfully" Oct 9 07:55:41.994928 kubelet[2734]: I1009 07:55:41.994883 2734 scope.go:117] "RemoveContainer" containerID="f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834" Oct 9 07:55:42.024938 containerd[1597]: time="2024-10-09T07:55:41.995477843Z" level=error msg="ContainerStatus for \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\": not found" Oct 9 07:55:42.047869 kubelet[2734]: E1009 07:55:42.047818 2734 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\": not found" containerID="f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834" Oct 9 07:55:42.059667 kubelet[2734]: I1009 07:55:42.059506 2734 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834"} err="failed to get container status \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\": rpc error: code = NotFound desc = an error occurred when try to find container \"f92f1fa9f5942af9a07e61d6d7d4583bdd81805f36e1fb6fdfc14d6e351f1834\": not found" Oct 9 07:55:42.059667 kubelet[2734]: I1009 07:55:42.059565 2734 scope.go:117] "RemoveContainer" containerID="679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f" Oct 9 07:55:42.061957 containerd[1597]: time="2024-10-09T07:55:42.061897909Z" level=info msg="RemoveContainer for \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\"" Oct 9 07:55:42.066217 containerd[1597]: time="2024-10-09T07:55:42.065963674Z" level=info msg="RemoveContainer for \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\" returns successfully" Oct 9 07:55:42.069455 kubelet[2734]: I1009 07:55:42.069297 2734 scope.go:117] "RemoveContainer" containerID="89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65" Oct 9 07:55:42.072308 containerd[1597]: time="2024-10-09T07:55:42.072058969Z" level=info msg="RemoveContainer for \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\"" Oct 9 07:55:42.091900 containerd[1597]: time="2024-10-09T07:55:42.091122232Z" level=info msg="RemoveContainer for \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\" returns successfully" Oct 9 07:55:42.092318 kubelet[2734]: I1009 07:55:42.092262 2734 scope.go:117] "RemoveContainer" containerID="8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46" Oct 9 07:55:42.094742 containerd[1597]: time="2024-10-09T07:55:42.094649140Z" level=info msg="RemoveContainer for \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\"" Oct 9 07:55:42.098140 containerd[1597]: time="2024-10-09T07:55:42.098038434Z" level=info msg="RemoveContainer for \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\" returns successfully" Oct 9 07:55:42.098655 kubelet[2734]: I1009 07:55:42.098378 2734 scope.go:117] "RemoveContainer" containerID="cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470" Oct 9 07:55:42.100634 containerd[1597]: time="2024-10-09T07:55:42.100585259Z" level=info msg="RemoveContainer for \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\"" Oct 9 07:55:42.103899 containerd[1597]: time="2024-10-09T07:55:42.103844535Z" level=info msg="RemoveContainer for \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\" returns successfully" Oct 9 07:55:42.104239 kubelet[2734]: I1009 07:55:42.104208 2734 scope.go:117] "RemoveContainer" containerID="a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a" Oct 9 07:55:42.106082 containerd[1597]: time="2024-10-09T07:55:42.106029924Z" level=info msg="RemoveContainer for \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\"" Oct 9 07:55:42.109480 containerd[1597]: time="2024-10-09T07:55:42.109413829Z" level=info msg="RemoveContainer for \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\" returns successfully" Oct 9 07:55:42.109979 kubelet[2734]: I1009 07:55:42.109765 2734 scope.go:117] "RemoveContainer" containerID="679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f" Oct 9 07:55:42.110313 containerd[1597]: time="2024-10-09T07:55:42.110258311Z" level=error msg="ContainerStatus for \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\": not found" Oct 9 07:55:42.110585 kubelet[2734]: E1009 07:55:42.110477 2734 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\": not found" containerID="679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f" Oct 9 07:55:42.110862 kubelet[2734]: I1009 07:55:42.110701 2734 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f"} err="failed to get container status \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\": rpc error: code = NotFound desc = an error occurred when try to find container \"679e13b615a2034898abfe5f5bfff6f0571f08b769ef2137a79194497b1eb05f\": not found" Oct 9 07:55:42.110862 kubelet[2734]: I1009 07:55:42.110762 2734 scope.go:117] "RemoveContainer" containerID="89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65" Oct 9 07:55:42.111184 containerd[1597]: time="2024-10-09T07:55:42.111136034Z" level=error msg="ContainerStatus for \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\": not found" Oct 9 07:55:42.111378 kubelet[2734]: E1009 07:55:42.111355 2734 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\": not found" containerID="89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65" Oct 9 07:55:42.111445 kubelet[2734]: I1009 07:55:42.111403 2734 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65"} err="failed to get container status \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\": rpc error: code = NotFound desc = an error occurred when try to find container \"89a5c85d09ef7800b4ff150059c11d16f933e3694f3d70e50e23dcf830e1cc65\": not found" Oct 9 07:55:42.111445 kubelet[2734]: I1009 07:55:42.111419 2734 scope.go:117] "RemoveContainer" containerID="8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46" Oct 9 07:55:42.111753 containerd[1597]: time="2024-10-09T07:55:42.111677496Z" level=error msg="ContainerStatus for \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\": not found" Oct 9 07:55:42.112022 kubelet[2734]: E1009 07:55:42.111906 2734 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\": not found" containerID="8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46" Oct 9 07:55:42.112022 kubelet[2734]: I1009 07:55:42.111936 2734 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46"} err="failed to get container status \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\": rpc error: code = NotFound desc = an error occurred when try to find container \"8726b1d35edfa1b087873cfbd76c85cec9299f6ef10ee349e5aca2822f6d3d46\": not found" Oct 9 07:55:42.112022 kubelet[2734]: I1009 07:55:42.111948 2734 scope.go:117] "RemoveContainer" containerID="cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470" Oct 9 07:55:42.112675 containerd[1597]: time="2024-10-09T07:55:42.112635740Z" level=error msg="ContainerStatus for \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\": not found" Oct 9 07:55:42.112957 kubelet[2734]: E1009 07:55:42.112814 2734 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\": not found" containerID="cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470" Oct 9 07:55:42.112957 kubelet[2734]: I1009 07:55:42.112844 2734 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470"} err="failed to get container status \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\": rpc error: code = NotFound desc = an error occurred when try to find container \"cdded0fa4b165465523611ea952925c9f573b1b676788c1b5c2196a5d50f9470\": not found" Oct 9 07:55:42.112957 kubelet[2734]: I1009 07:55:42.112855 2734 scope.go:117] "RemoveContainer" containerID="a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a" Oct 9 07:55:42.113296 containerd[1597]: time="2024-10-09T07:55:42.113255612Z" level=error msg="ContainerStatus for \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\": not found" Oct 9 07:55:42.113513 kubelet[2734]: E1009 07:55:42.113459 2734 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\": not found" containerID="a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a" Oct 9 07:55:42.113513 kubelet[2734]: I1009 07:55:42.113493 2734 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a"} err="failed to get container status \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\": rpc error: code = NotFound desc = an error occurred when try to find container \"a7de4e117c5cb6725528f254da37a1df6dee7a2e9b74aced1a91f06e4efa0a0a\": not found" Oct 9 07:55:42.322252 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4668bc9375977cefa7f33a7921eab4be53f1b3a37a21eb2d4675d2280b6cfc0a-rootfs.mount: Deactivated successfully. Oct 9 07:55:42.322453 systemd[1]: var-lib-kubelet-pods-515d30f1\x2da4f9\x2d465b\x2d95eb\x2db1a89e8c318f-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dqkbdq.mount: Deactivated successfully. Oct 9 07:55:42.322560 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3372dd49115dd178262f7c423e8be467f104875ca4a23a27bd2f2a9453ea4434-rootfs.mount: Deactivated successfully. Oct 9 07:55:42.322657 systemd[1]: var-lib-kubelet-pods-3481fcc3\x2de066\x2d4e00\x2da996\x2dec6227b2a6a2-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2djnn67.mount: Deactivated successfully. Oct 9 07:55:42.323451 systemd[1]: var-lib-kubelet-pods-3481fcc3\x2de066\x2d4e00\x2da996\x2dec6227b2a6a2-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Oct 9 07:55:42.323745 systemd[1]: var-lib-kubelet-pods-3481fcc3\x2de066\x2d4e00\x2da996\x2dec6227b2a6a2-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Oct 9 07:55:42.583590 kubelet[2734]: I1009 07:55:42.582911 2734 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" path="/var/lib/kubelet/pods/3481fcc3-e066-4e00-a996-ec6227b2a6a2/volumes" Oct 9 07:55:42.584851 kubelet[2734]: I1009 07:55:42.584816 2734 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="515d30f1-a4f9-465b-95eb-b1a89e8c318f" path="/var/lib/kubelet/pods/515d30f1-a4f9-465b-95eb-b1a89e8c318f/volumes" Oct 9 07:55:43.202066 sshd[4368]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:43.214312 systemd[1]: Started sshd@25-143.198.138.163:22-139.178.89.65:42168.service - OpenSSH per-connection server daemon (139.178.89.65:42168). Oct 9 07:55:43.215512 systemd[1]: sshd@24-143.198.138.163:22-139.178.89.65:42164.service: Deactivated successfully. Oct 9 07:55:43.227690 systemd[1]: session-25.scope: Deactivated successfully. Oct 9 07:55:43.229936 systemd-logind[1572]: Session 25 logged out. Waiting for processes to exit. Oct 9 07:55:43.239273 systemd-logind[1572]: Removed session 25. Oct 9 07:55:43.310799 sshd[4529]: Accepted publickey for core from 139.178.89.65 port 42168 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:43.312496 sshd[4529]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:43.320658 systemd-logind[1572]: New session 26 of user core. Oct 9 07:55:43.331314 systemd[1]: Started session-26.scope - Session 26 of User core. Oct 9 07:55:43.939979 sshd[4529]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:43.956171 systemd[1]: Started sshd@26-143.198.138.163:22-139.178.89.65:42172.service - OpenSSH per-connection server daemon (139.178.89.65:42172). Oct 9 07:55:43.960131 systemd[1]: sshd@25-143.198.138.163:22-139.178.89.65:42168.service: Deactivated successfully. Oct 9 07:55:43.975280 systemd[1]: session-26.scope: Deactivated successfully. Oct 9 07:55:43.979092 systemd-logind[1572]: Session 26 logged out. Waiting for processes to exit. Oct 9 07:55:43.987420 systemd-logind[1572]: Removed session 26. Oct 9 07:55:44.021190 sshd[4542]: Accepted publickey for core from 139.178.89.65 port 42172 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:44.025794 sshd[4542]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:44.045697 systemd-logind[1572]: New session 27 of user core. Oct 9 07:55:44.051171 systemd[1]: Started session-27.scope - Session 27 of User core. Oct 9 07:55:44.076297 kubelet[2734]: I1009 07:55:44.075848 2734 topology_manager.go:215] "Topology Admit Handler" podUID="ff12d98f-e7f0-4fcc-834d-31e54e576b5e" podNamespace="kube-system" podName="cilium-vphmh" Oct 9 07:55:44.081754 kubelet[2734]: E1009 07:55:44.081680 2734 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" containerName="mount-bpf-fs" Oct 9 07:55:44.081949 kubelet[2734]: E1009 07:55:44.081936 2734 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" containerName="cilium-agent" Oct 9 07:55:44.082090 kubelet[2734]: E1009 07:55:44.082066 2734 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" containerName="apply-sysctl-overwrites" Oct 9 07:55:44.082177 kubelet[2734]: E1009 07:55:44.082168 2734 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="515d30f1-a4f9-465b-95eb-b1a89e8c318f" containerName="cilium-operator" Oct 9 07:55:44.082254 kubelet[2734]: E1009 07:55:44.082246 2734 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" containerName="clean-cilium-state" Oct 9 07:55:44.082307 kubelet[2734]: E1009 07:55:44.082296 2734 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" containerName="mount-cgroup" Oct 9 07:55:44.082459 kubelet[2734]: I1009 07:55:44.082436 2734 memory_manager.go:354] "RemoveStaleState removing state" podUID="3481fcc3-e066-4e00-a996-ec6227b2a6a2" containerName="cilium-agent" Oct 9 07:55:44.082546 kubelet[2734]: I1009 07:55:44.082539 2734 memory_manager.go:354] "RemoveStaleState removing state" podUID="515d30f1-a4f9-465b-95eb-b1a89e8c318f" containerName="cilium-operator" Oct 9 07:55:44.124943 sshd[4542]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:44.138035 systemd[1]: Started sshd@27-143.198.138.163:22-139.178.89.65:42188.service - OpenSSH per-connection server daemon (139.178.89.65:42188). Oct 9 07:55:44.138554 systemd[1]: sshd@26-143.198.138.163:22-139.178.89.65:42172.service: Deactivated successfully. Oct 9 07:55:44.146537 systemd-logind[1572]: Session 27 logged out. Waiting for processes to exit. Oct 9 07:55:44.148015 systemd[1]: session-27.scope: Deactivated successfully. Oct 9 07:55:44.149955 systemd-logind[1572]: Removed session 27. Oct 9 07:55:44.212865 sshd[4552]: Accepted publickey for core from 139.178.89.65 port 42188 ssh2: RSA SHA256:nDg0UeSiwkxxSWtKMhQ+P+HuSx1Axr49vgnqaJCGl7o Oct 9 07:55:44.216035 sshd[4552]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 9 07:55:44.224179 systemd-logind[1572]: New session 28 of user core. Oct 9 07:55:44.232386 systemd[1]: Started session-28.scope - Session 28 of User core. Oct 9 07:55:44.283446 kubelet[2734]: I1009 07:55:44.283383 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-etc-cni-netd\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.285890 kubelet[2734]: I1009 07:55:44.285838 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-cilium-config-path\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.286017 kubelet[2734]: I1009 07:55:44.285964 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-lib-modules\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.286017 kubelet[2734]: I1009 07:55:44.285988 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-host-proc-sys-net\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.286017 kubelet[2734]: I1009 07:55:44.286014 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-cilium-run\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.287689 kubelet[2734]: I1009 07:55:44.286045 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-bpf-maps\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.287689 kubelet[2734]: I1009 07:55:44.286070 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42sbw\" (UniqueName: \"kubernetes.io/projected/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-kube-api-access-42sbw\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.292765 kubelet[2734]: I1009 07:55:44.292106 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-hostproc\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.304836 kubelet[2734]: I1009 07:55:44.304796 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-cilium-ipsec-secrets\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.305046 kubelet[2734]: I1009 07:55:44.305037 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-clustermesh-secrets\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.305196 kubelet[2734]: I1009 07:55:44.305184 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-hubble-tls\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.305304 kubelet[2734]: I1009 07:55:44.305295 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-cilium-cgroup\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.305396 kubelet[2734]: I1009 07:55:44.305385 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-cni-path\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.305484 kubelet[2734]: I1009 07:55:44.305469 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-xtables-lock\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.306744 kubelet[2734]: I1009 07:55:44.305548 2734 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ff12d98f-e7f0-4fcc-834d-31e54e576b5e-host-proc-sys-kernel\") pod \"cilium-vphmh\" (UID: \"ff12d98f-e7f0-4fcc-834d-31e54e576b5e\") " pod="kube-system/cilium-vphmh" Oct 9 07:55:44.442198 kubelet[2734]: E1009 07:55:44.442163 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:44.447732 containerd[1597]: time="2024-10-09T07:55:44.446263094Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vphmh,Uid:ff12d98f-e7f0-4fcc-834d-31e54e576b5e,Namespace:kube-system,Attempt:0,}" Oct 9 07:55:44.485233 containerd[1597]: time="2024-10-09T07:55:44.485027723Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Oct 9 07:55:44.485921 containerd[1597]: time="2024-10-09T07:55:44.485524242Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Oct 9 07:55:44.485921 containerd[1597]: time="2024-10-09T07:55:44.485574727Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:55:44.486392 containerd[1597]: time="2024-10-09T07:55:44.486260033Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Oct 9 07:55:44.549663 containerd[1597]: time="2024-10-09T07:55:44.549595473Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vphmh,Uid:ff12d98f-e7f0-4fcc-834d-31e54e576b5e,Namespace:kube-system,Attempt:0,} returns sandbox id \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\"" Oct 9 07:55:44.551252 kubelet[2734]: E1009 07:55:44.551045 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:44.554480 containerd[1597]: time="2024-10-09T07:55:44.554305326Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Oct 9 07:55:44.573332 containerd[1597]: time="2024-10-09T07:55:44.573182560Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"ff79d82d36823e6a7f994e968647989a63a0de575bb318494e8f7e672ef79701\"" Oct 9 07:55:44.576034 containerd[1597]: time="2024-10-09T07:55:44.575983721Z" level=info msg="StartContainer for \"ff79d82d36823e6a7f994e968647989a63a0de575bb318494e8f7e672ef79701\"" Oct 9 07:55:44.651026 containerd[1597]: time="2024-10-09T07:55:44.650985177Z" level=info msg="StartContainer for \"ff79d82d36823e6a7f994e968647989a63a0de575bb318494e8f7e672ef79701\" returns successfully" Oct 9 07:55:44.697499 containerd[1597]: time="2024-10-09T07:55:44.697380188Z" level=info msg="shim disconnected" id=ff79d82d36823e6a7f994e968647989a63a0de575bb318494e8f7e672ef79701 namespace=k8s.io Oct 9 07:55:44.697499 containerd[1597]: time="2024-10-09T07:55:44.697474772Z" level=warning msg="cleaning up after shim disconnected" id=ff79d82d36823e6a7f994e968647989a63a0de575bb318494e8f7e672ef79701 namespace=k8s.io Oct 9 07:55:44.698126 containerd[1597]: time="2024-10-09T07:55:44.697817753Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:44.773733 kubelet[2734]: E1009 07:55:44.773671 2734 kubelet.go:2892] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Oct 9 07:55:45.022772 kubelet[2734]: E1009 07:55:45.022107 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:45.030275 containerd[1597]: time="2024-10-09T07:55:45.030140490Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Oct 9 07:55:45.042960 containerd[1597]: time="2024-10-09T07:55:45.042914240Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"dd12e9a58317d35a818db94a6381ed086f03266dff95f5d71c91ee528abe1c75\"" Oct 9 07:55:45.043739 containerd[1597]: time="2024-10-09T07:55:45.043686627Z" level=info msg="StartContainer for \"dd12e9a58317d35a818db94a6381ed086f03266dff95f5d71c91ee528abe1c75\"" Oct 9 07:55:45.140913 containerd[1597]: time="2024-10-09T07:55:45.140022212Z" level=info msg="StartContainer for \"dd12e9a58317d35a818db94a6381ed086f03266dff95f5d71c91ee528abe1c75\" returns successfully" Oct 9 07:55:45.181206 containerd[1597]: time="2024-10-09T07:55:45.180963182Z" level=info msg="shim disconnected" id=dd12e9a58317d35a818db94a6381ed086f03266dff95f5d71c91ee528abe1c75 namespace=k8s.io Oct 9 07:55:45.181206 containerd[1597]: time="2024-10-09T07:55:45.181018423Z" level=warning msg="cleaning up after shim disconnected" id=dd12e9a58317d35a818db94a6381ed086f03266dff95f5d71c91ee528abe1c75 namespace=k8s.io Oct 9 07:55:45.181206 containerd[1597]: time="2024-10-09T07:55:45.181028561Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:45.198583 containerd[1597]: time="2024-10-09T07:55:45.198485100Z" level=warning msg="cleanup warnings time=\"2024-10-09T07:55:45Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Oct 9 07:55:46.025127 kubelet[2734]: E1009 07:55:46.025079 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:46.033423 containerd[1597]: time="2024-10-09T07:55:46.033081898Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Oct 9 07:55:46.063629 containerd[1597]: time="2024-10-09T07:55:46.063537126Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"5224d500173d3790cc968bcc1b592ae512c30660e4df530f506f5dccdf5de79f\"" Oct 9 07:55:46.067163 containerd[1597]: time="2024-10-09T07:55:46.066859833Z" level=info msg="StartContainer for \"5224d500173d3790cc968bcc1b592ae512c30660e4df530f506f5dccdf5de79f\"" Oct 9 07:55:46.147931 containerd[1597]: time="2024-10-09T07:55:46.147794315Z" level=info msg="StartContainer for \"5224d500173d3790cc968bcc1b592ae512c30660e4df530f506f5dccdf5de79f\" returns successfully" Oct 9 07:55:46.184058 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5224d500173d3790cc968bcc1b592ae512c30660e4df530f506f5dccdf5de79f-rootfs.mount: Deactivated successfully. Oct 9 07:55:46.188902 containerd[1597]: time="2024-10-09T07:55:46.188803418Z" level=info msg="shim disconnected" id=5224d500173d3790cc968bcc1b592ae512c30660e4df530f506f5dccdf5de79f namespace=k8s.io Oct 9 07:55:46.188902 containerd[1597]: time="2024-10-09T07:55:46.188903280Z" level=warning msg="cleaning up after shim disconnected" id=5224d500173d3790cc968bcc1b592ae512c30660e4df530f506f5dccdf5de79f namespace=k8s.io Oct 9 07:55:46.189143 containerd[1597]: time="2024-10-09T07:55:46.188918038Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:46.426673 kubelet[2734]: I1009 07:55:46.426498 2734 setters.go:568] "Node became not ready" node="ci-4081.1.0-c-f399a429fa" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-10-09T07:55:46Z","lastTransitionTime":"2024-10-09T07:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Oct 9 07:55:46.581807 kubelet[2734]: E1009 07:55:46.581425 2734 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-76f75df574-m9zpt" podUID="35e137ea-8f33-425b-a35d-8ab05164f05b" Oct 9 07:55:46.583936 kubelet[2734]: E1009 07:55:46.582396 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:47.031089 kubelet[2734]: E1009 07:55:47.031035 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:47.035459 containerd[1597]: time="2024-10-09T07:55:47.035227981Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Oct 9 07:55:47.047972 containerd[1597]: time="2024-10-09T07:55:47.047759818Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e12e5b177f7046f7a76fa0fa93732128102ff7cf0d22d741326bd42ca2a83873\"" Oct 9 07:55:47.050361 containerd[1597]: time="2024-10-09T07:55:47.049262073Z" level=info msg="StartContainer for \"e12e5b177f7046f7a76fa0fa93732128102ff7cf0d22d741326bd42ca2a83873\"" Oct 9 07:55:47.102208 systemd[1]: run-containerd-runc-k8s.io-e12e5b177f7046f7a76fa0fa93732128102ff7cf0d22d741326bd42ca2a83873-runc.B6PcGf.mount: Deactivated successfully. Oct 9 07:55:47.156105 containerd[1597]: time="2024-10-09T07:55:47.154793626Z" level=info msg="StartContainer for \"e12e5b177f7046f7a76fa0fa93732128102ff7cf0d22d741326bd42ca2a83873\" returns successfully" Oct 9 07:55:47.187702 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e12e5b177f7046f7a76fa0fa93732128102ff7cf0d22d741326bd42ca2a83873-rootfs.mount: Deactivated successfully. Oct 9 07:55:47.189883 containerd[1597]: time="2024-10-09T07:55:47.188692382Z" level=info msg="shim disconnected" id=e12e5b177f7046f7a76fa0fa93732128102ff7cf0d22d741326bd42ca2a83873 namespace=k8s.io Oct 9 07:55:47.189883 containerd[1597]: time="2024-10-09T07:55:47.189016497Z" level=warning msg="cleaning up after shim disconnected" id=e12e5b177f7046f7a76fa0fa93732128102ff7cf0d22d741326bd42ca2a83873 namespace=k8s.io Oct 9 07:55:47.189883 containerd[1597]: time="2024-10-09T07:55:47.189036173Z" level=info msg="cleaning up dead shim" namespace=k8s.io Oct 9 07:55:47.209236 containerd[1597]: time="2024-10-09T07:55:47.208949245Z" level=warning msg="cleanup warnings time=\"2024-10-09T07:55:47Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Oct 9 07:55:48.039767 kubelet[2734]: E1009 07:55:48.038625 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:48.055757 containerd[1597]: time="2024-10-09T07:55:48.050644230Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Oct 9 07:55:48.087290 containerd[1597]: time="2024-10-09T07:55:48.086932503Z" level=info msg="CreateContainer within sandbox \"d4d5dff1d6b5f104b0f74d56de61e85111240b9b0484286ccdf6f95673493aa6\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"21110c6e98c5fc069cf9539ee691c1b9092357e98c363ae7a83853d9bcfa9da3\"" Oct 9 07:55:48.090504 containerd[1597]: time="2024-10-09T07:55:48.089123507Z" level=info msg="StartContainer for \"21110c6e98c5fc069cf9539ee691c1b9092357e98c363ae7a83853d9bcfa9da3\"" Oct 9 07:55:48.254300 containerd[1597]: time="2024-10-09T07:55:48.254243041Z" level=info msg="StartContainer for \"21110c6e98c5fc069cf9539ee691c1b9092357e98c363ae7a83853d9bcfa9da3\" returns successfully" Oct 9 07:55:48.580272 kubelet[2734]: E1009 07:55:48.580203 2734 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-76f75df574-m9zpt" podUID="35e137ea-8f33-425b-a35d-8ab05164f05b" Oct 9 07:55:48.833085 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Oct 9 07:55:49.054700 kubelet[2734]: E1009 07:55:49.052266 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:49.091949 kubelet[2734]: I1009 07:55:49.090284 2734 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-vphmh" podStartSLOduration=6.089970099 podStartE2EDuration="6.089970099s" podCreationTimestamp="2024-10-09 07:55:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-10-09 07:55:49.089166771 +0000 UTC m=+104.709701195" watchObservedRunningTime="2024-10-09 07:55:49.089970099 +0000 UTC m=+104.710504543" Oct 9 07:55:50.446561 kubelet[2734]: E1009 07:55:50.446415 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:50.587544 kubelet[2734]: E1009 07:55:50.585069 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:52.579594 systemd-networkd[1226]: lxc_health: Link UP Oct 9 07:55:52.586060 systemd-networkd[1226]: lxc_health: Gained carrier Oct 9 07:55:53.133356 systemd[1]: run-containerd-runc-k8s.io-21110c6e98c5fc069cf9539ee691c1b9092357e98c363ae7a83853d9bcfa9da3-runc.rfyDUA.mount: Deactivated successfully. Oct 9 07:55:53.701049 systemd-networkd[1226]: lxc_health: Gained IPv6LL Oct 9 07:55:54.455746 kubelet[2734]: E1009 07:55:54.454856 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:55.074294 kubelet[2734]: E1009 07:55:55.074244 2734 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Oct 9 07:55:55.428671 kubelet[2734]: E1009 07:55:55.428118 2734 upgradeaware.go:425] Error proxying data from client to backend: readfrom tcp 127.0.0.1:39314->127.0.0.1:33445: write tcp 127.0.0.1:39314->127.0.0.1:33445: write: broken pipe Oct 9 07:55:57.531453 systemd[1]: run-containerd-runc-k8s.io-21110c6e98c5fc069cf9539ee691c1b9092357e98c363ae7a83853d9bcfa9da3-runc.Q7UTU9.mount: Deactivated successfully. Oct 9 07:55:57.608340 sshd[4552]: pam_unix(sshd:session): session closed for user core Oct 9 07:55:57.612549 systemd-logind[1572]: Session 28 logged out. Waiting for processes to exit. Oct 9 07:55:57.616089 systemd[1]: sshd@27-143.198.138.163:22-139.178.89.65:42188.service: Deactivated successfully. Oct 9 07:55:57.619600 systemd[1]: session-28.scope: Deactivated successfully. Oct 9 07:55:57.621393 systemd-logind[1572]: Removed session 28.