Feb 13 20:12:19.117899 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p1) 13.3.1 20240614, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Feb 13 17:44:05 -00 2025 Feb 13 20:12:19.117937 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 20:12:19.117951 kernel: BIOS-provided physical RAM map: Feb 13 20:12:19.117962 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Feb 13 20:12:19.117972 kernel: BIOS-e820: [mem 0x00000000000c0000-0x00000000000fffff] reserved Feb 13 20:12:19.117983 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003ff40fff] usable Feb 13 20:12:19.117996 kernel: BIOS-e820: [mem 0x000000003ff41000-0x000000003ffc8fff] reserved Feb 13 20:12:19.118010 kernel: BIOS-e820: [mem 0x000000003ffc9000-0x000000003fffafff] ACPI data Feb 13 20:12:19.118021 kernel: BIOS-e820: [mem 0x000000003fffb000-0x000000003fffefff] ACPI NVS Feb 13 20:12:19.118032 kernel: BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] usable Feb 13 20:12:19.118044 kernel: BIOS-e820: [mem 0x0000000100000000-0x00000002bfffffff] usable Feb 13 20:12:19.118055 kernel: printk: bootconsole [earlyser0] enabled Feb 13 20:12:19.118066 kernel: NX (Execute Disable) protection: active Feb 13 20:12:19.118078 kernel: APIC: Static calls initialized Feb 13 20:12:19.118095 kernel: efi: EFI v2.7 by Microsoft Feb 13 20:12:19.118108 kernel: efi: ACPI=0x3fffa000 ACPI 2.0=0x3fffa014 SMBIOS=0x3ff85000 SMBIOS 3.0=0x3ff83000 MEMATTR=0x3ee83a98 RNG=0x3ffd1018 Feb 13 20:12:19.118121 kernel: random: crng init done Feb 13 20:12:19.118133 kernel: secureboot: Secure boot disabled Feb 13 20:12:19.118145 kernel: SMBIOS 3.1.0 present. Feb 13 20:12:19.118158 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 03/08/2024 Feb 13 20:12:19.118171 kernel: Hypervisor detected: Microsoft Hyper-V Feb 13 20:12:19.118183 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0x64e24, misc 0xbed7b2 Feb 13 20:12:19.118196 kernel: Hyper-V: Host Build 10.0.20348.1799-1-0 Feb 13 20:12:19.118208 kernel: Hyper-V: Nested features: 0x1e0101 Feb 13 20:12:19.118223 kernel: Hyper-V: LAPIC Timer Frequency: 0x30d40 Feb 13 20:12:19.118235 kernel: Hyper-V: Using hypercall for remote TLB flush Feb 13 20:12:19.118248 kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Feb 13 20:12:19.118260 kernel: clocksource: hyperv_clocksource_msr: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Feb 13 20:12:19.118273 kernel: tsc: Marking TSC unstable due to running on Hyper-V Feb 13 20:12:19.118286 kernel: tsc: Detected 2593.907 MHz processor Feb 13 20:12:19.118299 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 13 20:12:19.118312 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 13 20:12:19.118325 kernel: last_pfn = 0x2c0000 max_arch_pfn = 0x400000000 Feb 13 20:12:19.118340 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Feb 13 20:12:19.118353 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 13 20:12:19.118366 kernel: e820: update [mem 0x40000000-0xffffffff] usable ==> reserved Feb 13 20:12:19.118378 kernel: last_pfn = 0x40000 max_arch_pfn = 0x400000000 Feb 13 20:12:19.118391 kernel: Using GB pages for direct mapping Feb 13 20:12:19.118403 kernel: ACPI: Early table checksum verification disabled Feb 13 20:12:19.118416 kernel: ACPI: RSDP 0x000000003FFFA014 000024 (v02 VRTUAL) Feb 13 20:12:19.118434 kernel: ACPI: XSDT 0x000000003FFF90E8 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118451 kernel: ACPI: FACP 0x000000003FFF8000 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118464 kernel: ACPI: DSDT 0x000000003FFD6000 01E184 (v02 MSFTVM DSDT01 00000001 MSFT 05000000) Feb 13 20:12:19.118477 kernel: ACPI: FACS 0x000000003FFFE000 000040 Feb 13 20:12:19.118491 kernel: ACPI: OEM0 0x000000003FFF7000 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118505 kernel: ACPI: SPCR 0x000000003FFF6000 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118518 kernel: ACPI: WAET 0x000000003FFF5000 000028 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118543 kernel: ACPI: APIC 0x000000003FFD5000 000058 (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118557 kernel: ACPI: SRAT 0x000000003FFD4000 0002D0 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118570 kernel: ACPI: BGRT 0x000000003FFD3000 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118584 kernel: ACPI: FPDT 0x000000003FFD2000 000034 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118598 kernel: ACPI: Reserving FACP table memory at [mem 0x3fff8000-0x3fff8113] Feb 13 20:12:19.118611 kernel: ACPI: Reserving DSDT table memory at [mem 0x3ffd6000-0x3fff4183] Feb 13 20:12:19.118625 kernel: ACPI: Reserving FACS table memory at [mem 0x3fffe000-0x3fffe03f] Feb 13 20:12:19.118639 kernel: ACPI: Reserving OEM0 table memory at [mem 0x3fff7000-0x3fff7063] Feb 13 20:12:19.118652 kernel: ACPI: Reserving SPCR table memory at [mem 0x3fff6000-0x3fff604f] Feb 13 20:12:19.118669 kernel: ACPI: Reserving WAET table memory at [mem 0x3fff5000-0x3fff5027] Feb 13 20:12:19.118682 kernel: ACPI: Reserving APIC table memory at [mem 0x3ffd5000-0x3ffd5057] Feb 13 20:12:19.118696 kernel: ACPI: Reserving SRAT table memory at [mem 0x3ffd4000-0x3ffd42cf] Feb 13 20:12:19.118710 kernel: ACPI: Reserving BGRT table memory at [mem 0x3ffd3000-0x3ffd3037] Feb 13 20:12:19.118723 kernel: ACPI: Reserving FPDT table memory at [mem 0x3ffd2000-0x3ffd2033] Feb 13 20:12:19.118737 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Feb 13 20:12:19.118750 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Feb 13 20:12:19.118764 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Feb 13 20:12:19.118780 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x2bfffffff] hotplug Feb 13 20:12:19.118794 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2c0000000-0xfdfffffff] hotplug Feb 13 20:12:19.118807 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Feb 13 20:12:19.118821 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Feb 13 20:12:19.118834 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Feb 13 20:12:19.118848 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Feb 13 20:12:19.118862 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Feb 13 20:12:19.118875 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Feb 13 20:12:19.118889 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Feb 13 20:12:19.118905 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Feb 13 20:12:19.118919 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Feb 13 20:12:19.118932 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000000-0x1ffffffffffff] hotplug Feb 13 20:12:19.118946 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2000000000000-0x3ffffffffffff] hotplug Feb 13 20:12:19.118959 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x4000000000000-0x7ffffffffffff] hotplug Feb 13 20:12:19.118973 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x8000000000000-0xfffffffffffff] hotplug Feb 13 20:12:19.118986 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x2bfffffff] -> [mem 0x00000000-0x2bfffffff] Feb 13 20:12:19.119000 kernel: NODE_DATA(0) allocated [mem 0x2bfffa000-0x2bfffffff] Feb 13 20:12:19.119013 kernel: Zone ranges: Feb 13 20:12:19.119029 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 13 20:12:19.119043 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Feb 13 20:12:19.119056 kernel: Normal [mem 0x0000000100000000-0x00000002bfffffff] Feb 13 20:12:19.119070 kernel: Movable zone start for each node Feb 13 20:12:19.119083 kernel: Early memory node ranges Feb 13 20:12:19.119096 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Feb 13 20:12:19.119110 kernel: node 0: [mem 0x0000000000100000-0x000000003ff40fff] Feb 13 20:12:19.119123 kernel: node 0: [mem 0x000000003ffff000-0x000000003fffffff] Feb 13 20:12:19.119137 kernel: node 0: [mem 0x0000000100000000-0x00000002bfffffff] Feb 13 20:12:19.119153 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x00000002bfffffff] Feb 13 20:12:19.119167 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 20:12:19.119180 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Feb 13 20:12:19.119193 kernel: On node 0, zone DMA32: 190 pages in unavailable ranges Feb 13 20:12:19.119207 kernel: ACPI: PM-Timer IO Port: 0x408 Feb 13 20:12:19.119220 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] dfl dfl lint[0x1]) Feb 13 20:12:19.119234 kernel: IOAPIC[0]: apic_id 2, version 17, address 0xfec00000, GSI 0-23 Feb 13 20:12:19.119248 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 13 20:12:19.119261 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 13 20:12:19.119277 kernel: ACPI: SPCR: console: uart,io,0x3f8,115200 Feb 13 20:12:19.119291 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Feb 13 20:12:19.119304 kernel: [mem 0x40000000-0xffffffff] available for PCI devices Feb 13 20:12:19.119318 kernel: Booting paravirtualized kernel on Hyper-V Feb 13 20:12:19.119331 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 13 20:12:19.119345 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Feb 13 20:12:19.119356 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Feb 13 20:12:19.119368 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Feb 13 20:12:19.119379 kernel: pcpu-alloc: [0] 0 1 Feb 13 20:12:19.119393 kernel: Hyper-V: PV spinlocks enabled Feb 13 20:12:19.119405 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Feb 13 20:12:19.119418 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 20:12:19.119426 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 20:12:19.119433 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Feb 13 20:12:19.119440 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 20:12:19.119448 kernel: Fallback order for Node 0: 0 Feb 13 20:12:19.119455 kernel: Built 1 zonelists, mobility grouping on. Total pages: 2062618 Feb 13 20:12:19.119466 kernel: Policy zone: Normal Feb 13 20:12:19.119481 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 20:12:19.119489 kernel: software IO TLB: area num 2. Feb 13 20:12:19.119500 kernel: Memory: 8077024K/8387460K available (12288K kernel code, 2301K rwdata, 22736K rodata, 42976K init, 2216K bss, 310180K reserved, 0K cma-reserved) Feb 13 20:12:19.119510 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 13 20:12:19.119522 kernel: ftrace: allocating 37923 entries in 149 pages Feb 13 20:12:19.119550 kernel: ftrace: allocated 149 pages with 4 groups Feb 13 20:12:19.119562 kernel: Dynamic Preempt: voluntary Feb 13 20:12:19.119575 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 20:12:19.119589 kernel: rcu: RCU event tracing is enabled. Feb 13 20:12:19.119601 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 13 20:12:19.119618 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 20:12:19.119630 kernel: Rude variant of Tasks RCU enabled. Feb 13 20:12:19.119643 kernel: Tracing variant of Tasks RCU enabled. Feb 13 20:12:19.119657 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 20:12:19.119671 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 13 20:12:19.119702 kernel: Using NULL legacy PIC Feb 13 20:12:19.119732 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 0 Feb 13 20:12:19.119748 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 20:12:19.119760 kernel: Console: colour dummy device 80x25 Feb 13 20:12:19.119774 kernel: printk: console [tty1] enabled Feb 13 20:12:19.119789 kernel: printk: console [ttyS0] enabled Feb 13 20:12:19.119804 kernel: printk: bootconsole [earlyser0] disabled Feb 13 20:12:19.119817 kernel: ACPI: Core revision 20230628 Feb 13 20:12:19.119829 kernel: Failed to register legacy timer interrupt Feb 13 20:12:19.119842 kernel: APIC: Switch to symmetric I/O mode setup Feb 13 20:12:19.119858 kernel: Hyper-V: enabling crash_kexec_post_notifiers Feb 13 20:12:19.119873 kernel: Hyper-V: Using IPI hypercalls Feb 13 20:12:19.119886 kernel: APIC: send_IPI() replaced with hv_send_ipi() Feb 13 20:12:19.119898 kernel: APIC: send_IPI_mask() replaced with hv_send_ipi_mask() Feb 13 20:12:19.119910 kernel: APIC: send_IPI_mask_allbutself() replaced with hv_send_ipi_mask_allbutself() Feb 13 20:12:19.119923 kernel: APIC: send_IPI_allbutself() replaced with hv_send_ipi_allbutself() Feb 13 20:12:19.119936 kernel: APIC: send_IPI_all() replaced with hv_send_ipi_all() Feb 13 20:12:19.119948 kernel: APIC: send_IPI_self() replaced with hv_send_ipi_self() Feb 13 20:12:19.119964 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 5187.81 BogoMIPS (lpj=2593907) Feb 13 20:12:19.119988 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Feb 13 20:12:19.119999 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Feb 13 20:12:19.120010 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 13 20:12:19.120021 kernel: Spectre V2 : Mitigation: Retpolines Feb 13 20:12:19.120033 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 13 20:12:19.120046 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 13 20:12:19.120060 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Feb 13 20:12:19.120071 kernel: RETBleed: Vulnerable Feb 13 20:12:19.120083 kernel: Speculative Store Bypass: Vulnerable Feb 13 20:12:19.120095 kernel: TAA: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 20:12:19.120112 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 20:12:19.120126 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 13 20:12:19.120140 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 13 20:12:19.120152 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 13 20:12:19.120164 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Feb 13 20:12:19.120179 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Feb 13 20:12:19.120194 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Feb 13 20:12:19.120209 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 13 20:12:19.120223 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Feb 13 20:12:19.120238 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Feb 13 20:12:19.120252 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Feb 13 20:12:19.120270 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Feb 13 20:12:19.120285 kernel: Freeing SMP alternatives memory: 32K Feb 13 20:12:19.120299 kernel: pid_max: default: 32768 minimum: 301 Feb 13 20:12:19.120314 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 20:12:19.120329 kernel: landlock: Up and running. Feb 13 20:12:19.120343 kernel: SELinux: Initializing. Feb 13 20:12:19.120358 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.120372 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.120388 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8272CL CPU @ 2.60GHz (family: 0x6, model: 0x55, stepping: 0x7) Feb 13 20:12:19.120403 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:12:19.120418 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:12:19.120437 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:12:19.120452 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Feb 13 20:12:19.120467 kernel: signal: max sigframe size: 3632 Feb 13 20:12:19.120483 kernel: rcu: Hierarchical SRCU implementation. Feb 13 20:12:19.120498 kernel: rcu: Max phase no-delay instances is 400. Feb 13 20:12:19.120513 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Feb 13 20:12:19.120526 kernel: smp: Bringing up secondary CPUs ... Feb 13 20:12:19.120552 kernel: smpboot: x86: Booting SMP configuration: Feb 13 20:12:19.120566 kernel: .... node #0, CPUs: #1 Feb 13 20:12:19.120584 kernel: TAA CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/tsx_async_abort.html for more details. Feb 13 20:12:19.120600 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Feb 13 20:12:19.120614 kernel: smp: Brought up 1 node, 2 CPUs Feb 13 20:12:19.120628 kernel: smpboot: Max logical packages: 1 Feb 13 20:12:19.120642 kernel: smpboot: Total of 2 processors activated (10375.62 BogoMIPS) Feb 13 20:12:19.120656 kernel: devtmpfs: initialized Feb 13 20:12:19.120670 kernel: x86/mm: Memory block size: 128MB Feb 13 20:12:19.120684 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x3fffb000-0x3fffefff] (16384 bytes) Feb 13 20:12:19.120701 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 20:12:19.120716 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 13 20:12:19.120728 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 20:12:19.120740 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 20:12:19.120754 kernel: audit: initializing netlink subsys (disabled) Feb 13 20:12:19.120768 kernel: audit: type=2000 audit(1739477537.028:1): state=initialized audit_enabled=0 res=1 Feb 13 20:12:19.120782 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 20:12:19.120796 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 13 20:12:19.120810 kernel: cpuidle: using governor menu Feb 13 20:12:19.120826 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 20:12:19.120840 kernel: dca service started, version 1.12.1 Feb 13 20:12:19.120854 kernel: e820: reserve RAM buffer [mem 0x3ff41000-0x3fffffff] Feb 13 20:12:19.120869 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 13 20:12:19.120882 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 20:12:19.120897 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 20:12:19.120911 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 20:12:19.120928 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 20:12:19.120942 kernel: ACPI: Added _OSI(Module Device) Feb 13 20:12:19.120958 kernel: ACPI: Added _OSI(Processor Device) Feb 13 20:12:19.120973 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 20:12:19.120987 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 20:12:19.121001 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 20:12:19.121015 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Feb 13 20:12:19.121028 kernel: ACPI: Interpreter enabled Feb 13 20:12:19.121042 kernel: ACPI: PM: (supports S0 S5) Feb 13 20:12:19.121056 kernel: ACPI: Using IOAPIC for interrupt routing Feb 13 20:12:19.121070 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 13 20:12:19.121087 kernel: PCI: Ignoring E820 reservations for host bridge windows Feb 13 20:12:19.121101 kernel: ACPI: Enabled 1 GPEs in block 00 to 0F Feb 13 20:12:19.121115 kernel: iommu: Default domain type: Translated Feb 13 20:12:19.121129 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 13 20:12:19.121143 kernel: efivars: Registered efivars operations Feb 13 20:12:19.121157 kernel: PCI: Using ACPI for IRQ routing Feb 13 20:12:19.121171 kernel: PCI: System does not support PCI Feb 13 20:12:19.121184 kernel: vgaarb: loaded Feb 13 20:12:19.121198 kernel: clocksource: Switched to clocksource hyperv_clocksource_tsc_page Feb 13 20:12:19.121215 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 20:12:19.121229 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 20:12:19.121243 kernel: pnp: PnP ACPI init Feb 13 20:12:19.121257 kernel: pnp: PnP ACPI: found 3 devices Feb 13 20:12:19.121271 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 13 20:12:19.121285 kernel: NET: Registered PF_INET protocol family Feb 13 20:12:19.121299 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Feb 13 20:12:19.121313 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Feb 13 20:12:19.121327 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 20:12:19.121347 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 20:12:19.121361 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Feb 13 20:12:19.121375 kernel: TCP: Hash tables configured (established 65536 bind 65536) Feb 13 20:12:19.121388 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.121402 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.121417 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 20:12:19.121429 kernel: NET: Registered PF_XDP protocol family Feb 13 20:12:19.121441 kernel: PCI: CLS 0 bytes, default 64 Feb 13 20:12:19.121454 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Feb 13 20:12:19.121471 kernel: software IO TLB: mapped [mem 0x000000003ad8c000-0x000000003ed8c000] (64MB) Feb 13 20:12:19.121485 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Feb 13 20:12:19.121498 kernel: Initialise system trusted keyrings Feb 13 20:12:19.121525 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Feb 13 20:12:19.121549 kernel: Key type asymmetric registered Feb 13 20:12:19.121560 kernel: Asymmetric key parser 'x509' registered Feb 13 20:12:19.121572 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Feb 13 20:12:19.121586 kernel: io scheduler mq-deadline registered Feb 13 20:12:19.121601 kernel: io scheduler kyber registered Feb 13 20:12:19.121621 kernel: io scheduler bfq registered Feb 13 20:12:19.121635 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 13 20:12:19.121649 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 20:12:19.121663 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 13 20:12:19.121678 kernel: 00:01: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Feb 13 20:12:19.121693 kernel: i8042: PNP: No PS/2 controller found. Feb 13 20:12:19.121883 kernel: rtc_cmos 00:02: registered as rtc0 Feb 13 20:12:19.122007 kernel: rtc_cmos 00:02: setting system clock to 2025-02-13T20:12:18 UTC (1739477538) Feb 13 20:12:19.122118 kernel: rtc_cmos 00:02: alarms up to one month, 114 bytes nvram Feb 13 20:12:19.122135 kernel: intel_pstate: CPU model not supported Feb 13 20:12:19.122149 kernel: efifb: probing for efifb Feb 13 20:12:19.122164 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Feb 13 20:12:19.122177 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Feb 13 20:12:19.122191 kernel: efifb: scrolling: redraw Feb 13 20:12:19.122205 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Feb 13 20:12:19.122219 kernel: Console: switching to colour frame buffer device 128x48 Feb 13 20:12:19.122236 kernel: fb0: EFI VGA frame buffer device Feb 13 20:12:19.122250 kernel: pstore: Using crash dump compression: deflate Feb 13 20:12:19.122264 kernel: pstore: Registered efi_pstore as persistent store backend Feb 13 20:12:19.122278 kernel: NET: Registered PF_INET6 protocol family Feb 13 20:12:19.122291 kernel: Segment Routing with IPv6 Feb 13 20:12:19.122305 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 20:12:19.122319 kernel: NET: Registered PF_PACKET protocol family Feb 13 20:12:19.122333 kernel: Key type dns_resolver registered Feb 13 20:12:19.122347 kernel: IPI shorthand broadcast: enabled Feb 13 20:12:19.122361 kernel: sched_clock: Marking stable (902002900, 54173500)->(1197844200, -241667800) Feb 13 20:12:19.122378 kernel: registered taskstats version 1 Feb 13 20:12:19.122392 kernel: Loading compiled-in X.509 certificates Feb 13 20:12:19.122405 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 0cc219a306b9e46e583adebba1820decbdc4307b' Feb 13 20:12:19.122419 kernel: Key type .fscrypt registered Feb 13 20:12:19.122432 kernel: Key type fscrypt-provisioning registered Feb 13 20:12:19.122446 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 20:12:19.122460 kernel: ima: Allocated hash algorithm: sha1 Feb 13 20:12:19.122474 kernel: ima: No architecture policies found Feb 13 20:12:19.122490 kernel: clk: Disabling unused clocks Feb 13 20:12:19.122505 kernel: Freeing unused kernel image (initmem) memory: 42976K Feb 13 20:12:19.122518 kernel: Write protecting the kernel read-only data: 36864k Feb 13 20:12:19.122544 kernel: Freeing unused kernel image (rodata/data gap) memory: 1840K Feb 13 20:12:19.122567 kernel: Run /init as init process Feb 13 20:12:19.122581 kernel: with arguments: Feb 13 20:12:19.122594 kernel: /init Feb 13 20:12:19.122608 kernel: with environment: Feb 13 20:12:19.122621 kernel: HOME=/ Feb 13 20:12:19.122634 kernel: TERM=linux Feb 13 20:12:19.122651 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 20:12:19.122668 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 20:12:19.122684 systemd[1]: Detected virtualization microsoft. Feb 13 20:12:19.122709 systemd[1]: Detected architecture x86-64. Feb 13 20:12:19.122722 systemd[1]: Running in initrd. Feb 13 20:12:19.122737 systemd[1]: No hostname configured, using default hostname. Feb 13 20:12:19.122750 systemd[1]: Hostname set to . Feb 13 20:12:19.122774 systemd[1]: Initializing machine ID from random generator. Feb 13 20:12:19.122789 systemd[1]: Queued start job for default target initrd.target. Feb 13 20:12:19.122805 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:12:19.122821 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:12:19.122838 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 20:12:19.122855 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 20:12:19.122870 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 20:12:19.122886 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 20:12:19.122908 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 20:12:19.122923 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 20:12:19.122940 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:12:19.122956 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:12:19.122971 systemd[1]: Reached target paths.target - Path Units. Feb 13 20:12:19.122987 systemd[1]: Reached target slices.target - Slice Units. Feb 13 20:12:19.123003 systemd[1]: Reached target swap.target - Swaps. Feb 13 20:12:19.123022 systemd[1]: Reached target timers.target - Timer Units. Feb 13 20:12:19.123037 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 20:12:19.123053 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 20:12:19.123070 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 20:12:19.123086 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Feb 13 20:12:19.123102 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:12:19.123118 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 20:12:19.123133 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:12:19.123152 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 20:12:19.123168 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 20:12:19.123184 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 20:12:19.123200 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 20:12:19.123216 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 20:12:19.123232 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 20:12:19.123248 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 20:12:19.123264 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:19.123303 systemd-journald[177]: Collecting audit messages is disabled. Feb 13 20:12:19.123342 systemd-journald[177]: Journal started Feb 13 20:12:19.123375 systemd-journald[177]: Runtime Journal (/run/log/journal/0238720c1c9a40a98bc505e0f3e0b423) is 8.0M, max 158.8M, 150.8M free. Feb 13 20:12:19.124346 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 20:12:19.120010 systemd-modules-load[178]: Inserted module 'overlay' Feb 13 20:12:19.137623 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 20:12:19.141601 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:12:19.149975 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 20:12:19.166828 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 20:12:19.178341 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 20:12:19.178378 kernel: Bridge firewalling registered Feb 13 20:12:19.180623 systemd-modules-load[178]: Inserted module 'br_netfilter' Feb 13 20:12:19.183117 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 20:12:19.189366 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 20:12:19.195066 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:19.201198 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 20:12:19.206763 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:12:19.218698 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:12:19.225833 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 20:12:19.233980 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 20:12:19.250465 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:12:19.257347 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:12:19.266691 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 20:12:19.272079 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:19.279727 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 20:12:19.302214 dracut-cmdline[213]: dracut-dracut-053 Feb 13 20:12:19.306603 dracut-cmdline[213]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 20:12:19.314471 systemd-resolved[208]: Positive Trust Anchors: Feb 13 20:12:19.314487 systemd-resolved[208]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 20:12:19.314525 systemd-resolved[208]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 20:12:19.317164 systemd-resolved[208]: Defaulting to hostname 'linux'. Feb 13 20:12:19.318152 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 20:12:19.326765 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:12:19.415558 kernel: SCSI subsystem initialized Feb 13 20:12:19.425551 kernel: Loading iSCSI transport class v2.0-870. Feb 13 20:12:19.436556 kernel: iscsi: registered transport (tcp) Feb 13 20:12:19.457521 kernel: iscsi: registered transport (qla4xxx) Feb 13 20:12:19.457605 kernel: QLogic iSCSI HBA Driver Feb 13 20:12:19.492273 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 20:12:19.505798 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 20:12:19.535774 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 20:12:19.535884 kernel: device-mapper: uevent: version 1.0.3 Feb 13 20:12:19.539916 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 20:12:19.581562 kernel: raid6: avx512x4 gen() 18512 MB/s Feb 13 20:12:19.600545 kernel: raid6: avx512x2 gen() 18467 MB/s Feb 13 20:12:19.619540 kernel: raid6: avx512x1 gen() 18291 MB/s Feb 13 20:12:19.638552 kernel: raid6: avx2x4 gen() 18336 MB/s Feb 13 20:12:19.657545 kernel: raid6: avx2x2 gen() 18313 MB/s Feb 13 20:12:19.677722 kernel: raid6: avx2x1 gen() 13940 MB/s Feb 13 20:12:19.677769 kernel: raid6: using algorithm avx512x4 gen() 18512 MB/s Feb 13 20:12:19.699591 kernel: raid6: .... xor() 6457 MB/s, rmw enabled Feb 13 20:12:19.699629 kernel: raid6: using avx512x2 recovery algorithm Feb 13 20:12:19.721560 kernel: xor: automatically using best checksumming function avx Feb 13 20:12:19.872560 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 20:12:19.881760 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 20:12:19.892696 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:12:19.907573 systemd-udevd[395]: Using default interface naming scheme 'v255'. Feb 13 20:12:19.912121 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:12:19.930730 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 20:12:19.944892 dracut-pre-trigger[405]: rd.md=0: removing MD RAID activation Feb 13 20:12:19.972261 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 20:12:19.997707 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 20:12:20.040444 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:12:20.051698 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 20:12:20.069927 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 20:12:20.077999 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 20:12:20.084800 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:12:20.087934 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 20:12:20.102818 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 20:12:20.124996 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 20:12:20.137366 kernel: cryptd: max_cpu_qlen set to 1000 Feb 13 20:12:20.161192 kernel: AVX2 version of gcm_enc/dec engaged. Feb 13 20:12:20.161270 kernel: AES CTR mode by8 optimization enabled Feb 13 20:12:20.163307 kernel: hv_vmbus: Vmbus version:5.2 Feb 13 20:12:20.172294 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 20:12:20.177786 kernel: hv_vmbus: registering driver hyperv_keyboard Feb 13 20:12:20.173319 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:20.207067 kernel: pps_core: LinuxPPS API ver. 1 registered Feb 13 20:12:20.207162 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/VMBUS:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Feb 13 20:12:20.207189 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Feb 13 20:12:20.193101 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:12:20.212160 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:12:20.212469 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:20.224440 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:20.237012 kernel: PTP clock support registered Feb 13 20:12:20.235990 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:20.248841 kernel: hv_utils: Registering HyperV Utility Driver Feb 13 20:12:20.248892 kernel: hv_vmbus: registering driver hv_utils Feb 13 20:12:20.248906 kernel: hv_utils: Heartbeat IC version 3.0 Feb 13 20:12:20.251552 kernel: hv_utils: Shutdown IC version 3.2 Feb 13 20:12:20.801236 kernel: hv_utils: TimeSync IC version 4.0 Feb 13 20:12:20.800279 systemd-resolved[208]: Clock change detected. Flushing caches. Feb 13 20:12:20.813641 kernel: hid: raw HID events driver (C) Jiri Kosina Feb 13 20:12:20.818401 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:12:20.848955 kernel: hv_vmbus: registering driver hv_storvsc Feb 13 20:12:20.848987 kernel: hv_vmbus: registering driver hv_netvsc Feb 13 20:12:20.849007 kernel: scsi host1: storvsc_host_t Feb 13 20:12:20.849214 kernel: hv_vmbus: registering driver hid_hyperv Feb 13 20:12:20.849233 kernel: scsi host0: storvsc_host_t Feb 13 20:12:20.849524 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Feb 13 20:12:20.849545 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Feb 13 20:12:20.855732 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Feb 13 20:12:20.855783 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Feb 13 20:12:20.818523 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:20.850364 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:20.887875 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Feb 13 20:12:20.890119 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Feb 13 20:12:20.890145 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Feb 13 20:12:20.888376 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:20.905205 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Feb 13 20:12:20.923400 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Feb 13 20:12:20.923588 kernel: sd 0:0:0:0: [sda] Write Protect is off Feb 13 20:12:20.923763 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Feb 13 20:12:20.923930 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Feb 13 20:12:20.924107 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:20.924128 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Feb 13 20:12:20.905971 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:12:20.935324 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:20.972186 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: VF slot 1 added Feb 13 20:12:20.980290 kernel: hv_vmbus: registering driver hv_pci Feb 13 20:12:20.986033 kernel: hv_pci a9d58489-ac61-4534-a070-abd92e6109ea: PCI VMBus probing: Using version 0x10004 Feb 13 20:12:21.031583 kernel: hv_pci a9d58489-ac61-4534-a070-abd92e6109ea: PCI host bridge to bus ac61:00 Feb 13 20:12:21.031879 kernel: pci_bus ac61:00: root bus resource [mem 0xfe0000000-0xfe00fffff window] Feb 13 20:12:21.032131 kernel: pci_bus ac61:00: No busn resource found for root bus, will use [bus 00-ff] Feb 13 20:12:21.032370 kernel: pci ac61:00:02.0: [15b3:1016] type 00 class 0x020000 Feb 13 20:12:21.032623 kernel: pci ac61:00:02.0: reg 0x10: [mem 0xfe0000000-0xfe00fffff 64bit pref] Feb 13 20:12:21.032838 kernel: pci ac61:00:02.0: enabling Extended Tags Feb 13 20:12:21.033066 kernel: pci ac61:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at ac61:00:02.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Feb 13 20:12:21.033321 kernel: pci_bus ac61:00: busn_res: [bus 00-ff] end is updated to 00 Feb 13 20:12:21.033562 kernel: pci ac61:00:02.0: BAR 0: assigned [mem 0xfe0000000-0xfe00fffff 64bit pref] Feb 13 20:12:21.206478 kernel: mlx5_core ac61:00:02.0: enabling device (0000 -> 0002) Feb 13 20:12:21.442179 kernel: mlx5_core ac61:00:02.0: firmware version: 14.30.5000 Feb 13 20:12:21.442437 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: VF registering: eth1 Feb 13 20:12:21.442909 kernel: mlx5_core ac61:00:02.0 eth1: joined to eth0 Feb 13 20:12:21.443660 kernel: mlx5_core ac61:00:02.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Feb 13 20:12:21.407916 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Feb 13 20:12:21.450273 kernel: mlx5_core ac61:00:02.0 enP44129s1: renamed from eth1 Feb 13 20:12:21.522311 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by (udev-worker) (447) Feb 13 20:12:21.538114 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Feb 13 20:12:21.555285 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Feb 13 20:12:21.570791 kernel: BTRFS: device fsid e9c87d9f-3864-4b45-9be4-80a5397f1fc6 devid 1 transid 38 /dev/sda3 scanned by (udev-worker) (441) Feb 13 20:12:21.589007 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Feb 13 20:12:21.592296 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Feb 13 20:12:21.608411 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 20:12:21.624283 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:21.630268 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:22.637670 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:22.637750 disk-uuid[601]: The operation has completed successfully. Feb 13 20:12:22.709138 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 20:12:22.709264 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 20:12:22.738607 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 20:12:22.747761 sh[687]: Success Feb 13 20:12:22.781571 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Feb 13 20:12:22.977608 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 20:12:22.992387 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 20:12:22.997387 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 20:12:23.012268 kernel: BTRFS info (device dm-0): first mount of filesystem e9c87d9f-3864-4b45-9be4-80a5397f1fc6 Feb 13 20:12:23.012330 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:23.017909 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 20:12:23.020573 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 20:12:23.023041 kernel: BTRFS info (device dm-0): using free space tree Feb 13 20:12:23.305474 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 20:12:23.310916 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 20:12:23.326441 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 20:12:23.334292 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 20:12:23.348399 kernel: BTRFS info (device sda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:23.348457 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:23.348474 kernel: BTRFS info (device sda6): using free space tree Feb 13 20:12:23.426727 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 20:12:23.438285 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 20:12:23.440731 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 20:12:23.452222 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 20:12:23.458439 kernel: BTRFS info (device sda6): last unmount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:23.465047 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 20:12:23.473969 systemd-networkd[861]: lo: Link UP Feb 13 20:12:23.474137 systemd-networkd[861]: lo: Gained carrier Feb 13 20:12:23.475457 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 20:12:23.479154 systemd-networkd[861]: Enumeration completed Feb 13 20:12:23.481829 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 20:12:23.482671 systemd-networkd[861]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:12:23.482676 systemd-networkd[861]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 20:12:23.490732 systemd[1]: Reached target network.target - Network. Feb 13 20:12:23.551280 kernel: mlx5_core ac61:00:02.0 enP44129s1: Link up Feb 13 20:12:23.591393 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: Data path switched to VF: enP44129s1 Feb 13 20:12:23.591551 systemd-networkd[861]: enP44129s1: Link UP Feb 13 20:12:23.591689 systemd-networkd[861]: eth0: Link UP Feb 13 20:12:23.596894 systemd-networkd[861]: eth0: Gained carrier Feb 13 20:12:23.596909 systemd-networkd[861]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:12:23.607514 systemd-networkd[861]: enP44129s1: Gained carrier Feb 13 20:12:23.635335 systemd-networkd[861]: eth0: DHCPv4 address 10.200.4.48/24, gateway 10.200.4.1 acquired from 168.63.129.16 Feb 13 20:12:24.445969 ignition[871]: Ignition 2.20.0 Feb 13 20:12:24.445981 ignition[871]: Stage: fetch-offline Feb 13 20:12:24.446032 ignition[871]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.446042 ignition[871]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.446170 ignition[871]: parsed url from cmdline: "" Feb 13 20:12:24.446175 ignition[871]: no config URL provided Feb 13 20:12:24.446186 ignition[871]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 20:12:24.446196 ignition[871]: no config at "/usr/lib/ignition/user.ign" Feb 13 20:12:24.446205 ignition[871]: failed to fetch config: resource requires networking Feb 13 20:12:24.462531 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 20:12:24.446535 ignition[871]: Ignition finished successfully Feb 13 20:12:24.484537 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Feb 13 20:12:24.502806 ignition[880]: Ignition 2.20.0 Feb 13 20:12:24.502818 ignition[880]: Stage: fetch Feb 13 20:12:24.503035 ignition[880]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.503048 ignition[880]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.503438 ignition[880]: parsed url from cmdline: "" Feb 13 20:12:24.503443 ignition[880]: no config URL provided Feb 13 20:12:24.503448 ignition[880]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 20:12:24.503457 ignition[880]: no config at "/usr/lib/ignition/user.ign" Feb 13 20:12:24.503482 ignition[880]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Feb 13 20:12:24.595897 ignition[880]: GET result: OK Feb 13 20:12:24.596031 ignition[880]: config has been read from IMDS userdata Feb 13 20:12:24.596072 ignition[880]: parsing config with SHA512: 34218719cc285cb68475ae095c7d73c3f3dd2930cb0ce913866ad9bbd923829ca17f50ff3b0bcaa227592c93966bca55f0102758c079459dd0db3b87cd78c264 Feb 13 20:12:24.605212 unknown[880]: fetched base config from "system" Feb 13 20:12:24.605551 ignition[880]: fetch: fetch complete Feb 13 20:12:24.605218 unknown[880]: fetched base config from "system" Feb 13 20:12:24.605556 ignition[880]: fetch: fetch passed Feb 13 20:12:24.605224 unknown[880]: fetched user config from "azure" Feb 13 20:12:24.605600 ignition[880]: Ignition finished successfully Feb 13 20:12:24.607319 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Feb 13 20:12:24.627460 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 20:12:24.646625 ignition[886]: Ignition 2.20.0 Feb 13 20:12:24.646637 ignition[886]: Stage: kargs Feb 13 20:12:24.646853 ignition[886]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.646866 ignition[886]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.647752 ignition[886]: kargs: kargs passed Feb 13 20:12:24.647805 ignition[886]: Ignition finished successfully Feb 13 20:12:24.658355 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 20:12:24.668563 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 20:12:24.681303 ignition[892]: Ignition 2.20.0 Feb 13 20:12:24.681314 ignition[892]: Stage: disks Feb 13 20:12:24.683434 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 20:12:24.681543 ignition[892]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.681557 ignition[892]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.682464 ignition[892]: disks: disks passed Feb 13 20:12:24.682512 ignition[892]: Ignition finished successfully Feb 13 20:12:24.698235 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 20:12:24.701105 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 20:12:24.704734 systemd-networkd[861]: eth0: Gained IPv6LL Feb 13 20:12:24.707593 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 20:12:24.707922 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 20:12:24.708324 systemd[1]: Reached target basic.target - Basic System. Feb 13 20:12:24.721529 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 20:12:24.785872 systemd-fsck[900]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Feb 13 20:12:24.792633 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 20:12:24.805564 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 20:12:24.903300 kernel: EXT4-fs (sda9): mounted filesystem c5993b0e-9201-4b44-aa01-79dc9d6c9fc9 r/w with ordered data mode. Quota mode: none. Feb 13 20:12:24.905182 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 20:12:24.912665 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 20:12:24.962406 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 20:12:24.971650 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 20:12:24.979954 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (911) Feb 13 20:12:24.985864 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Feb 13 20:12:25.003937 kernel: BTRFS info (device sda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:25.003976 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:25.003993 kernel: BTRFS info (device sda6): using free space tree Feb 13 20:12:25.004011 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 20:12:24.996159 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 20:12:24.996197 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 20:12:25.012567 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 20:12:25.015099 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 20:12:25.024440 systemd-networkd[861]: enP44129s1: Gained IPv6LL Feb 13 20:12:25.028412 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 20:12:25.572517 coreos-metadata[913]: Feb 13 20:12:25.572 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Feb 13 20:12:25.579576 coreos-metadata[913]: Feb 13 20:12:25.579 INFO Fetch successful Feb 13 20:12:25.579576 coreos-metadata[913]: Feb 13 20:12:25.579 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Feb 13 20:12:25.593007 coreos-metadata[913]: Feb 13 20:12:25.592 INFO Fetch successful Feb 13 20:12:25.607907 coreos-metadata[913]: Feb 13 20:12:25.607 INFO wrote hostname ci-4152.2.1-a-2754074fca to /sysroot/etc/hostname Feb 13 20:12:25.616139 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Feb 13 20:12:25.681859 initrd-setup-root[942]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 20:12:25.718110 initrd-setup-root[949]: cut: /sysroot/etc/group: No such file or directory Feb 13 20:12:25.744482 initrd-setup-root[956]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 20:12:25.750067 initrd-setup-root[963]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 20:12:26.698894 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 20:12:26.708573 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 20:12:26.715464 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 20:12:26.728211 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 20:12:26.734353 kernel: BTRFS info (device sda6): last unmount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:26.749655 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 20:12:26.761879 ignition[1031]: INFO : Ignition 2.20.0 Feb 13 20:12:26.761879 ignition[1031]: INFO : Stage: mount Feb 13 20:12:26.768533 ignition[1031]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:26.768533 ignition[1031]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:26.768533 ignition[1031]: INFO : mount: mount passed Feb 13 20:12:26.768533 ignition[1031]: INFO : Ignition finished successfully Feb 13 20:12:26.764067 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 20:12:26.778530 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 20:12:26.796508 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 20:12:26.809285 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/sda6 scanned by mount (1042) Feb 13 20:12:26.815906 kernel: BTRFS info (device sda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:26.816002 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:26.818418 kernel: BTRFS info (device sda6): using free space tree Feb 13 20:12:26.823276 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 20:12:26.825741 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 20:12:26.850885 ignition[1059]: INFO : Ignition 2.20.0 Feb 13 20:12:26.850885 ignition[1059]: INFO : Stage: files Feb 13 20:12:26.855206 ignition[1059]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:26.855206 ignition[1059]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:26.855206 ignition[1059]: DEBUG : files: compiled without relabeling support, skipping Feb 13 20:12:26.868896 ignition[1059]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 20:12:26.872735 ignition[1059]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 20:12:26.922582 ignition[1059]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 20:12:26.926815 ignition[1059]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 20:12:26.926815 ignition[1059]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 20:12:26.923163 unknown[1059]: wrote ssh authorized keys file for user: core Feb 13 20:12:26.941009 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 20:12:26.946383 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Feb 13 20:12:26.983834 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 20:12:27.112299 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 20:12:27.118676 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Feb 13 20:12:27.118676 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Feb 13 20:12:27.614810 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Feb 13 20:12:27.697614 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Feb 13 20:12:27.702334 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Feb 13 20:12:27.706600 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 20:12:27.706600 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 20:12:27.715238 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 20:12:27.719662 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 20:12:27.723923 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 20:12:27.728234 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 20:12:27.736650 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 20:12:27.741281 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 20:12:27.745728 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 20:12:27.750025 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:27.756355 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:27.762496 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:27.767818 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 Feb 13 20:12:28.242895 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Feb 13 20:12:28.557755 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:28.557755 ignition[1059]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Feb 13 20:12:28.591382 ignition[1059]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: files passed Feb 13 20:12:28.602094 ignition[1059]: INFO : Ignition finished successfully Feb 13 20:12:28.593287 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 20:12:28.610540 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 20:12:28.617813 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 20:12:28.664998 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 20:12:28.665119 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 20:12:28.675955 initrd-setup-root-after-ignition[1088]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:12:28.675955 initrd-setup-root-after-ignition[1088]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:12:28.687549 initrd-setup-root-after-ignition[1092]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:12:28.679849 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 20:12:28.684289 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 20:12:28.699731 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 20:12:28.722974 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 20:12:28.723096 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 20:12:28.729132 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 20:12:28.737028 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 20:12:28.740513 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 20:12:28.754503 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 20:12:28.768558 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 20:12:28.778455 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 20:12:28.791672 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:12:28.799238 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:12:28.801407 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 20:12:28.801905 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 20:12:28.802039 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 20:12:28.802776 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 20:12:28.803165 systemd[1]: Stopped target basic.target - Basic System. Feb 13 20:12:28.804019 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 20:12:28.804838 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 20:12:28.805357 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 20:12:28.805873 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 20:12:28.806286 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 20:12:28.806707 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 20:12:28.807107 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 20:12:28.807570 systemd[1]: Stopped target swap.target - Swaps. Feb 13 20:12:28.807940 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 20:12:28.808091 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 20:12:28.809341 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:12:28.809862 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:12:28.810242 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 20:12:28.848292 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:12:28.902679 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 20:12:28.902975 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 20:12:28.909110 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 20:12:28.909332 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 20:12:28.921223 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 20:12:28.924407 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 20:12:28.931307 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Feb 13 20:12:28.931499 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Feb 13 20:12:28.942577 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 20:12:28.945009 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 20:12:28.945178 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:12:28.954475 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 20:12:28.961794 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 20:12:28.967492 ignition[1112]: INFO : Ignition 2.20.0 Feb 13 20:12:28.967492 ignition[1112]: INFO : Stage: umount Feb 13 20:12:28.967492 ignition[1112]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:28.967492 ignition[1112]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:28.967492 ignition[1112]: INFO : umount: umount passed Feb 13 20:12:28.967492 ignition[1112]: INFO : Ignition finished successfully Feb 13 20:12:28.961980 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:12:28.967315 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 20:12:28.967474 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 20:12:28.973531 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 20:12:28.973622 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 20:12:29.003165 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 20:12:29.004061 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 20:12:29.011024 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 20:12:29.011153 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 20:12:29.019694 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 20:12:29.022294 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 20:12:29.025552 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 13 20:12:29.025613 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Feb 13 20:12:29.048487 systemd[1]: Stopped target network.target - Network. Feb 13 20:12:29.058765 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 20:12:29.059415 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 20:12:29.066432 systemd[1]: Stopped target paths.target - Path Units. Feb 13 20:12:29.072650 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 20:12:29.076175 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:12:29.084549 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 20:12:29.091978 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 20:12:29.098314 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 20:12:29.098630 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 20:12:29.110361 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 20:12:29.110433 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 20:12:29.116705 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 20:12:29.119843 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 20:12:29.128956 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 20:12:29.129041 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 20:12:29.134128 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 20:12:29.140839 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 20:12:29.161473 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 20:12:29.165305 systemd-networkd[861]: eth0: DHCPv6 lease lost Feb 13 20:12:29.169692 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 20:12:29.170048 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 20:12:29.177672 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 20:12:29.177793 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 20:12:29.185425 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 20:12:29.185499 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:12:29.196402 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 20:12:29.199559 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 20:12:29.199614 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 20:12:29.204465 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 20:12:29.204520 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:12:29.211209 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 20:12:29.211265 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 20:12:29.215820 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 20:12:29.215870 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:12:29.222547 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:12:29.246864 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 20:12:29.247043 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:12:29.252758 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 20:12:29.252795 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 20:12:29.253705 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 20:12:29.253735 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:12:29.254088 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 20:12:29.254128 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 20:12:29.257420 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 20:12:29.257462 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 20:12:29.258220 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 20:12:29.258597 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:29.301428 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 20:12:29.306872 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 20:12:29.306975 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:12:29.316170 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:12:29.319283 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:29.326653 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 20:12:29.326745 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 20:12:29.340269 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: Data path switched from VF: enP44129s1 Feb 13 20:12:29.358217 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 20:12:29.358357 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 20:12:30.170126 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 20:12:30.170278 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 20:12:30.173462 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 20:12:30.180437 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 20:12:30.180510 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 20:12:30.198418 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 20:12:30.207020 systemd[1]: Switching root. Feb 13 20:12:30.441945 systemd-journald[177]: Journal stopped Feb 13 20:12:19.117899 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p1) 13.3.1 20240614, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Feb 13 17:44:05 -00 2025 Feb 13 20:12:19.117937 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 20:12:19.117951 kernel: BIOS-provided physical RAM map: Feb 13 20:12:19.117962 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Feb 13 20:12:19.117972 kernel: BIOS-e820: [mem 0x00000000000c0000-0x00000000000fffff] reserved Feb 13 20:12:19.117983 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003ff40fff] usable Feb 13 20:12:19.117996 kernel: BIOS-e820: [mem 0x000000003ff41000-0x000000003ffc8fff] reserved Feb 13 20:12:19.118010 kernel: BIOS-e820: [mem 0x000000003ffc9000-0x000000003fffafff] ACPI data Feb 13 20:12:19.118021 kernel: BIOS-e820: [mem 0x000000003fffb000-0x000000003fffefff] ACPI NVS Feb 13 20:12:19.118032 kernel: BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] usable Feb 13 20:12:19.118044 kernel: BIOS-e820: [mem 0x0000000100000000-0x00000002bfffffff] usable Feb 13 20:12:19.118055 kernel: printk: bootconsole [earlyser0] enabled Feb 13 20:12:19.118066 kernel: NX (Execute Disable) protection: active Feb 13 20:12:19.118078 kernel: APIC: Static calls initialized Feb 13 20:12:19.118095 kernel: efi: EFI v2.7 by Microsoft Feb 13 20:12:19.118108 kernel: efi: ACPI=0x3fffa000 ACPI 2.0=0x3fffa014 SMBIOS=0x3ff85000 SMBIOS 3.0=0x3ff83000 MEMATTR=0x3ee83a98 RNG=0x3ffd1018 Feb 13 20:12:19.118121 kernel: random: crng init done Feb 13 20:12:19.118133 kernel: secureboot: Secure boot disabled Feb 13 20:12:19.118145 kernel: SMBIOS 3.1.0 present. Feb 13 20:12:19.118158 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 03/08/2024 Feb 13 20:12:19.118171 kernel: Hypervisor detected: Microsoft Hyper-V Feb 13 20:12:19.118183 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0x64e24, misc 0xbed7b2 Feb 13 20:12:19.118196 kernel: Hyper-V: Host Build 10.0.20348.1799-1-0 Feb 13 20:12:19.118208 kernel: Hyper-V: Nested features: 0x1e0101 Feb 13 20:12:19.118223 kernel: Hyper-V: LAPIC Timer Frequency: 0x30d40 Feb 13 20:12:19.118235 kernel: Hyper-V: Using hypercall for remote TLB flush Feb 13 20:12:19.118248 kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Feb 13 20:12:19.118260 kernel: clocksource: hyperv_clocksource_msr: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Feb 13 20:12:19.118273 kernel: tsc: Marking TSC unstable due to running on Hyper-V Feb 13 20:12:19.118286 kernel: tsc: Detected 2593.907 MHz processor Feb 13 20:12:19.118299 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 13 20:12:19.118312 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 13 20:12:19.118325 kernel: last_pfn = 0x2c0000 max_arch_pfn = 0x400000000 Feb 13 20:12:19.118340 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Feb 13 20:12:19.118353 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 13 20:12:19.118366 kernel: e820: update [mem 0x40000000-0xffffffff] usable ==> reserved Feb 13 20:12:19.118378 kernel: last_pfn = 0x40000 max_arch_pfn = 0x400000000 Feb 13 20:12:19.118391 kernel: Using GB pages for direct mapping Feb 13 20:12:19.118403 kernel: ACPI: Early table checksum verification disabled Feb 13 20:12:19.118416 kernel: ACPI: RSDP 0x000000003FFFA014 000024 (v02 VRTUAL) Feb 13 20:12:19.118434 kernel: ACPI: XSDT 0x000000003FFF90E8 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118451 kernel: ACPI: FACP 0x000000003FFF8000 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118464 kernel: ACPI: DSDT 0x000000003FFD6000 01E184 (v02 MSFTVM DSDT01 00000001 MSFT 05000000) Feb 13 20:12:19.118477 kernel: ACPI: FACS 0x000000003FFFE000 000040 Feb 13 20:12:19.118491 kernel: ACPI: OEM0 0x000000003FFF7000 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118505 kernel: ACPI: SPCR 0x000000003FFF6000 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118518 kernel: ACPI: WAET 0x000000003FFF5000 000028 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118543 kernel: ACPI: APIC 0x000000003FFD5000 000058 (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118557 kernel: ACPI: SRAT 0x000000003FFD4000 0002D0 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118570 kernel: ACPI: BGRT 0x000000003FFD3000 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118584 kernel: ACPI: FPDT 0x000000003FFD2000 000034 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Feb 13 20:12:19.118598 kernel: ACPI: Reserving FACP table memory at [mem 0x3fff8000-0x3fff8113] Feb 13 20:12:19.118611 kernel: ACPI: Reserving DSDT table memory at [mem 0x3ffd6000-0x3fff4183] Feb 13 20:12:19.118625 kernel: ACPI: Reserving FACS table memory at [mem 0x3fffe000-0x3fffe03f] Feb 13 20:12:19.118639 kernel: ACPI: Reserving OEM0 table memory at [mem 0x3fff7000-0x3fff7063] Feb 13 20:12:19.118652 kernel: ACPI: Reserving SPCR table memory at [mem 0x3fff6000-0x3fff604f] Feb 13 20:12:19.118669 kernel: ACPI: Reserving WAET table memory at [mem 0x3fff5000-0x3fff5027] Feb 13 20:12:19.118682 kernel: ACPI: Reserving APIC table memory at [mem 0x3ffd5000-0x3ffd5057] Feb 13 20:12:19.118696 kernel: ACPI: Reserving SRAT table memory at [mem 0x3ffd4000-0x3ffd42cf] Feb 13 20:12:19.118710 kernel: ACPI: Reserving BGRT table memory at [mem 0x3ffd3000-0x3ffd3037] Feb 13 20:12:19.118723 kernel: ACPI: Reserving FPDT table memory at [mem 0x3ffd2000-0x3ffd2033] Feb 13 20:12:19.118737 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Feb 13 20:12:19.118750 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Feb 13 20:12:19.118764 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Feb 13 20:12:19.118780 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x2bfffffff] hotplug Feb 13 20:12:19.118794 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2c0000000-0xfdfffffff] hotplug Feb 13 20:12:19.118807 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Feb 13 20:12:19.118821 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Feb 13 20:12:19.118834 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Feb 13 20:12:19.118848 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Feb 13 20:12:19.118862 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Feb 13 20:12:19.118875 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Feb 13 20:12:19.118889 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Feb 13 20:12:19.118905 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Feb 13 20:12:19.118919 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Feb 13 20:12:19.118932 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000000-0x1ffffffffffff] hotplug Feb 13 20:12:19.118946 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2000000000000-0x3ffffffffffff] hotplug Feb 13 20:12:19.118959 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x4000000000000-0x7ffffffffffff] hotplug Feb 13 20:12:19.118973 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x8000000000000-0xfffffffffffff] hotplug Feb 13 20:12:19.118986 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x2bfffffff] -> [mem 0x00000000-0x2bfffffff] Feb 13 20:12:19.119000 kernel: NODE_DATA(0) allocated [mem 0x2bfffa000-0x2bfffffff] Feb 13 20:12:19.119013 kernel: Zone ranges: Feb 13 20:12:19.119029 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 13 20:12:19.119043 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Feb 13 20:12:19.119056 kernel: Normal [mem 0x0000000100000000-0x00000002bfffffff] Feb 13 20:12:19.119070 kernel: Movable zone start for each node Feb 13 20:12:19.119083 kernel: Early memory node ranges Feb 13 20:12:19.119096 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Feb 13 20:12:19.119110 kernel: node 0: [mem 0x0000000000100000-0x000000003ff40fff] Feb 13 20:12:19.119123 kernel: node 0: [mem 0x000000003ffff000-0x000000003fffffff] Feb 13 20:12:19.119137 kernel: node 0: [mem 0x0000000100000000-0x00000002bfffffff] Feb 13 20:12:19.119153 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x00000002bfffffff] Feb 13 20:12:19.119167 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 20:12:19.119180 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Feb 13 20:12:19.119193 kernel: On node 0, zone DMA32: 190 pages in unavailable ranges Feb 13 20:12:19.119207 kernel: ACPI: PM-Timer IO Port: 0x408 Feb 13 20:12:19.119220 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] dfl dfl lint[0x1]) Feb 13 20:12:19.119234 kernel: IOAPIC[0]: apic_id 2, version 17, address 0xfec00000, GSI 0-23 Feb 13 20:12:19.119248 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 13 20:12:19.119261 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 13 20:12:19.119277 kernel: ACPI: SPCR: console: uart,io,0x3f8,115200 Feb 13 20:12:19.119291 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Feb 13 20:12:19.119304 kernel: [mem 0x40000000-0xffffffff] available for PCI devices Feb 13 20:12:19.119318 kernel: Booting paravirtualized kernel on Hyper-V Feb 13 20:12:19.119331 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 13 20:12:19.119345 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Feb 13 20:12:19.119356 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Feb 13 20:12:19.119368 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Feb 13 20:12:19.119379 kernel: pcpu-alloc: [0] 0 1 Feb 13 20:12:19.119393 kernel: Hyper-V: PV spinlocks enabled Feb 13 20:12:19.119405 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Feb 13 20:12:19.119418 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 20:12:19.119426 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 20:12:19.119433 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Feb 13 20:12:19.119440 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 20:12:19.119448 kernel: Fallback order for Node 0: 0 Feb 13 20:12:19.119455 kernel: Built 1 zonelists, mobility grouping on. Total pages: 2062618 Feb 13 20:12:19.119466 kernel: Policy zone: Normal Feb 13 20:12:19.119481 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 20:12:19.119489 kernel: software IO TLB: area num 2. Feb 13 20:12:19.119500 kernel: Memory: 8077024K/8387460K available (12288K kernel code, 2301K rwdata, 22736K rodata, 42976K init, 2216K bss, 310180K reserved, 0K cma-reserved) Feb 13 20:12:19.119510 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 13 20:12:19.119522 kernel: ftrace: allocating 37923 entries in 149 pages Feb 13 20:12:19.119550 kernel: ftrace: allocated 149 pages with 4 groups Feb 13 20:12:19.119562 kernel: Dynamic Preempt: voluntary Feb 13 20:12:19.119575 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 20:12:19.119589 kernel: rcu: RCU event tracing is enabled. Feb 13 20:12:19.119601 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 13 20:12:19.119618 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 20:12:19.119630 kernel: Rude variant of Tasks RCU enabled. Feb 13 20:12:19.119643 kernel: Tracing variant of Tasks RCU enabled. Feb 13 20:12:19.119657 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 20:12:19.119671 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 13 20:12:19.119702 kernel: Using NULL legacy PIC Feb 13 20:12:19.119732 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 0 Feb 13 20:12:19.119748 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 20:12:19.119760 kernel: Console: colour dummy device 80x25 Feb 13 20:12:19.119774 kernel: printk: console [tty1] enabled Feb 13 20:12:19.119789 kernel: printk: console [ttyS0] enabled Feb 13 20:12:19.119804 kernel: printk: bootconsole [earlyser0] disabled Feb 13 20:12:19.119817 kernel: ACPI: Core revision 20230628 Feb 13 20:12:19.119829 kernel: Failed to register legacy timer interrupt Feb 13 20:12:19.119842 kernel: APIC: Switch to symmetric I/O mode setup Feb 13 20:12:19.119858 kernel: Hyper-V: enabling crash_kexec_post_notifiers Feb 13 20:12:19.119873 kernel: Hyper-V: Using IPI hypercalls Feb 13 20:12:19.119886 kernel: APIC: send_IPI() replaced with hv_send_ipi() Feb 13 20:12:19.119898 kernel: APIC: send_IPI_mask() replaced with hv_send_ipi_mask() Feb 13 20:12:19.119910 kernel: APIC: send_IPI_mask_allbutself() replaced with hv_send_ipi_mask_allbutself() Feb 13 20:12:19.119923 kernel: APIC: send_IPI_allbutself() replaced with hv_send_ipi_allbutself() Feb 13 20:12:19.119936 kernel: APIC: send_IPI_all() replaced with hv_send_ipi_all() Feb 13 20:12:19.119948 kernel: APIC: send_IPI_self() replaced with hv_send_ipi_self() Feb 13 20:12:19.119964 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 5187.81 BogoMIPS (lpj=2593907) Feb 13 20:12:19.119988 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Feb 13 20:12:19.119999 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Feb 13 20:12:19.120010 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 13 20:12:19.120021 kernel: Spectre V2 : Mitigation: Retpolines Feb 13 20:12:19.120033 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 13 20:12:19.120046 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 13 20:12:19.120060 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Feb 13 20:12:19.120071 kernel: RETBleed: Vulnerable Feb 13 20:12:19.120083 kernel: Speculative Store Bypass: Vulnerable Feb 13 20:12:19.120095 kernel: TAA: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 20:12:19.120112 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 20:12:19.120126 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 13 20:12:19.120140 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 13 20:12:19.120152 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 13 20:12:19.120164 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Feb 13 20:12:19.120179 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Feb 13 20:12:19.120194 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Feb 13 20:12:19.120209 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 13 20:12:19.120223 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Feb 13 20:12:19.120238 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Feb 13 20:12:19.120252 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Feb 13 20:12:19.120270 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Feb 13 20:12:19.120285 kernel: Freeing SMP alternatives memory: 32K Feb 13 20:12:19.120299 kernel: pid_max: default: 32768 minimum: 301 Feb 13 20:12:19.120314 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 20:12:19.120329 kernel: landlock: Up and running. Feb 13 20:12:19.120343 kernel: SELinux: Initializing. Feb 13 20:12:19.120358 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.120372 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.120388 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8272CL CPU @ 2.60GHz (family: 0x6, model: 0x55, stepping: 0x7) Feb 13 20:12:19.120403 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:12:19.120418 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:12:19.120437 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:12:19.120452 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Feb 13 20:12:19.120467 kernel: signal: max sigframe size: 3632 Feb 13 20:12:19.120483 kernel: rcu: Hierarchical SRCU implementation. Feb 13 20:12:19.120498 kernel: rcu: Max phase no-delay instances is 400. Feb 13 20:12:19.120513 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Feb 13 20:12:19.120526 kernel: smp: Bringing up secondary CPUs ... Feb 13 20:12:19.120552 kernel: smpboot: x86: Booting SMP configuration: Feb 13 20:12:19.120566 kernel: .... node #0, CPUs: #1 Feb 13 20:12:19.120584 kernel: TAA CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/tsx_async_abort.html for more details. Feb 13 20:12:19.120600 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Feb 13 20:12:19.120614 kernel: smp: Brought up 1 node, 2 CPUs Feb 13 20:12:19.120628 kernel: smpboot: Max logical packages: 1 Feb 13 20:12:19.120642 kernel: smpboot: Total of 2 processors activated (10375.62 BogoMIPS) Feb 13 20:12:19.120656 kernel: devtmpfs: initialized Feb 13 20:12:19.120670 kernel: x86/mm: Memory block size: 128MB Feb 13 20:12:19.120684 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x3fffb000-0x3fffefff] (16384 bytes) Feb 13 20:12:19.120701 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 20:12:19.120716 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 13 20:12:19.120728 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 20:12:19.120740 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 20:12:19.120754 kernel: audit: initializing netlink subsys (disabled) Feb 13 20:12:19.120768 kernel: audit: type=2000 audit(1739477537.028:1): state=initialized audit_enabled=0 res=1 Feb 13 20:12:19.120782 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 20:12:19.120796 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 13 20:12:19.120810 kernel: cpuidle: using governor menu Feb 13 20:12:19.120826 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 20:12:19.120840 kernel: dca service started, version 1.12.1 Feb 13 20:12:19.120854 kernel: e820: reserve RAM buffer [mem 0x3ff41000-0x3fffffff] Feb 13 20:12:19.120869 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 13 20:12:19.120882 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 20:12:19.120897 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 20:12:19.120911 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 20:12:19.120928 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 20:12:19.120942 kernel: ACPI: Added _OSI(Module Device) Feb 13 20:12:19.120958 kernel: ACPI: Added _OSI(Processor Device) Feb 13 20:12:19.120973 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 20:12:19.120987 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 20:12:19.121001 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 20:12:19.121015 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Feb 13 20:12:19.121028 kernel: ACPI: Interpreter enabled Feb 13 20:12:19.121042 kernel: ACPI: PM: (supports S0 S5) Feb 13 20:12:19.121056 kernel: ACPI: Using IOAPIC for interrupt routing Feb 13 20:12:19.121070 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 13 20:12:19.121087 kernel: PCI: Ignoring E820 reservations for host bridge windows Feb 13 20:12:19.121101 kernel: ACPI: Enabled 1 GPEs in block 00 to 0F Feb 13 20:12:19.121115 kernel: iommu: Default domain type: Translated Feb 13 20:12:19.121129 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 13 20:12:19.121143 kernel: efivars: Registered efivars operations Feb 13 20:12:19.121157 kernel: PCI: Using ACPI for IRQ routing Feb 13 20:12:19.121171 kernel: PCI: System does not support PCI Feb 13 20:12:19.121184 kernel: vgaarb: loaded Feb 13 20:12:19.121198 kernel: clocksource: Switched to clocksource hyperv_clocksource_tsc_page Feb 13 20:12:19.121215 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 20:12:19.121229 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 20:12:19.121243 kernel: pnp: PnP ACPI init Feb 13 20:12:19.121257 kernel: pnp: PnP ACPI: found 3 devices Feb 13 20:12:19.121271 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 13 20:12:19.121285 kernel: NET: Registered PF_INET protocol family Feb 13 20:12:19.121299 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Feb 13 20:12:19.121313 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Feb 13 20:12:19.121327 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 20:12:19.121347 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 20:12:19.121361 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Feb 13 20:12:19.121375 kernel: TCP: Hash tables configured (established 65536 bind 65536) Feb 13 20:12:19.121388 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.121402 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Feb 13 20:12:19.121417 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 20:12:19.121429 kernel: NET: Registered PF_XDP protocol family Feb 13 20:12:19.121441 kernel: PCI: CLS 0 bytes, default 64 Feb 13 20:12:19.121454 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Feb 13 20:12:19.121471 kernel: software IO TLB: mapped [mem 0x000000003ad8c000-0x000000003ed8c000] (64MB) Feb 13 20:12:19.121485 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Feb 13 20:12:19.121498 kernel: Initialise system trusted keyrings Feb 13 20:12:19.121525 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Feb 13 20:12:19.121549 kernel: Key type asymmetric registered Feb 13 20:12:19.121560 kernel: Asymmetric key parser 'x509' registered Feb 13 20:12:19.121572 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Feb 13 20:12:19.121586 kernel: io scheduler mq-deadline registered Feb 13 20:12:19.121601 kernel: io scheduler kyber registered Feb 13 20:12:19.121621 kernel: io scheduler bfq registered Feb 13 20:12:19.121635 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 13 20:12:19.121649 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 20:12:19.121663 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 13 20:12:19.121678 kernel: 00:01: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Feb 13 20:12:19.121693 kernel: i8042: PNP: No PS/2 controller found. Feb 13 20:12:19.121883 kernel: rtc_cmos 00:02: registered as rtc0 Feb 13 20:12:19.122007 kernel: rtc_cmos 00:02: setting system clock to 2025-02-13T20:12:18 UTC (1739477538) Feb 13 20:12:19.122118 kernel: rtc_cmos 00:02: alarms up to one month, 114 bytes nvram Feb 13 20:12:19.122135 kernel: intel_pstate: CPU model not supported Feb 13 20:12:19.122149 kernel: efifb: probing for efifb Feb 13 20:12:19.122164 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Feb 13 20:12:19.122177 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Feb 13 20:12:19.122191 kernel: efifb: scrolling: redraw Feb 13 20:12:19.122205 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Feb 13 20:12:19.122219 kernel: Console: switching to colour frame buffer device 128x48 Feb 13 20:12:19.122236 kernel: fb0: EFI VGA frame buffer device Feb 13 20:12:19.122250 kernel: pstore: Using crash dump compression: deflate Feb 13 20:12:19.122264 kernel: pstore: Registered efi_pstore as persistent store backend Feb 13 20:12:19.122278 kernel: NET: Registered PF_INET6 protocol family Feb 13 20:12:19.122291 kernel: Segment Routing with IPv6 Feb 13 20:12:19.122305 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 20:12:19.122319 kernel: NET: Registered PF_PACKET protocol family Feb 13 20:12:19.122333 kernel: Key type dns_resolver registered Feb 13 20:12:19.122347 kernel: IPI shorthand broadcast: enabled Feb 13 20:12:19.122361 kernel: sched_clock: Marking stable (902002900, 54173500)->(1197844200, -241667800) Feb 13 20:12:19.122378 kernel: registered taskstats version 1 Feb 13 20:12:19.122392 kernel: Loading compiled-in X.509 certificates Feb 13 20:12:19.122405 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 0cc219a306b9e46e583adebba1820decbdc4307b' Feb 13 20:12:19.122419 kernel: Key type .fscrypt registered Feb 13 20:12:19.122432 kernel: Key type fscrypt-provisioning registered Feb 13 20:12:19.122446 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 20:12:19.122460 kernel: ima: Allocated hash algorithm: sha1 Feb 13 20:12:19.122474 kernel: ima: No architecture policies found Feb 13 20:12:19.122490 kernel: clk: Disabling unused clocks Feb 13 20:12:19.122505 kernel: Freeing unused kernel image (initmem) memory: 42976K Feb 13 20:12:19.122518 kernel: Write protecting the kernel read-only data: 36864k Feb 13 20:12:19.122544 kernel: Freeing unused kernel image (rodata/data gap) memory: 1840K Feb 13 20:12:19.122567 kernel: Run /init as init process Feb 13 20:12:19.122581 kernel: with arguments: Feb 13 20:12:19.122594 kernel: /init Feb 13 20:12:19.122608 kernel: with environment: Feb 13 20:12:19.122621 kernel: HOME=/ Feb 13 20:12:19.122634 kernel: TERM=linux Feb 13 20:12:19.122651 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 20:12:19.122668 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 20:12:19.122684 systemd[1]: Detected virtualization microsoft. Feb 13 20:12:19.122709 systemd[1]: Detected architecture x86-64. Feb 13 20:12:19.122722 systemd[1]: Running in initrd. Feb 13 20:12:19.122737 systemd[1]: No hostname configured, using default hostname. Feb 13 20:12:19.122750 systemd[1]: Hostname set to . Feb 13 20:12:19.122774 systemd[1]: Initializing machine ID from random generator. Feb 13 20:12:19.122789 systemd[1]: Queued start job for default target initrd.target. Feb 13 20:12:19.122805 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:12:19.122821 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:12:19.122838 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 20:12:19.122855 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 20:12:19.122870 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 20:12:19.122886 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 20:12:19.122908 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 20:12:19.122923 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 20:12:19.122940 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:12:19.122956 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:12:19.122971 systemd[1]: Reached target paths.target - Path Units. Feb 13 20:12:19.122987 systemd[1]: Reached target slices.target - Slice Units. Feb 13 20:12:19.123003 systemd[1]: Reached target swap.target - Swaps. Feb 13 20:12:19.123022 systemd[1]: Reached target timers.target - Timer Units. Feb 13 20:12:19.123037 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 20:12:19.123053 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 20:12:19.123070 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 20:12:19.123086 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Feb 13 20:12:19.123102 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:12:19.123118 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 20:12:19.123133 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:12:19.123152 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 20:12:19.123168 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 20:12:19.123184 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 20:12:19.123200 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 20:12:19.123216 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 20:12:19.123232 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 20:12:19.123248 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 20:12:19.123264 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:19.123303 systemd-journald[177]: Collecting audit messages is disabled. Feb 13 20:12:19.123342 systemd-journald[177]: Journal started Feb 13 20:12:19.123375 systemd-journald[177]: Runtime Journal (/run/log/journal/0238720c1c9a40a98bc505e0f3e0b423) is 8.0M, max 158.8M, 150.8M free. Feb 13 20:12:19.124346 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 20:12:19.120010 systemd-modules-load[178]: Inserted module 'overlay' Feb 13 20:12:19.137623 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 20:12:19.141601 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:12:19.149975 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 20:12:19.166828 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 20:12:19.178341 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 20:12:19.178378 kernel: Bridge firewalling registered Feb 13 20:12:19.180623 systemd-modules-load[178]: Inserted module 'br_netfilter' Feb 13 20:12:19.183117 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 20:12:19.189366 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 20:12:19.195066 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:19.201198 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 20:12:19.206763 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:12:19.218698 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:12:19.225833 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 20:12:19.233980 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 20:12:19.250465 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:12:19.257347 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:12:19.266691 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 20:12:19.272079 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:19.279727 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 20:12:19.302214 dracut-cmdline[213]: dracut-dracut-053 Feb 13 20:12:19.306603 dracut-cmdline[213]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 20:12:19.314471 systemd-resolved[208]: Positive Trust Anchors: Feb 13 20:12:19.314487 systemd-resolved[208]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 20:12:19.314525 systemd-resolved[208]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 20:12:19.317164 systemd-resolved[208]: Defaulting to hostname 'linux'. Feb 13 20:12:19.318152 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 20:12:19.326765 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:12:19.415558 kernel: SCSI subsystem initialized Feb 13 20:12:19.425551 kernel: Loading iSCSI transport class v2.0-870. Feb 13 20:12:19.436556 kernel: iscsi: registered transport (tcp) Feb 13 20:12:19.457521 kernel: iscsi: registered transport (qla4xxx) Feb 13 20:12:19.457605 kernel: QLogic iSCSI HBA Driver Feb 13 20:12:19.492273 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 20:12:19.505798 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 20:12:19.535774 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 20:12:19.535884 kernel: device-mapper: uevent: version 1.0.3 Feb 13 20:12:19.539916 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 20:12:19.581562 kernel: raid6: avx512x4 gen() 18512 MB/s Feb 13 20:12:19.600545 kernel: raid6: avx512x2 gen() 18467 MB/s Feb 13 20:12:19.619540 kernel: raid6: avx512x1 gen() 18291 MB/s Feb 13 20:12:19.638552 kernel: raid6: avx2x4 gen() 18336 MB/s Feb 13 20:12:19.657545 kernel: raid6: avx2x2 gen() 18313 MB/s Feb 13 20:12:19.677722 kernel: raid6: avx2x1 gen() 13940 MB/s Feb 13 20:12:19.677769 kernel: raid6: using algorithm avx512x4 gen() 18512 MB/s Feb 13 20:12:19.699591 kernel: raid6: .... xor() 6457 MB/s, rmw enabled Feb 13 20:12:19.699629 kernel: raid6: using avx512x2 recovery algorithm Feb 13 20:12:19.721560 kernel: xor: automatically using best checksumming function avx Feb 13 20:12:19.872560 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 20:12:19.881760 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 20:12:19.892696 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:12:19.907573 systemd-udevd[395]: Using default interface naming scheme 'v255'. Feb 13 20:12:19.912121 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:12:19.930730 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 20:12:19.944892 dracut-pre-trigger[405]: rd.md=0: removing MD RAID activation Feb 13 20:12:19.972261 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 20:12:19.997707 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 20:12:20.040444 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:12:20.051698 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 20:12:20.069927 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 20:12:20.077999 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 20:12:20.084800 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:12:20.087934 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 20:12:20.102818 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 20:12:20.124996 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 20:12:20.137366 kernel: cryptd: max_cpu_qlen set to 1000 Feb 13 20:12:20.161192 kernel: AVX2 version of gcm_enc/dec engaged. Feb 13 20:12:20.161270 kernel: AES CTR mode by8 optimization enabled Feb 13 20:12:20.163307 kernel: hv_vmbus: Vmbus version:5.2 Feb 13 20:12:20.172294 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 20:12:20.177786 kernel: hv_vmbus: registering driver hyperv_keyboard Feb 13 20:12:20.173319 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:20.207067 kernel: pps_core: LinuxPPS API ver. 1 registered Feb 13 20:12:20.207162 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/VMBUS:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Feb 13 20:12:20.207189 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Feb 13 20:12:20.193101 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:12:20.212160 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:12:20.212469 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:20.224440 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:20.237012 kernel: PTP clock support registered Feb 13 20:12:20.235990 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:20.248841 kernel: hv_utils: Registering HyperV Utility Driver Feb 13 20:12:20.248892 kernel: hv_vmbus: registering driver hv_utils Feb 13 20:12:20.248906 kernel: hv_utils: Heartbeat IC version 3.0 Feb 13 20:12:20.251552 kernel: hv_utils: Shutdown IC version 3.2 Feb 13 20:12:20.801236 kernel: hv_utils: TimeSync IC version 4.0 Feb 13 20:12:20.800279 systemd-resolved[208]: Clock change detected. Flushing caches. Feb 13 20:12:20.813641 kernel: hid: raw HID events driver (C) Jiri Kosina Feb 13 20:12:20.818401 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:12:20.848955 kernel: hv_vmbus: registering driver hv_storvsc Feb 13 20:12:20.848987 kernel: hv_vmbus: registering driver hv_netvsc Feb 13 20:12:20.849007 kernel: scsi host1: storvsc_host_t Feb 13 20:12:20.849214 kernel: hv_vmbus: registering driver hid_hyperv Feb 13 20:12:20.849233 kernel: scsi host0: storvsc_host_t Feb 13 20:12:20.849524 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Feb 13 20:12:20.849545 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Feb 13 20:12:20.855732 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Feb 13 20:12:20.855783 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Feb 13 20:12:20.818523 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:20.850364 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:20.887875 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Feb 13 20:12:20.890119 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Feb 13 20:12:20.890145 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Feb 13 20:12:20.888376 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:20.905205 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Feb 13 20:12:20.923400 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Feb 13 20:12:20.923588 kernel: sd 0:0:0:0: [sda] Write Protect is off Feb 13 20:12:20.923763 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Feb 13 20:12:20.923930 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Feb 13 20:12:20.924107 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:20.924128 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Feb 13 20:12:20.905971 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:12:20.935324 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:20.972186 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: VF slot 1 added Feb 13 20:12:20.980290 kernel: hv_vmbus: registering driver hv_pci Feb 13 20:12:20.986033 kernel: hv_pci a9d58489-ac61-4534-a070-abd92e6109ea: PCI VMBus probing: Using version 0x10004 Feb 13 20:12:21.031583 kernel: hv_pci a9d58489-ac61-4534-a070-abd92e6109ea: PCI host bridge to bus ac61:00 Feb 13 20:12:21.031879 kernel: pci_bus ac61:00: root bus resource [mem 0xfe0000000-0xfe00fffff window] Feb 13 20:12:21.032131 kernel: pci_bus ac61:00: No busn resource found for root bus, will use [bus 00-ff] Feb 13 20:12:21.032370 kernel: pci ac61:00:02.0: [15b3:1016] type 00 class 0x020000 Feb 13 20:12:21.032623 kernel: pci ac61:00:02.0: reg 0x10: [mem 0xfe0000000-0xfe00fffff 64bit pref] Feb 13 20:12:21.032838 kernel: pci ac61:00:02.0: enabling Extended Tags Feb 13 20:12:21.033066 kernel: pci ac61:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at ac61:00:02.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Feb 13 20:12:21.033321 kernel: pci_bus ac61:00: busn_res: [bus 00-ff] end is updated to 00 Feb 13 20:12:21.033562 kernel: pci ac61:00:02.0: BAR 0: assigned [mem 0xfe0000000-0xfe00fffff 64bit pref] Feb 13 20:12:21.206478 kernel: mlx5_core ac61:00:02.0: enabling device (0000 -> 0002) Feb 13 20:12:21.442179 kernel: mlx5_core ac61:00:02.0: firmware version: 14.30.5000 Feb 13 20:12:21.442437 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: VF registering: eth1 Feb 13 20:12:21.442909 kernel: mlx5_core ac61:00:02.0 eth1: joined to eth0 Feb 13 20:12:21.443660 kernel: mlx5_core ac61:00:02.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Feb 13 20:12:21.407916 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Feb 13 20:12:21.450273 kernel: mlx5_core ac61:00:02.0 enP44129s1: renamed from eth1 Feb 13 20:12:21.522311 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by (udev-worker) (447) Feb 13 20:12:21.538114 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Feb 13 20:12:21.555285 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Feb 13 20:12:21.570791 kernel: BTRFS: device fsid e9c87d9f-3864-4b45-9be4-80a5397f1fc6 devid 1 transid 38 /dev/sda3 scanned by (udev-worker) (441) Feb 13 20:12:21.589007 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Feb 13 20:12:21.592296 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Feb 13 20:12:21.608411 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 20:12:21.624283 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:21.630268 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:22.637670 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 20:12:22.637750 disk-uuid[601]: The operation has completed successfully. Feb 13 20:12:22.709138 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 20:12:22.709264 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 20:12:22.738607 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 20:12:22.747761 sh[687]: Success Feb 13 20:12:22.781571 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Feb 13 20:12:22.977608 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 20:12:22.992387 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 20:12:22.997387 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 20:12:23.012268 kernel: BTRFS info (device dm-0): first mount of filesystem e9c87d9f-3864-4b45-9be4-80a5397f1fc6 Feb 13 20:12:23.012330 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:23.017909 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 20:12:23.020573 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 20:12:23.023041 kernel: BTRFS info (device dm-0): using free space tree Feb 13 20:12:23.305474 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 20:12:23.310916 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 20:12:23.326441 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 20:12:23.334292 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 20:12:23.348399 kernel: BTRFS info (device sda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:23.348457 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:23.348474 kernel: BTRFS info (device sda6): using free space tree Feb 13 20:12:23.426727 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 20:12:23.438285 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 20:12:23.440731 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 20:12:23.452222 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 20:12:23.458439 kernel: BTRFS info (device sda6): last unmount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:23.465047 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 20:12:23.473969 systemd-networkd[861]: lo: Link UP Feb 13 20:12:23.474137 systemd-networkd[861]: lo: Gained carrier Feb 13 20:12:23.475457 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 20:12:23.479154 systemd-networkd[861]: Enumeration completed Feb 13 20:12:23.481829 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 20:12:23.482671 systemd-networkd[861]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:12:23.482676 systemd-networkd[861]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 20:12:23.490732 systemd[1]: Reached target network.target - Network. Feb 13 20:12:23.551280 kernel: mlx5_core ac61:00:02.0 enP44129s1: Link up Feb 13 20:12:23.591393 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: Data path switched to VF: enP44129s1 Feb 13 20:12:23.591551 systemd-networkd[861]: enP44129s1: Link UP Feb 13 20:12:23.591689 systemd-networkd[861]: eth0: Link UP Feb 13 20:12:23.596894 systemd-networkd[861]: eth0: Gained carrier Feb 13 20:12:23.596909 systemd-networkd[861]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:12:23.607514 systemd-networkd[861]: enP44129s1: Gained carrier Feb 13 20:12:23.635335 systemd-networkd[861]: eth0: DHCPv4 address 10.200.4.48/24, gateway 10.200.4.1 acquired from 168.63.129.16 Feb 13 20:12:24.445969 ignition[871]: Ignition 2.20.0 Feb 13 20:12:24.445981 ignition[871]: Stage: fetch-offline Feb 13 20:12:24.446032 ignition[871]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.446042 ignition[871]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.446170 ignition[871]: parsed url from cmdline: "" Feb 13 20:12:24.446175 ignition[871]: no config URL provided Feb 13 20:12:24.446186 ignition[871]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 20:12:24.446196 ignition[871]: no config at "/usr/lib/ignition/user.ign" Feb 13 20:12:24.446205 ignition[871]: failed to fetch config: resource requires networking Feb 13 20:12:24.462531 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 20:12:24.446535 ignition[871]: Ignition finished successfully Feb 13 20:12:24.484537 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Feb 13 20:12:24.502806 ignition[880]: Ignition 2.20.0 Feb 13 20:12:24.502818 ignition[880]: Stage: fetch Feb 13 20:12:24.503035 ignition[880]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.503048 ignition[880]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.503438 ignition[880]: parsed url from cmdline: "" Feb 13 20:12:24.503443 ignition[880]: no config URL provided Feb 13 20:12:24.503448 ignition[880]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 20:12:24.503457 ignition[880]: no config at "/usr/lib/ignition/user.ign" Feb 13 20:12:24.503482 ignition[880]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Feb 13 20:12:24.595897 ignition[880]: GET result: OK Feb 13 20:12:24.596031 ignition[880]: config has been read from IMDS userdata Feb 13 20:12:24.596072 ignition[880]: parsing config with SHA512: 34218719cc285cb68475ae095c7d73c3f3dd2930cb0ce913866ad9bbd923829ca17f50ff3b0bcaa227592c93966bca55f0102758c079459dd0db3b87cd78c264 Feb 13 20:12:24.605212 unknown[880]: fetched base config from "system" Feb 13 20:12:24.605551 ignition[880]: fetch: fetch complete Feb 13 20:12:24.605218 unknown[880]: fetched base config from "system" Feb 13 20:12:24.605556 ignition[880]: fetch: fetch passed Feb 13 20:12:24.605224 unknown[880]: fetched user config from "azure" Feb 13 20:12:24.605600 ignition[880]: Ignition finished successfully Feb 13 20:12:24.607319 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Feb 13 20:12:24.627460 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 20:12:24.646625 ignition[886]: Ignition 2.20.0 Feb 13 20:12:24.646637 ignition[886]: Stage: kargs Feb 13 20:12:24.646853 ignition[886]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.646866 ignition[886]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.647752 ignition[886]: kargs: kargs passed Feb 13 20:12:24.647805 ignition[886]: Ignition finished successfully Feb 13 20:12:24.658355 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 20:12:24.668563 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 20:12:24.681303 ignition[892]: Ignition 2.20.0 Feb 13 20:12:24.681314 ignition[892]: Stage: disks Feb 13 20:12:24.683434 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 20:12:24.681543 ignition[892]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:24.681557 ignition[892]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:24.682464 ignition[892]: disks: disks passed Feb 13 20:12:24.682512 ignition[892]: Ignition finished successfully Feb 13 20:12:24.698235 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 20:12:24.701105 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 20:12:24.704734 systemd-networkd[861]: eth0: Gained IPv6LL Feb 13 20:12:24.707593 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 20:12:24.707922 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 20:12:24.708324 systemd[1]: Reached target basic.target - Basic System. Feb 13 20:12:24.721529 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 20:12:24.785872 systemd-fsck[900]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Feb 13 20:12:24.792633 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 20:12:24.805564 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 20:12:24.903300 kernel: EXT4-fs (sda9): mounted filesystem c5993b0e-9201-4b44-aa01-79dc9d6c9fc9 r/w with ordered data mode. Quota mode: none. Feb 13 20:12:24.905182 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 20:12:24.912665 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 20:12:24.962406 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 20:12:24.971650 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 20:12:24.979954 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (911) Feb 13 20:12:24.985864 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Feb 13 20:12:25.003937 kernel: BTRFS info (device sda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:25.003976 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:25.003993 kernel: BTRFS info (device sda6): using free space tree Feb 13 20:12:25.004011 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 20:12:24.996159 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 20:12:24.996197 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 20:12:25.012567 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 20:12:25.015099 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 20:12:25.024440 systemd-networkd[861]: enP44129s1: Gained IPv6LL Feb 13 20:12:25.028412 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 20:12:25.572517 coreos-metadata[913]: Feb 13 20:12:25.572 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Feb 13 20:12:25.579576 coreos-metadata[913]: Feb 13 20:12:25.579 INFO Fetch successful Feb 13 20:12:25.579576 coreos-metadata[913]: Feb 13 20:12:25.579 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Feb 13 20:12:25.593007 coreos-metadata[913]: Feb 13 20:12:25.592 INFO Fetch successful Feb 13 20:12:25.607907 coreos-metadata[913]: Feb 13 20:12:25.607 INFO wrote hostname ci-4152.2.1-a-2754074fca to /sysroot/etc/hostname Feb 13 20:12:25.616139 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Feb 13 20:12:25.681859 initrd-setup-root[942]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 20:12:25.718110 initrd-setup-root[949]: cut: /sysroot/etc/group: No such file or directory Feb 13 20:12:25.744482 initrd-setup-root[956]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 20:12:25.750067 initrd-setup-root[963]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 20:12:26.698894 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 20:12:26.708573 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 20:12:26.715464 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 20:12:26.728211 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 20:12:26.734353 kernel: BTRFS info (device sda6): last unmount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:26.749655 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 20:12:26.761879 ignition[1031]: INFO : Ignition 2.20.0 Feb 13 20:12:26.761879 ignition[1031]: INFO : Stage: mount Feb 13 20:12:26.768533 ignition[1031]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:26.768533 ignition[1031]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:26.768533 ignition[1031]: INFO : mount: mount passed Feb 13 20:12:26.768533 ignition[1031]: INFO : Ignition finished successfully Feb 13 20:12:26.764067 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 20:12:26.778530 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 20:12:26.796508 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 20:12:26.809285 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/sda6 scanned by mount (1042) Feb 13 20:12:26.815906 kernel: BTRFS info (device sda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 20:12:26.816002 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:12:26.818418 kernel: BTRFS info (device sda6): using free space tree Feb 13 20:12:26.823276 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 20:12:26.825741 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 20:12:26.850885 ignition[1059]: INFO : Ignition 2.20.0 Feb 13 20:12:26.850885 ignition[1059]: INFO : Stage: files Feb 13 20:12:26.855206 ignition[1059]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:26.855206 ignition[1059]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:26.855206 ignition[1059]: DEBUG : files: compiled without relabeling support, skipping Feb 13 20:12:26.868896 ignition[1059]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 20:12:26.872735 ignition[1059]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 20:12:26.922582 ignition[1059]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 20:12:26.926815 ignition[1059]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 20:12:26.926815 ignition[1059]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 20:12:26.923163 unknown[1059]: wrote ssh authorized keys file for user: core Feb 13 20:12:26.941009 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 20:12:26.946383 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Feb 13 20:12:26.983834 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 20:12:27.112299 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 20:12:27.118676 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Feb 13 20:12:27.118676 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Feb 13 20:12:27.614810 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Feb 13 20:12:27.697614 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Feb 13 20:12:27.702334 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Feb 13 20:12:27.706600 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 20:12:27.706600 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 20:12:27.715238 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 20:12:27.719662 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 20:12:27.723923 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 20:12:27.728234 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 20:12:27.736650 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 20:12:27.741281 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 20:12:27.745728 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 20:12:27.750025 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:27.756355 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:27.762496 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:27.767818 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 Feb 13 20:12:28.242895 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Feb 13 20:12:28.557755 ignition[1059]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:12:28.557755 ignition[1059]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Feb 13 20:12:28.591382 ignition[1059]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 20:12:28.602094 ignition[1059]: INFO : files: files passed Feb 13 20:12:28.602094 ignition[1059]: INFO : Ignition finished successfully Feb 13 20:12:28.593287 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 20:12:28.610540 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 20:12:28.617813 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 20:12:28.664998 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 20:12:28.665119 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 20:12:28.675955 initrd-setup-root-after-ignition[1088]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:12:28.675955 initrd-setup-root-after-ignition[1088]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:12:28.687549 initrd-setup-root-after-ignition[1092]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:12:28.679849 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 20:12:28.684289 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 20:12:28.699731 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 20:12:28.722974 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 20:12:28.723096 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 20:12:28.729132 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 20:12:28.737028 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 20:12:28.740513 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 20:12:28.754503 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 20:12:28.768558 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 20:12:28.778455 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 20:12:28.791672 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:12:28.799238 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:12:28.801407 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 20:12:28.801905 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 20:12:28.802039 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 20:12:28.802776 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 20:12:28.803165 systemd[1]: Stopped target basic.target - Basic System. Feb 13 20:12:28.804019 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 20:12:28.804838 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 20:12:28.805357 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 20:12:28.805873 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 20:12:28.806286 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 20:12:28.806707 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 20:12:28.807107 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 20:12:28.807570 systemd[1]: Stopped target swap.target - Swaps. Feb 13 20:12:28.807940 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 20:12:28.808091 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 20:12:28.809341 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:12:28.809862 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:12:28.810242 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 20:12:28.848292 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:12:28.902679 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 20:12:28.902975 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 20:12:28.909110 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 20:12:28.909332 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 20:12:28.921223 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 20:12:28.924407 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 20:12:28.931307 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Feb 13 20:12:28.931499 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Feb 13 20:12:28.942577 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 20:12:28.945009 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 20:12:28.945178 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:12:28.954475 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 20:12:28.961794 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 20:12:28.967492 ignition[1112]: INFO : Ignition 2.20.0 Feb 13 20:12:28.967492 ignition[1112]: INFO : Stage: umount Feb 13 20:12:28.967492 ignition[1112]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:12:28.967492 ignition[1112]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Feb 13 20:12:28.967492 ignition[1112]: INFO : umount: umount passed Feb 13 20:12:28.967492 ignition[1112]: INFO : Ignition finished successfully Feb 13 20:12:28.961980 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:12:28.967315 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 20:12:28.967474 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 20:12:28.973531 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 20:12:28.973622 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 20:12:29.003165 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 20:12:29.004061 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 20:12:29.011024 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 20:12:29.011153 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 20:12:29.019694 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 20:12:29.022294 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 20:12:29.025552 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 13 20:12:29.025613 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Feb 13 20:12:29.048487 systemd[1]: Stopped target network.target - Network. Feb 13 20:12:29.058765 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 20:12:29.059415 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 20:12:29.066432 systemd[1]: Stopped target paths.target - Path Units. Feb 13 20:12:29.072650 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 20:12:29.076175 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:12:29.084549 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 20:12:29.091978 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 20:12:29.098314 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 20:12:29.098630 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 20:12:29.110361 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 20:12:29.110433 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 20:12:29.116705 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 20:12:29.119843 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 20:12:29.128956 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 20:12:29.129041 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 20:12:29.134128 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 20:12:29.140839 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 20:12:29.161473 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 20:12:29.165305 systemd-networkd[861]: eth0: DHCPv6 lease lost Feb 13 20:12:29.169692 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 20:12:29.170048 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 20:12:29.177672 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 20:12:29.177793 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 20:12:29.185425 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 20:12:29.185499 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:12:29.196402 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 20:12:29.199559 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 20:12:29.199614 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 20:12:29.204465 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 20:12:29.204520 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:12:29.211209 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 20:12:29.211265 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 20:12:29.215820 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 20:12:29.215870 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:12:29.222547 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:12:29.246864 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 20:12:29.247043 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:12:29.252758 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 20:12:29.252795 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 20:12:29.253705 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 20:12:29.253735 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:12:29.254088 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 20:12:29.254128 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 20:12:29.257420 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 20:12:29.257462 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 20:12:29.258220 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 20:12:29.258597 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:12:29.301428 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 20:12:29.306872 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 20:12:29.306975 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:12:29.316170 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:12:29.319283 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:29.326653 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 20:12:29.326745 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 20:12:29.340269 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: Data path switched from VF: enP44129s1 Feb 13 20:12:29.358217 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 20:12:29.358357 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 20:12:30.170126 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 20:12:30.170278 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 20:12:30.173462 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 20:12:30.180437 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 20:12:30.180510 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 20:12:30.198418 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 20:12:30.207020 systemd[1]: Switching root. Feb 13 20:12:30.441945 systemd-journald[177]: Journal stopped Feb 13 20:12:36.875551 systemd-journald[177]: Received SIGTERM from PID 1 (systemd). Feb 13 20:12:36.875626 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 20:12:36.875648 kernel: SELinux: policy capability open_perms=1 Feb 13 20:12:36.875661 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 20:12:36.875673 kernel: SELinux: policy capability always_check_network=0 Feb 13 20:12:36.875686 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 20:12:36.875702 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 20:12:36.875720 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 20:12:36.875736 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 20:12:36.875751 kernel: audit: type=1403 audit(1739477551.694:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 20:12:36.875769 systemd[1]: Successfully loaded SELinux policy in 175.362ms. Feb 13 20:12:36.875788 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 9.847ms. Feb 13 20:12:36.875809 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 20:12:36.875829 systemd[1]: Detected virtualization microsoft. Feb 13 20:12:36.875848 systemd[1]: Detected architecture x86-64. Feb 13 20:12:36.875862 systemd[1]: Detected first boot. Feb 13 20:12:36.875872 systemd[1]: Hostname set to . Feb 13 20:12:36.875884 systemd[1]: Initializing machine ID from random generator. Feb 13 20:12:36.875896 zram_generator::config[1155]: No configuration found. Feb 13 20:12:36.875912 systemd[1]: Populated /etc with preset unit settings. Feb 13 20:12:36.875926 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 20:12:36.875939 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 20:12:36.875952 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 20:12:36.875963 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 20:12:36.875976 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 20:12:36.875987 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 20:12:36.876002 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 20:12:36.876014 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 20:12:36.876025 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 20:12:36.876037 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 20:12:36.876048 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 20:12:36.876059 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:12:36.876071 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:12:36.876082 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 20:12:36.876096 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 20:12:36.876108 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 20:12:36.876120 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 20:12:36.876130 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Feb 13 20:12:36.876142 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:12:36.876153 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 20:12:36.876169 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 20:12:36.876179 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 20:12:36.876193 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 20:12:36.876204 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:12:36.876217 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 20:12:36.876227 systemd[1]: Reached target slices.target - Slice Units. Feb 13 20:12:36.876239 systemd[1]: Reached target swap.target - Swaps. Feb 13 20:12:36.876250 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 20:12:36.876276 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 20:12:36.876291 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:12:36.876302 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 20:12:36.876316 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:12:36.876327 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 20:12:36.876340 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 20:12:36.876355 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 20:12:36.876366 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 20:12:36.876379 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:36.876390 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 20:12:36.876403 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 20:12:36.876414 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 20:12:36.876428 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 20:12:36.876441 systemd[1]: Reached target machines.target - Containers. Feb 13 20:12:36.876455 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 20:12:36.876469 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:12:36.876480 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 20:12:36.876493 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 20:12:36.876504 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 20:12:36.876517 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 20:12:36.876529 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 20:12:36.876540 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 20:12:36.876552 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 20:12:36.876567 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 20:12:36.876581 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 20:12:36.876592 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 20:12:36.876605 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 20:12:36.876615 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 20:12:36.876629 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 20:12:36.876640 kernel: fuse: init (API version 7.39) Feb 13 20:12:36.876652 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 20:12:36.876666 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 20:12:36.876678 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 20:12:36.876690 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 20:12:36.876702 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 20:12:36.876714 systemd[1]: Stopped verity-setup.service. Feb 13 20:12:36.876728 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:36.876741 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 20:12:36.876753 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 20:12:36.876767 kernel: loop: module loaded Feb 13 20:12:36.876780 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 20:12:36.876790 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 20:12:36.876804 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 20:12:36.876849 systemd-journald[1244]: Collecting audit messages is disabled. Feb 13 20:12:36.876882 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 20:12:36.876896 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 20:12:36.876907 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:12:36.876920 kernel: ACPI: bus type drm_connector registered Feb 13 20:12:36.876930 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 20:12:36.876944 systemd-journald[1244]: Journal started Feb 13 20:12:36.876971 systemd-journald[1244]: Runtime Journal (/run/log/journal/3b88622a88394c38912e5e12fea0836b) is 8.0M, max 158.8M, 150.8M free. Feb 13 20:12:35.960179 systemd[1]: Queued start job for default target multi-user.target. Feb 13 20:12:36.192737 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Feb 13 20:12:36.193128 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 20:12:36.887893 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 20:12:36.894564 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 20:12:36.896139 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 20:12:36.896618 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 20:12:36.900414 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 20:12:36.900671 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 20:12:36.904504 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 20:12:36.904843 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 20:12:36.909246 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 20:12:36.909559 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 20:12:36.913331 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 20:12:36.913617 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 20:12:36.917368 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 20:12:36.921391 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 20:12:36.927932 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 20:12:36.946680 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:12:36.951501 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 20:12:36.960561 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 20:12:36.969355 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 20:12:36.973448 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 20:12:36.973648 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 20:12:36.978285 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Feb 13 20:12:36.990462 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 20:12:36.995653 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 20:12:36.998528 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:12:37.034516 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 20:12:37.040626 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 20:12:37.043818 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 20:12:37.045475 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 20:12:37.048679 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 20:12:37.052431 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 20:12:37.056752 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 20:12:37.065481 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 20:12:37.077339 systemd-journald[1244]: Time spent on flushing to /var/log/journal/3b88622a88394c38912e5e12fea0836b is 20.678ms for 959 entries. Feb 13 20:12:37.077339 systemd-journald[1244]: System Journal (/var/log/journal/3b88622a88394c38912e5e12fea0836b) is 8.0M, max 2.6G, 2.6G free. Feb 13 20:12:37.150523 systemd-journald[1244]: Received client request to flush runtime journal. Feb 13 20:12:37.081115 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 20:12:37.088103 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 20:12:37.100443 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 20:12:37.105003 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 20:12:37.118470 udevadm[1294]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Feb 13 20:12:37.131608 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 20:12:37.138236 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 20:12:37.157565 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Feb 13 20:12:37.161894 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 20:12:37.175308 kernel: loop0: detected capacity change from 0 to 138184 Feb 13 20:12:37.201822 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 20:12:37.203023 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Feb 13 20:12:37.213637 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:12:37.545673 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 20:12:37.556593 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 20:12:37.675350 systemd-tmpfiles[1308]: ACLs are not supported, ignoring. Feb 13 20:12:37.675383 systemd-tmpfiles[1308]: ACLs are not supported, ignoring. Feb 13 20:12:37.688500 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:12:37.729326 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 20:12:37.797292 kernel: loop1: detected capacity change from 0 to 28272 Feb 13 20:12:38.237104 kernel: loop2: detected capacity change from 0 to 140992 Feb 13 20:12:38.566765 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 20:12:38.576551 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:12:38.602466 systemd-udevd[1315]: Using default interface naming scheme 'v255'. Feb 13 20:12:38.783313 kernel: loop3: detected capacity change from 0 to 218376 Feb 13 20:12:38.823281 kernel: loop4: detected capacity change from 0 to 138184 Feb 13 20:12:38.845372 kernel: loop5: detected capacity change from 0 to 28272 Feb 13 20:12:38.839874 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:12:38.853749 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 20:12:38.871292 kernel: loop6: detected capacity change from 0 to 140992 Feb 13 20:12:38.891342 kernel: loop7: detected capacity change from 0 to 218376 Feb 13 20:12:38.904191 (sd-merge)[1318]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Feb 13 20:12:38.904881 (sd-merge)[1318]: Merged extensions into '/usr'. Feb 13 20:12:38.935834 systemd[1]: Reloading requested from client PID 1292 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 20:12:38.935853 systemd[1]: Reloading... Feb 13 20:12:39.038339 zram_generator::config[1363]: No configuration found. Feb 13 20:12:39.148288 kernel: mousedev: PS/2 mouse device common for all mice Feb 13 20:12:39.153720 kernel: hv_vmbus: registering driver hv_balloon Feb 13 20:12:39.163418 kernel: hv_vmbus: registering driver hyperv_fb Feb 13 20:12:39.163515 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Feb 13 20:12:39.188313 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Feb 13 20:12:39.195380 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Feb 13 20:12:39.203130 kernel: Console: switching to colour dummy device 80x25 Feb 13 20:12:39.207313 kernel: Console: switching to colour frame buffer device 128x48 Feb 13 20:12:39.485109 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1323) Feb 13 20:12:39.543336 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:12:39.626285 kernel: kvm_intel: Using Hyper-V Enlightened VMCS Feb 13 20:12:39.714792 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Feb 13 20:12:39.715588 systemd[1]: Reloading finished in 779 ms. Feb 13 20:12:39.745223 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 20:12:39.775583 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Feb 13 20:12:39.792118 systemd[1]: Starting ensure-sysext.service... Feb 13 20:12:39.798357 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 20:12:39.803662 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 20:12:39.810480 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 20:12:39.824940 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:12:39.835657 systemd[1]: Reloading requested from client PID 1501 ('systemctl') (unit ensure-sysext.service)... Feb 13 20:12:39.835847 systemd[1]: Reloading... Feb 13 20:12:39.880285 systemd-tmpfiles[1503]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 20:12:39.880789 systemd-tmpfiles[1503]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 20:12:39.884224 systemd-tmpfiles[1503]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 20:12:39.887744 systemd-tmpfiles[1503]: ACLs are not supported, ignoring. Feb 13 20:12:39.887848 systemd-tmpfiles[1503]: ACLs are not supported, ignoring. Feb 13 20:12:39.919009 systemd-tmpfiles[1503]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 20:12:39.919025 systemd-tmpfiles[1503]: Skipping /boot Feb 13 20:12:39.960296 zram_generator::config[1543]: No configuration found. Feb 13 20:12:39.960736 systemd-tmpfiles[1503]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 20:12:39.960877 systemd-tmpfiles[1503]: Skipping /boot Feb 13 20:12:40.130483 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:12:40.145209 systemd-networkd[1325]: lo: Link UP Feb 13 20:12:40.145224 systemd-networkd[1325]: lo: Gained carrier Feb 13 20:12:40.149126 systemd-networkd[1325]: Enumeration completed Feb 13 20:12:40.149831 systemd-networkd[1325]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:12:40.149836 systemd-networkd[1325]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 20:12:40.206310 kernel: mlx5_core ac61:00:02.0 enP44129s1: Link up Feb 13 20:12:40.236007 systemd[1]: Reloading finished in 399 ms. Feb 13 20:12:40.237275 kernel: hv_netvsc 000d3ad6-6c1a-000d-3ad6-6c1a000d3ad6 eth0: Data path switched to VF: enP44129s1 Feb 13 20:12:40.237741 systemd-networkd[1325]: enP44129s1: Link UP Feb 13 20:12:40.237830 systemd-networkd[1325]: eth0: Link UP Feb 13 20:12:40.237834 systemd-networkd[1325]: eth0: Gained carrier Feb 13 20:12:40.237850 systemd-networkd[1325]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:12:40.241672 systemd-networkd[1325]: enP44129s1: Gained carrier Feb 13 20:12:40.252794 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 20:12:40.257226 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 20:12:40.263337 systemd-networkd[1325]: eth0: DHCPv4 address 10.200.4.48/24, gateway 10.200.4.1 acquired from 168.63.129.16 Feb 13 20:12:40.264853 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 20:12:40.271784 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 20:12:40.275777 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:12:40.279469 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:12:40.293542 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 20:12:40.341592 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 20:12:40.347299 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 20:12:40.356730 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 20:12:40.365758 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 20:12:40.374445 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 20:12:40.386000 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 20:12:40.395666 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:40.396012 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:12:40.401163 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 20:12:40.414718 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 20:12:40.419719 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 20:12:40.423848 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:12:40.424099 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:40.426147 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 20:12:40.427374 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 20:12:40.434076 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 20:12:40.434922 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 20:12:40.446197 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 20:12:40.446713 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 20:12:40.455114 lvm[1612]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 20:12:40.455948 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 20:12:40.471487 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:40.471768 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:12:40.483553 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 20:12:40.493650 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 20:12:40.513927 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 20:12:40.518050 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:12:40.518281 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:40.526611 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 20:12:40.534123 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 20:12:40.534426 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 20:12:40.538480 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 20:12:40.542318 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 20:12:40.542500 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 20:12:40.546360 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 20:12:40.546544 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 20:12:40.554632 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:12:40.556281 systemd-resolved[1615]: Positive Trust Anchors: Feb 13 20:12:40.556299 systemd-resolved[1615]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 20:12:40.556336 systemd-resolved[1615]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 20:12:40.557892 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:40.558153 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:12:40.563074 systemd-resolved[1615]: Using system hostname 'ci-4152.2.1-a-2754074fca'. Feb 13 20:12:40.565417 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 20:12:40.572390 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 20:12:40.575019 lvm[1647]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 20:12:40.576853 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:12:40.576911 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 20:12:40.576973 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 20:12:40.577022 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 20:12:40.580147 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:12:40.580709 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 20:12:40.584218 systemd[1]: Finished ensure-sysext.service. Feb 13 20:12:40.586796 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 20:12:40.588017 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 20:12:40.595756 systemd[1]: Reached target network.target - Network. Feb 13 20:12:40.598472 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:12:40.606715 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 20:12:40.636772 augenrules[1656]: No rules Feb 13 20:12:40.639859 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 20:12:40.640104 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 20:12:40.902541 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 20:12:40.906623 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 20:12:41.472419 systemd-networkd[1325]: enP44129s1: Gained IPv6LL Feb 13 20:12:41.728466 systemd-networkd[1325]: eth0: Gained IPv6LL Feb 13 20:12:41.731874 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 20:12:41.740000 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 20:12:43.954115 ldconfig[1287]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 20:12:43.963975 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 20:12:43.972456 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 20:12:43.984616 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 20:12:43.988147 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 20:12:43.991027 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 20:12:43.993955 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 20:12:43.997208 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 20:12:43.999794 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 20:12:44.002844 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 20:12:44.006019 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 20:12:44.006072 systemd[1]: Reached target paths.target - Path Units. Feb 13 20:12:44.008358 systemd[1]: Reached target timers.target - Timer Units. Feb 13 20:12:44.011759 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 20:12:44.015771 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 20:12:44.026300 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 20:12:44.029568 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 20:12:44.032193 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 20:12:44.034496 systemd[1]: Reached target basic.target - Basic System. Feb 13 20:12:44.036803 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 20:12:44.036842 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 20:12:44.066456 systemd[1]: Starting chronyd.service - NTP client/server... Feb 13 20:12:44.071420 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 20:12:44.077705 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Feb 13 20:12:44.088418 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 20:12:44.094376 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 20:12:44.100462 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 20:12:44.102931 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 20:12:44.102980 systemd[1]: hv_fcopy_daemon.service - Hyper-V FCOPY daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_fcopy). Feb 13 20:12:44.105696 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Feb 13 20:12:44.109428 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Feb 13 20:12:44.118344 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:12:44.122137 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 20:12:44.128336 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 20:12:44.142423 jq[1673]: false Feb 13 20:12:44.140120 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 20:12:44.147727 (chronyd)[1669]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Feb 13 20:12:44.150933 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 20:12:44.155494 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 20:12:44.163443 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 20:12:44.167146 KVP[1678]: KVP starting; pid is:1678 Feb 13 20:12:44.170248 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 20:12:44.170892 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 20:12:44.176407 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 20:12:44.186373 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 20:12:44.193530 jq[1690]: true Feb 13 20:12:44.195701 chronyd[1692]: chronyd version 4.6 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) Feb 13 20:12:44.203342 kernel: hv_utils: KVP IC version 4.0 Feb 13 20:12:44.203365 KVP[1678]: KVP LIC Version: 3.1 Feb 13 20:12:44.204686 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 20:12:44.204927 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 20:12:44.216324 chronyd[1692]: Timezone right/UTC failed leap second check, ignoring Feb 13 20:12:44.217415 chronyd[1692]: Loaded seccomp filter (level 2) Feb 13 20:12:44.228404 systemd[1]: Started chronyd.service - NTP client/server. Feb 13 20:12:44.241721 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 20:12:44.241965 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 20:12:44.244897 extend-filesystems[1674]: Found loop4 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found loop5 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found loop6 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found loop7 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda1 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda2 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda3 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found usr Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda4 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda6 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda7 Feb 13 20:12:44.244897 extend-filesystems[1674]: Found sda9 Feb 13 20:12:44.244897 extend-filesystems[1674]: Checking size of /dev/sda9 Feb 13 20:12:44.309829 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 20:12:44.314603 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 20:12:44.315445 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 20:12:44.320460 extend-filesystems[1674]: Old size kept for /dev/sda9 Feb 13 20:12:44.320460 extend-filesystems[1674]: Found sr0 Feb 13 20:12:44.324802 (ntainerd)[1714]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 20:12:44.337023 jq[1699]: true Feb 13 20:12:44.334162 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 20:12:44.335461 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 20:12:44.339592 update_engine[1686]: I20250213 20:12:44.339174 1686 main.cc:92] Flatcar Update Engine starting Feb 13 20:12:44.398282 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1750) Feb 13 20:12:44.400188 tar[1698]: linux-amd64/LICENSE Feb 13 20:12:44.400535 tar[1698]: linux-amd64/helm Feb 13 20:12:44.423205 systemd-logind[1685]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Feb 13 20:12:44.423917 systemd-logind[1685]: New seat seat0. Feb 13 20:12:44.426013 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 20:12:44.441229 bash[1752]: Updated "/home/core/.ssh/authorized_keys" Feb 13 20:12:44.443016 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 20:12:44.455077 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Feb 13 20:12:44.521738 dbus-daemon[1672]: [system] SELinux support is enabled Feb 13 20:12:44.522015 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 20:12:44.532109 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 20:12:44.532156 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 20:12:44.539437 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 20:12:44.539476 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 20:12:44.550037 dbus-daemon[1672]: [system] Successfully activated service 'org.freedesktop.systemd1' Feb 13 20:12:44.554800 systemd[1]: Started update-engine.service - Update Engine. Feb 13 20:12:44.557768 update_engine[1686]: I20250213 20:12:44.557705 1686 update_check_scheduler.cc:74] Next update check in 5m11s Feb 13 20:12:44.568444 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 20:12:44.602670 sshd_keygen[1722]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 20:12:44.638572 coreos-metadata[1671]: Feb 13 20:12:44.638 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Feb 13 20:12:44.645754 coreos-metadata[1671]: Feb 13 20:12:44.644 INFO Fetch successful Feb 13 20:12:44.645754 coreos-metadata[1671]: Feb 13 20:12:44.644 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Feb 13 20:12:44.650484 coreos-metadata[1671]: Feb 13 20:12:44.650 INFO Fetch successful Feb 13 20:12:44.650484 coreos-metadata[1671]: Feb 13 20:12:44.650 INFO Fetching http://168.63.129.16/machine/d810c8ae-94a2-4f44-8734-eea33dad517c/739bcef5%2D0fc0%2D4b80%2D8c4b%2D9a6fe6372542.%5Fci%2D4152.2.1%2Da%2D2754074fca?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Feb 13 20:12:44.655080 coreos-metadata[1671]: Feb 13 20:12:44.654 INFO Fetch successful Feb 13 20:12:44.655080 coreos-metadata[1671]: Feb 13 20:12:44.654 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Feb 13 20:12:44.664547 coreos-metadata[1671]: Feb 13 20:12:44.664 INFO Fetch successful Feb 13 20:12:44.758713 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 20:12:44.789803 locksmithd[1789]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 20:12:44.790675 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 20:12:44.804351 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Feb 13 20:12:44.807903 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Feb 13 20:12:44.813968 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 20:12:44.840951 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 20:12:44.841194 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 20:12:44.862736 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 20:12:44.870439 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Feb 13 20:12:44.891275 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 20:12:44.902557 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 20:12:44.912583 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Feb 13 20:12:44.916483 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 20:12:45.317407 tar[1698]: linux-amd64/README.md Feb 13 20:12:45.332415 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 20:12:45.890621 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:12:45.911763 (kubelet)[1858]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:12:45.946838 containerd[1714]: time="2025-02-13T20:12:45.946512700Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Feb 13 20:12:45.997637 containerd[1714]: time="2025-02-13T20:12:45.997572100Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:12:45.999772 containerd[1714]: time="2025-02-13T20:12:45.999715100Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:12:45.999772 containerd[1714]: time="2025-02-13T20:12:45.999759600Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 20:12:45.999905 containerd[1714]: time="2025-02-13T20:12:45.999782500Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 20:12:46.000009 containerd[1714]: time="2025-02-13T20:12:45.999971000Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 20:12:46.000009 containerd[1714]: time="2025-02-13T20:12:45.999996300Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000117 containerd[1714]: time="2025-02-13T20:12:46.000085800Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000117 containerd[1714]: time="2025-02-13T20:12:46.000107900Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000919 containerd[1714]: time="2025-02-13T20:12:46.000357200Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000919 containerd[1714]: time="2025-02-13T20:12:46.000386300Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000919 containerd[1714]: time="2025-02-13T20:12:46.000407100Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000919 containerd[1714]: time="2025-02-13T20:12:46.000421600Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000919 containerd[1714]: time="2025-02-13T20:12:46.000524800Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:12:46.000919 containerd[1714]: time="2025-02-13T20:12:46.000873700Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:12:46.001175 containerd[1714]: time="2025-02-13T20:12:46.001042700Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:12:46.001175 containerd[1714]: time="2025-02-13T20:12:46.001064600Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 20:12:46.001175 containerd[1714]: time="2025-02-13T20:12:46.001170500Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 20:12:46.001527 containerd[1714]: time="2025-02-13T20:12:46.001228300Z" level=info msg="metadata content store policy set" policy=shared Feb 13 20:12:46.018968 containerd[1714]: time="2025-02-13T20:12:46.018916100Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019172200Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019259500Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019285700Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019349000Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019603800Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019810600Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019906000Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019919300Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019932400Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019944900Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019958300Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019973000Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019986600Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.021314 containerd[1714]: time="2025-02-13T20:12:46.019999500Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020012300Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020024800Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020036200Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020053900Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020067200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020086400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020100300Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020111100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020122700Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020132400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020144500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020155900Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020186000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.022806 containerd[1714]: time="2025-02-13T20:12:46.020200200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020210100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020219800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020232500Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020273000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020289600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020299800Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020342500Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020359300Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020369100Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020381000Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020389400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020400600Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020409500Z" level=info msg="NRI interface is disabled by configuration." Feb 13 20:12:46.023507 containerd[1714]: time="2025-02-13T20:12:46.020419500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.020681800Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.020724800Z" level=info msg="Connect containerd service" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.020757600Z" level=info msg="using legacy CRI server" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.020765300Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.020870800Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.021756800Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.022237900Z" level=info msg="Start subscribing containerd event" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.022322900Z" level=info msg="Start recovering state" Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.022240000Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 20:12:46.024119 containerd[1714]: time="2025-02-13T20:12:46.022437300Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 20:12:46.024866 containerd[1714]: time="2025-02-13T20:12:46.024845400Z" level=info msg="Start event monitor" Feb 13 20:12:46.025001 containerd[1714]: time="2025-02-13T20:12:46.024932500Z" level=info msg="Start snapshots syncer" Feb 13 20:12:46.025001 containerd[1714]: time="2025-02-13T20:12:46.024949800Z" level=info msg="Start cni network conf syncer for default" Feb 13 20:12:46.025001 containerd[1714]: time="2025-02-13T20:12:46.024961300Z" level=info msg="Start streaming server" Feb 13 20:12:46.028807 containerd[1714]: time="2025-02-13T20:12:46.025324900Z" level=info msg="containerd successfully booted in 0.079964s" Feb 13 20:12:46.025497 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 20:12:46.030387 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 20:12:46.036853 systemd[1]: Startup finished in 759ms (firmware) + 31.260s (loader) + 1.042s (kernel) + 12.281s (initrd) + 14.513s (userspace) = 59.857s. Feb 13 20:12:46.517665 login[1844]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Feb 13 20:12:46.520829 login[1845]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Feb 13 20:12:46.532407 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 20:12:46.541655 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 20:12:46.547683 systemd-logind[1685]: New session 2 of user core. Feb 13 20:12:46.557891 systemd-logind[1685]: New session 1 of user core. Feb 13 20:12:46.563579 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 20:12:46.568644 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 20:12:46.599460 (systemd)[1873]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 20:12:46.661659 kubelet[1858]: E0213 20:12:46.661611 1858 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:12:46.664340 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:12:46.664532 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:12:46.917639 systemd[1873]: Queued start job for default target default.target. Feb 13 20:12:46.924622 systemd[1873]: Created slice app.slice - User Application Slice. Feb 13 20:12:46.924661 systemd[1873]: Reached target paths.target - Paths. Feb 13 20:12:46.924681 systemd[1873]: Reached target timers.target - Timers. Feb 13 20:12:46.925993 systemd[1873]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 20:12:46.938521 systemd[1873]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 20:12:46.938800 systemd[1873]: Reached target sockets.target - Sockets. Feb 13 20:12:46.938915 systemd[1873]: Reached target basic.target - Basic System. Feb 13 20:12:46.939029 systemd[1873]: Reached target default.target - Main User Target. Feb 13 20:12:46.939222 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 20:12:46.939283 systemd[1873]: Startup finished in 329ms. Feb 13 20:12:46.949408 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 20:12:46.950386 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 20:12:47.245390 waagent[1841]: 2025-02-13T20:12:47.245150Z INFO Daemon Daemon Azure Linux Agent Version: 2.9.1.1 Feb 13 20:12:47.254130 waagent[1841]: 2025-02-13T20:12:47.247563Z INFO Daemon Daemon OS: flatcar 4152.2.1 Feb 13 20:12:47.254130 waagent[1841]: 2025-02-13T20:12:47.251390Z INFO Daemon Daemon Python: 3.11.10 Feb 13 20:12:47.254348 waagent[1841]: 2025-02-13T20:12:47.254165Z INFO Daemon Daemon Run daemon Feb 13 20:12:47.275634 waagent[1841]: 2025-02-13T20:12:47.256750Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4152.2.1' Feb 13 20:12:47.275634 waagent[1841]: 2025-02-13T20:12:47.258571Z INFO Daemon Daemon Using waagent for provisioning Feb 13 20:12:47.275634 waagent[1841]: 2025-02-13T20:12:47.261418Z INFO Daemon Daemon Activate resource disk Feb 13 20:12:47.275634 waagent[1841]: 2025-02-13T20:12:47.263695Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Feb 13 20:12:47.275934 waagent[1841]: 2025-02-13T20:12:47.275845Z INFO Daemon Daemon Found device: None Feb 13 20:12:47.278628 waagent[1841]: 2025-02-13T20:12:47.278532Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Feb 13 20:12:47.288300 waagent[1841]: 2025-02-13T20:12:47.282119Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Feb 13 20:12:47.288300 waagent[1841]: 2025-02-13T20:12:47.285629Z INFO Daemon Daemon Clean protocol and wireserver endpoint Feb 13 20:12:47.288300 waagent[1841]: 2025-02-13T20:12:47.287002Z INFO Daemon Daemon Running default provisioning handler Feb 13 20:12:47.297812 waagent[1841]: 2025-02-13T20:12:47.297710Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Feb 13 20:12:47.312122 waagent[1841]: 2025-02-13T20:12:47.299793Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Feb 13 20:12:47.312122 waagent[1841]: 2025-02-13T20:12:47.300421Z INFO Daemon Daemon cloud-init is enabled: False Feb 13 20:12:47.312122 waagent[1841]: 2025-02-13T20:12:47.300922Z INFO Daemon Daemon Copying ovf-env.xml Feb 13 20:12:47.390291 waagent[1841]: 2025-02-13T20:12:47.386574Z INFO Daemon Daemon Successfully mounted dvd Feb 13 20:12:47.424806 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Feb 13 20:12:47.427451 waagent[1841]: 2025-02-13T20:12:47.427364Z INFO Daemon Daemon Detect protocol endpoint Feb 13 20:12:47.430193 waagent[1841]: 2025-02-13T20:12:47.430124Z INFO Daemon Daemon Clean protocol and wireserver endpoint Feb 13 20:12:47.433742 waagent[1841]: 2025-02-13T20:12:47.433680Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Feb 13 20:12:47.437191 waagent[1841]: 2025-02-13T20:12:47.437132Z INFO Daemon Daemon Test for route to 168.63.129.16 Feb 13 20:12:47.440189 waagent[1841]: 2025-02-13T20:12:47.440135Z INFO Daemon Daemon Route to 168.63.129.16 exists Feb 13 20:12:47.443385 waagent[1841]: 2025-02-13T20:12:47.442390Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Feb 13 20:12:47.503553 waagent[1841]: 2025-02-13T20:12:47.503386Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Feb 13 20:12:47.514332 waagent[1841]: 2025-02-13T20:12:47.506242Z INFO Daemon Daemon Wire protocol version:2012-11-30 Feb 13 20:12:47.514332 waagent[1841]: 2025-02-13T20:12:47.508433Z INFO Daemon Daemon Server preferred version:2015-04-05 Feb 13 20:12:47.664963 waagent[1841]: 2025-02-13T20:12:47.664842Z INFO Daemon Daemon Initializing goal state during protocol detection Feb 13 20:12:47.671129 waagent[1841]: 2025-02-13T20:12:47.666471Z INFO Daemon Daemon Forcing an update of the goal state. Feb 13 20:12:47.673703 waagent[1841]: 2025-02-13T20:12:47.673646Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Feb 13 20:12:47.705088 waagent[1841]: 2025-02-13T20:12:47.705008Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.162 Feb 13 20:12:47.734543 waagent[1841]: 2025-02-13T20:12:47.707111Z INFO Daemon Feb 13 20:12:47.734543 waagent[1841]: 2025-02-13T20:12:47.708929Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: 6324d238-5172-4dac-8576-aac62548d5f8 eTag: 7863965257946900533 source: Fabric] Feb 13 20:12:47.734543 waagent[1841]: 2025-02-13T20:12:47.710475Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Feb 13 20:12:47.734543 waagent[1841]: 2025-02-13T20:12:47.711508Z INFO Daemon Feb 13 20:12:47.734543 waagent[1841]: 2025-02-13T20:12:47.712277Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Feb 13 20:12:47.737743 waagent[1841]: 2025-02-13T20:12:47.737685Z INFO Daemon Daemon Downloading artifacts profile blob Feb 13 20:12:47.830607 waagent[1841]: 2025-02-13T20:12:47.830453Z INFO Daemon Downloaded certificate {'thumbprint': '9A2EC5E2858C44923678467949F4ABCF454FEABA', 'hasPrivateKey': True} Feb 13 20:12:47.835793 waagent[1841]: 2025-02-13T20:12:47.835725Z INFO Daemon Fetch goal state completed Feb 13 20:12:47.844073 waagent[1841]: 2025-02-13T20:12:47.844011Z INFO Daemon Daemon Starting provisioning Feb 13 20:12:47.851010 waagent[1841]: 2025-02-13T20:12:47.845517Z INFO Daemon Daemon Handle ovf-env.xml. Feb 13 20:12:47.851010 waagent[1841]: 2025-02-13T20:12:47.846414Z INFO Daemon Daemon Set hostname [ci-4152.2.1-a-2754074fca] Feb 13 20:12:47.852347 waagent[1841]: 2025-02-13T20:12:47.852294Z INFO Daemon Daemon Publish hostname [ci-4152.2.1-a-2754074fca] Feb 13 20:12:47.859674 waagent[1841]: 2025-02-13T20:12:47.853874Z INFO Daemon Daemon Examine /proc/net/route for primary interface Feb 13 20:12:47.859674 waagent[1841]: 2025-02-13T20:12:47.854361Z INFO Daemon Daemon Primary interface is [eth0] Feb 13 20:12:47.881558 systemd-networkd[1325]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:12:47.881569 systemd-networkd[1325]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 20:12:47.881624 systemd-networkd[1325]: eth0: DHCP lease lost Feb 13 20:12:47.883163 waagent[1841]: 2025-02-13T20:12:47.883056Z INFO Daemon Daemon Create user account if not exists Feb 13 20:12:47.901363 waagent[1841]: 2025-02-13T20:12:47.884493Z INFO Daemon Daemon User core already exists, skip useradd Feb 13 20:12:47.901363 waagent[1841]: 2025-02-13T20:12:47.885556Z INFO Daemon Daemon Configure sudoer Feb 13 20:12:47.901363 waagent[1841]: 2025-02-13T20:12:47.886361Z INFO Daemon Daemon Configure sshd Feb 13 20:12:47.901363 waagent[1841]: 2025-02-13T20:12:47.887183Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Feb 13 20:12:47.901363 waagent[1841]: 2025-02-13T20:12:47.887576Z INFO Daemon Daemon Deploy ssh public key. Feb 13 20:12:47.901656 systemd-networkd[1325]: eth0: DHCPv6 lease lost Feb 13 20:12:47.940364 systemd-networkd[1325]: eth0: DHCPv4 address 10.200.4.48/24, gateway 10.200.4.1 acquired from 168.63.129.16 Feb 13 20:12:48.990502 waagent[1841]: 2025-02-13T20:12:48.990433Z INFO Daemon Daemon Provisioning complete Feb 13 20:12:49.002456 waagent[1841]: 2025-02-13T20:12:49.002391Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Feb 13 20:12:49.009426 waagent[1841]: 2025-02-13T20:12:49.003565Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Feb 13 20:12:49.009426 waagent[1841]: 2025-02-13T20:12:49.004371Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.9.1.1 is the most current agent Feb 13 20:12:49.135904 waagent[1926]: 2025-02-13T20:12:49.135763Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.9.1.1) Feb 13 20:12:49.136435 waagent[1926]: 2025-02-13T20:12:49.135966Z INFO ExtHandler ExtHandler OS: flatcar 4152.2.1 Feb 13 20:12:49.136435 waagent[1926]: 2025-02-13T20:12:49.136055Z INFO ExtHandler ExtHandler Python: 3.11.10 Feb 13 20:12:49.182496 waagent[1926]: 2025-02-13T20:12:49.182383Z INFO ExtHandler ExtHandler Distro: flatcar-4152.2.1; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.10; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Feb 13 20:12:49.182749 waagent[1926]: 2025-02-13T20:12:49.182694Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Feb 13 20:12:49.182840 waagent[1926]: 2025-02-13T20:12:49.182804Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Feb 13 20:12:49.191338 waagent[1926]: 2025-02-13T20:12:49.191221Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Feb 13 20:12:49.197743 waagent[1926]: 2025-02-13T20:12:49.197666Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.162 Feb 13 20:12:49.198339 waagent[1926]: 2025-02-13T20:12:49.198275Z INFO ExtHandler Feb 13 20:12:49.198445 waagent[1926]: 2025-02-13T20:12:49.198391Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 16a26682-611f-48c4-871b-b755e51cfad1 eTag: 7863965257946900533 source: Fabric] Feb 13 20:12:49.198780 waagent[1926]: 2025-02-13T20:12:49.198724Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Feb 13 20:12:49.199407 waagent[1926]: 2025-02-13T20:12:49.199348Z INFO ExtHandler Feb 13 20:12:49.199486 waagent[1926]: 2025-02-13T20:12:49.199435Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Feb 13 20:12:49.203442 waagent[1926]: 2025-02-13T20:12:49.203389Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Feb 13 20:12:49.294794 waagent[1926]: 2025-02-13T20:12:49.294623Z INFO ExtHandler Downloaded certificate {'thumbprint': '9A2EC5E2858C44923678467949F4ABCF454FEABA', 'hasPrivateKey': True} Feb 13 20:12:49.295377 waagent[1926]: 2025-02-13T20:12:49.295308Z INFO ExtHandler Fetch goal state completed Feb 13 20:12:49.310548 waagent[1926]: 2025-02-13T20:12:49.310458Z INFO ExtHandler ExtHandler WALinuxAgent-2.9.1.1 running as process 1926 Feb 13 20:12:49.310740 waagent[1926]: 2025-02-13T20:12:49.310688Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Feb 13 20:12:49.312483 waagent[1926]: 2025-02-13T20:12:49.312418Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '4152.2.1', '', 'Flatcar Container Linux by Kinvolk'] Feb 13 20:12:49.312857 waagent[1926]: 2025-02-13T20:12:49.312804Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Feb 13 20:12:49.433909 waagent[1926]: 2025-02-13T20:12:49.433845Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Feb 13 20:12:49.434189 waagent[1926]: 2025-02-13T20:12:49.434130Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Feb 13 20:12:49.441902 waagent[1926]: 2025-02-13T20:12:49.441853Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Feb 13 20:12:49.451554 systemd[1]: Reloading requested from client PID 1939 ('systemctl') (unit waagent.service)... Feb 13 20:12:49.451573 systemd[1]: Reloading... Feb 13 20:12:49.564294 zram_generator::config[1979]: No configuration found. Feb 13 20:12:49.686920 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:12:49.772434 systemd[1]: Reloading finished in 320 ms. Feb 13 20:12:49.799831 waagent[1926]: 2025-02-13T20:12:49.797307Z INFO ExtHandler ExtHandler Executing systemctl daemon-reload for setting up waagent-network-setup.service Feb 13 20:12:49.806645 systemd[1]: Reloading requested from client PID 2030 ('systemctl') (unit waagent.service)... Feb 13 20:12:49.806664 systemd[1]: Reloading... Feb 13 20:12:49.900310 zram_generator::config[2064]: No configuration found. Feb 13 20:12:50.033140 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:12:50.119988 systemd[1]: Reloading finished in 312 ms. Feb 13 20:12:50.152678 waagent[1926]: 2025-02-13T20:12:50.149568Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Feb 13 20:12:50.152678 waagent[1926]: 2025-02-13T20:12:50.149794Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Feb 13 20:12:51.026059 waagent[1926]: 2025-02-13T20:12:51.025948Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Feb 13 20:12:51.028970 waagent[1926]: 2025-02-13T20:12:51.028866Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [True], cgroups enabled [False], python supported: [True] Feb 13 20:12:51.030089 waagent[1926]: 2025-02-13T20:12:51.029999Z INFO ExtHandler ExtHandler Starting env monitor service. Feb 13 20:12:51.030312 waagent[1926]: 2025-02-13T20:12:51.030187Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Feb 13 20:12:51.031074 waagent[1926]: 2025-02-13T20:12:51.031002Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Feb 13 20:12:51.031242 waagent[1926]: 2025-02-13T20:12:51.031145Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Feb 13 20:12:51.031328 waagent[1926]: 2025-02-13T20:12:51.031248Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Feb 13 20:12:51.031477 waagent[1926]: 2025-02-13T20:12:51.031415Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Feb 13 20:12:51.031876 waagent[1926]: 2025-02-13T20:12:51.031819Z INFO EnvHandler ExtHandler Configure routes Feb 13 20:12:51.032109 waagent[1926]: 2025-02-13T20:12:51.031948Z INFO EnvHandler ExtHandler Gateway:None Feb 13 20:12:51.032235 waagent[1926]: 2025-02-13T20:12:51.032182Z INFO EnvHandler ExtHandler Routes:None Feb 13 20:12:51.032637 waagent[1926]: 2025-02-13T20:12:51.032580Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Feb 13 20:12:51.033643 waagent[1926]: 2025-02-13T20:12:51.033465Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Feb 13 20:12:51.033736 waagent[1926]: 2025-02-13T20:12:51.033644Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Feb 13 20:12:51.033736 waagent[1926]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Feb 13 20:12:51.033736 waagent[1926]: eth0 00000000 0104C80A 0003 0 0 1024 00000000 0 0 0 Feb 13 20:12:51.033736 waagent[1926]: eth0 0004C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Feb 13 20:12:51.033736 waagent[1926]: eth0 0104C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Feb 13 20:12:51.033736 waagent[1926]: eth0 10813FA8 0104C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Feb 13 20:12:51.033736 waagent[1926]: eth0 FEA9FEA9 0104C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Feb 13 20:12:51.033995 waagent[1926]: 2025-02-13T20:12:51.033766Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Feb 13 20:12:51.035321 waagent[1926]: 2025-02-13T20:12:51.035172Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Feb 13 20:12:51.035399 waagent[1926]: 2025-02-13T20:12:51.035061Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Feb 13 20:12:51.035875 waagent[1926]: 2025-02-13T20:12:51.035825Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Feb 13 20:12:51.046853 waagent[1926]: 2025-02-13T20:12:51.046779Z INFO ExtHandler ExtHandler Feb 13 20:12:51.046952 waagent[1926]: 2025-02-13T20:12:51.046904Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: 0070662b-3536-4f42-a818-e1aa56799763 correlation 73b966f0-ef12-4e7b-8cdf-c218c0c0baeb created: 2025-02-13T20:11:36.870377Z] Feb 13 20:12:51.049282 waagent[1926]: 2025-02-13T20:12:51.047522Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Feb 13 20:12:51.049282 waagent[1926]: 2025-02-13T20:12:51.049159Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 2 ms] Feb 13 20:12:51.089054 waagent[1926]: 2025-02-13T20:12:51.088981Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.9.1.1 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: AEF1FC8B-1E1C-41AB-AC5B-23AB486194B2;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 0] Feb 13 20:12:51.126013 waagent[1926]: 2025-02-13T20:12:51.125915Z INFO MonitorHandler ExtHandler Network interfaces: Feb 13 20:12:51.126013 waagent[1926]: Executing ['ip', '-a', '-o', 'link']: Feb 13 20:12:51.126013 waagent[1926]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Feb 13 20:12:51.126013 waagent[1926]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:d6:6c:1a brd ff:ff:ff:ff:ff:ff Feb 13 20:12:51.126013 waagent[1926]: 3: enP44129s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:d6:6c:1a brd ff:ff:ff:ff:ff:ff\ altname enP44129p0s2 Feb 13 20:12:51.126013 waagent[1926]: Executing ['ip', '-4', '-a', '-o', 'address']: Feb 13 20:12:51.126013 waagent[1926]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Feb 13 20:12:51.126013 waagent[1926]: 2: eth0 inet 10.200.4.48/24 metric 1024 brd 10.200.4.255 scope global eth0\ valid_lft forever preferred_lft forever Feb 13 20:12:51.126013 waagent[1926]: Executing ['ip', '-6', '-a', '-o', 'address']: Feb 13 20:12:51.126013 waagent[1926]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Feb 13 20:12:51.126013 waagent[1926]: 2: eth0 inet6 fe80::20d:3aff:fed6:6c1a/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Feb 13 20:12:51.126013 waagent[1926]: 3: enP44129s1 inet6 fe80::20d:3aff:fed6:6c1a/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Feb 13 20:12:51.156884 waagent[1926]: 2025-02-13T20:12:51.156797Z INFO EnvHandler ExtHandler Successfully added Azure fabric firewall rules. Current Firewall rules: Feb 13 20:12:51.156884 waagent[1926]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Feb 13 20:12:51.156884 waagent[1926]: pkts bytes target prot opt in out source destination Feb 13 20:12:51.156884 waagent[1926]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Feb 13 20:12:51.156884 waagent[1926]: pkts bytes target prot opt in out source destination Feb 13 20:12:51.156884 waagent[1926]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Feb 13 20:12:51.156884 waagent[1926]: pkts bytes target prot opt in out source destination Feb 13 20:12:51.156884 waagent[1926]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Feb 13 20:12:51.156884 waagent[1926]: 7 938 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Feb 13 20:12:51.156884 waagent[1926]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Feb 13 20:12:51.160764 waagent[1926]: 2025-02-13T20:12:51.160705Z INFO EnvHandler ExtHandler Current Firewall rules: Feb 13 20:12:51.160764 waagent[1926]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Feb 13 20:12:51.160764 waagent[1926]: pkts bytes target prot opt in out source destination Feb 13 20:12:51.160764 waagent[1926]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Feb 13 20:12:51.160764 waagent[1926]: pkts bytes target prot opt in out source destination Feb 13 20:12:51.160764 waagent[1926]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Feb 13 20:12:51.160764 waagent[1926]: pkts bytes target prot opt in out source destination Feb 13 20:12:51.160764 waagent[1926]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Feb 13 20:12:51.160764 waagent[1926]: 11 1353 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Feb 13 20:12:51.160764 waagent[1926]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Feb 13 20:12:51.161139 waagent[1926]: 2025-02-13T20:12:51.161056Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Feb 13 20:12:56.833973 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 20:12:56.843562 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:12:56.967925 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:12:56.982665 (kubelet)[2160]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:12:57.699786 kubelet[2160]: E0213 20:12:57.699721 2160 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:12:57.703522 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:12:57.703750 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:13:07.834209 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 13 20:13:07.846574 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:07.957033 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:07.967657 (kubelet)[2175]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:13:08.016560 chronyd[1692]: Selected source PHC0 Feb 13 20:13:08.545790 kubelet[2175]: E0213 20:13:08.545727 2175 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:13:08.548537 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:13:08.548754 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:13:14.230190 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 20:13:14.238598 systemd[1]: Started sshd@0-10.200.4.48:22-10.200.16.10:48676.service - OpenSSH per-connection server daemon (10.200.16.10:48676). Feb 13 20:13:15.257607 sshd[2183]: Accepted publickey for core from 10.200.16.10 port 48676 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:13:15.259359 sshd-session[2183]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:13:15.264032 systemd-logind[1685]: New session 3 of user core. Feb 13 20:13:15.274491 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 20:13:15.815764 systemd[1]: Started sshd@1-10.200.4.48:22-10.200.16.10:48678.service - OpenSSH per-connection server daemon (10.200.16.10:48678). Feb 13 20:13:16.399276 sshd[2188]: Accepted publickey for core from 10.200.16.10 port 48678 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:13:16.401025 sshd-session[2188]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:13:16.406303 systemd-logind[1685]: New session 4 of user core. Feb 13 20:13:16.411480 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 20:13:16.824030 sshd[2190]: Connection closed by 10.200.16.10 port 48678 Feb 13 20:13:16.826192 sshd-session[2188]: pam_unix(sshd:session): session closed for user core Feb 13 20:13:16.829535 systemd[1]: sshd@1-10.200.4.48:22-10.200.16.10:48678.service: Deactivated successfully. Feb 13 20:13:16.832000 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 20:13:16.833788 systemd-logind[1685]: Session 4 logged out. Waiting for processes to exit. Feb 13 20:13:16.834915 systemd-logind[1685]: Removed session 4. Feb 13 20:13:16.931424 systemd[1]: Started sshd@2-10.200.4.48:22-10.200.16.10:48694.service - OpenSSH per-connection server daemon (10.200.16.10:48694). Feb 13 20:13:17.520142 sshd[2195]: Accepted publickey for core from 10.200.16.10 port 48694 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:13:17.521875 sshd-session[2195]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:13:17.527383 systemd-logind[1685]: New session 5 of user core. Feb 13 20:13:17.537434 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 20:13:17.966501 sshd[2197]: Connection closed by 10.200.16.10 port 48694 Feb 13 20:13:17.970106 sshd-session[2195]: pam_unix(sshd:session): session closed for user core Feb 13 20:13:17.974201 systemd-logind[1685]: Session 5 logged out. Waiting for processes to exit. Feb 13 20:13:17.974962 systemd[1]: sshd@2-10.200.4.48:22-10.200.16.10:48694.service: Deactivated successfully. Feb 13 20:13:17.977183 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 20:13:17.978774 systemd-logind[1685]: Removed session 5. Feb 13 20:13:18.097100 systemd[1]: Started sshd@3-10.200.4.48:22-10.200.16.10:48696.service - OpenSSH per-connection server daemon (10.200.16.10:48696). Feb 13 20:13:18.578167 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Feb 13 20:13:18.583570 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:18.682587 sshd[2202]: Accepted publickey for core from 10.200.16.10 port 48696 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:13:18.685299 sshd-session[2202]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:13:18.706115 systemd-logind[1685]: New session 6 of user core. Feb 13 20:13:18.708154 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:18.713017 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 20:13:18.719587 (kubelet)[2211]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:13:19.135307 sshd[2213]: Connection closed by 10.200.16.10 port 48696 Feb 13 20:13:19.136672 sshd-session[2202]: pam_unix(sshd:session): session closed for user core Feb 13 20:13:19.140927 systemd[1]: sshd@3-10.200.4.48:22-10.200.16.10:48696.service: Deactivated successfully. Feb 13 20:13:19.143747 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 20:13:19.145609 systemd-logind[1685]: Session 6 logged out. Waiting for processes to exit. Feb 13 20:13:19.146694 systemd-logind[1685]: Removed session 6. Feb 13 20:13:19.244706 systemd[1]: Started sshd@4-10.200.4.48:22-10.200.16.10:42406.service - OpenSSH per-connection server daemon (10.200.16.10:42406). Feb 13 20:13:19.299451 kubelet[2211]: E0213 20:13:19.299382 2211 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:13:19.302245 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:13:19.302456 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:13:19.840034 sshd[2222]: Accepted publickey for core from 10.200.16.10 port 42406 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:13:19.841720 sshd-session[2222]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:13:19.847569 systemd-logind[1685]: New session 7 of user core. Feb 13 20:13:19.858459 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 20:13:20.635305 sudo[2228]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Feb 13 20:13:20.635804 sudo[2228]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:13:20.668080 sudo[2228]: pam_unix(sudo:session): session closed for user root Feb 13 20:13:20.772821 sshd[2227]: Connection closed by 10.200.16.10 port 42406 Feb 13 20:13:20.774037 sshd-session[2222]: pam_unix(sshd:session): session closed for user core Feb 13 20:13:20.777939 systemd[1]: sshd@4-10.200.4.48:22-10.200.16.10:42406.service: Deactivated successfully. Feb 13 20:13:20.780533 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 20:13:20.782388 systemd-logind[1685]: Session 7 logged out. Waiting for processes to exit. Feb 13 20:13:20.783465 systemd-logind[1685]: Removed session 7. Feb 13 20:13:20.889668 systemd[1]: Started sshd@5-10.200.4.48:22-10.200.16.10:42418.service - OpenSSH per-connection server daemon (10.200.16.10:42418). Feb 13 20:13:21.476477 sshd[2233]: Accepted publickey for core from 10.200.16.10 port 42418 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:13:21.478344 sshd-session[2233]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:13:21.482939 systemd-logind[1685]: New session 8 of user core. Feb 13 20:13:21.492446 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 20:13:21.800331 sudo[2237]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Feb 13 20:13:21.800778 sudo[2237]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:13:21.804974 sudo[2237]: pam_unix(sudo:session): session closed for user root Feb 13 20:13:21.809938 sudo[2236]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Feb 13 20:13:21.810288 sudo[2236]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:13:21.828683 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 20:13:21.853869 augenrules[2259]: No rules Feb 13 20:13:21.855224 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 20:13:21.855557 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 20:13:21.856725 sudo[2236]: pam_unix(sudo:session): session closed for user root Feb 13 20:13:21.985604 sshd[2235]: Connection closed by 10.200.16.10 port 42418 Feb 13 20:13:21.986761 sshd-session[2233]: pam_unix(sshd:session): session closed for user core Feb 13 20:13:21.990465 systemd[1]: sshd@5-10.200.4.48:22-10.200.16.10:42418.service: Deactivated successfully. Feb 13 20:13:21.994798 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 20:13:21.996709 systemd-logind[1685]: Session 8 logged out. Waiting for processes to exit. Feb 13 20:13:21.997941 systemd-logind[1685]: Removed session 8. Feb 13 20:13:22.089488 systemd[1]: Started sshd@6-10.200.4.48:22-10.200.16.10:42434.service - OpenSSH per-connection server daemon (10.200.16.10:42434). Feb 13 20:13:22.684706 sshd[2267]: Accepted publickey for core from 10.200.16.10 port 42434 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:13:22.686568 sshd-session[2267]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:13:22.692118 systemd-logind[1685]: New session 9 of user core. Feb 13 20:13:22.697458 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 20:13:23.012318 sudo[2270]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 20:13:23.012744 sudo[2270]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:13:24.414610 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 20:13:24.414693 (dockerd)[2287]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 20:13:25.936847 dockerd[2287]: time="2025-02-13T20:13:25.936772722Z" level=info msg="Starting up" Feb 13 20:13:26.469299 dockerd[2287]: time="2025-02-13T20:13:26.469234866Z" level=info msg="Loading containers: start." Feb 13 20:13:26.708321 kernel: Initializing XFRM netlink socket Feb 13 20:13:26.876112 systemd-networkd[1325]: docker0: Link UP Feb 13 20:13:26.929556 dockerd[2287]: time="2025-02-13T20:13:26.929501628Z" level=info msg="Loading containers: done." Feb 13 20:13:26.991087 dockerd[2287]: time="2025-02-13T20:13:26.991028047Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 20:13:26.991575 dockerd[2287]: time="2025-02-13T20:13:26.991173951Z" level=info msg="Docker daemon" commit=8b539b8df24032dabeaaa099cf1d0535ef0286a3 containerd-snapshotter=false storage-driver=overlay2 version=27.2.1 Feb 13 20:13:26.991575 dockerd[2287]: time="2025-02-13T20:13:26.991352355Z" level=info msg="Daemon has completed initialization" Feb 13 20:13:27.049892 dockerd[2287]: time="2025-02-13T20:13:27.049821299Z" level=info msg="API listen on /run/docker.sock" Feb 13 20:13:27.050158 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 20:13:27.283345 kernel: hv_balloon: Max. dynamic memory size: 8192 MB Feb 13 20:13:27.755465 containerd[1714]: time="2025-02-13T20:13:27.755425517Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\"" Feb 13 20:13:28.709460 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2398306114.mount: Deactivated successfully. Feb 13 20:13:29.333914 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Feb 13 20:13:29.340503 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:30.101935 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:30.115793 (kubelet)[2520]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:13:30.164295 kubelet[2520]: E0213 20:13:30.163286 2520 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:13:30.164817 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:13:30.164993 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:13:30.203434 update_engine[1686]: I20250213 20:13:30.203348 1686 update_attempter.cc:509] Updating boot flags... Feb 13 20:13:30.295563 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (2554) Feb 13 20:13:31.085111 containerd[1714]: time="2025-02-13T20:13:31.085053111Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:31.088970 containerd[1714]: time="2025-02-13T20:13:31.088906496Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.2: active requests=0, bytes read=28673939" Feb 13 20:13:31.092555 containerd[1714]: time="2025-02-13T20:13:31.092499275Z" level=info msg="ImageCreate event name:\"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:31.098688 containerd[1714]: time="2025-02-13T20:13:31.098637510Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:31.099665 containerd[1714]: time="2025-02-13T20:13:31.099633932Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.2\" with image id \"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\", size \"28670731\" in 3.344170314s" Feb 13 20:13:31.100354 containerd[1714]: time="2025-02-13T20:13:31.099782235Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\" returns image reference \"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\"" Feb 13 20:13:31.100705 containerd[1714]: time="2025-02-13T20:13:31.100582253Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\"" Feb 13 20:13:32.711441 containerd[1714]: time="2025-02-13T20:13:32.711371948Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:32.713565 containerd[1714]: time="2025-02-13T20:13:32.713400091Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.2: active requests=0, bytes read=24771792" Feb 13 20:13:32.716726 containerd[1714]: time="2025-02-13T20:13:32.716675060Z" level=info msg="ImageCreate event name:\"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:32.721815 containerd[1714]: time="2025-02-13T20:13:32.721757366Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:32.722897 containerd[1714]: time="2025-02-13T20:13:32.722736387Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.2\" with image id \"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\", size \"26259392\" in 1.622050632s" Feb 13 20:13:32.722897 containerd[1714]: time="2025-02-13T20:13:32.722776188Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\" returns image reference \"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\"" Feb 13 20:13:32.723732 containerd[1714]: time="2025-02-13T20:13:32.723702307Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\"" Feb 13 20:13:34.080737 containerd[1714]: time="2025-02-13T20:13:34.080675701Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:34.082672 containerd[1714]: time="2025-02-13T20:13:34.082477439Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.2: active requests=0, bytes read=19170284" Feb 13 20:13:34.087286 containerd[1714]: time="2025-02-13T20:13:34.085690506Z" level=info msg="ImageCreate event name:\"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:34.092027 containerd[1714]: time="2025-02-13T20:13:34.091986638Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:34.093818 containerd[1714]: time="2025-02-13T20:13:34.092993560Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.2\" with image id \"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\", size \"20657902\" in 1.369259051s" Feb 13 20:13:34.093818 containerd[1714]: time="2025-02-13T20:13:34.093033460Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\" returns image reference \"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\"" Feb 13 20:13:34.093977 containerd[1714]: time="2025-02-13T20:13:34.093948080Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\"" Feb 13 20:13:35.394079 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1071687345.mount: Deactivated successfully. Feb 13 20:13:35.971024 containerd[1714]: time="2025-02-13T20:13:35.970967094Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:35.973594 containerd[1714]: time="2025-02-13T20:13:35.973522647Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.2: active requests=0, bytes read=30908847" Feb 13 20:13:35.977438 containerd[1714]: time="2025-02-13T20:13:35.977352528Z" level=info msg="ImageCreate event name:\"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:35.983043 containerd[1714]: time="2025-02-13T20:13:35.982956345Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:35.986555 containerd[1714]: time="2025-02-13T20:13:35.985287594Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.2\" with image id \"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\", repo tag \"registry.k8s.io/kube-proxy:v1.32.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\", size \"30907858\" in 1.891288614s" Feb 13 20:13:35.986555 containerd[1714]: time="2025-02-13T20:13:35.985329695Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\" returns image reference \"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\"" Feb 13 20:13:35.988705 containerd[1714]: time="2025-02-13T20:13:35.988684666Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Feb 13 20:13:36.648528 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1296994164.mount: Deactivated successfully. Feb 13 20:13:37.969351 containerd[1714]: time="2025-02-13T20:13:37.969281254Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:37.971156 containerd[1714]: time="2025-02-13T20:13:37.971088092Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565249" Feb 13 20:13:37.973655 containerd[1714]: time="2025-02-13T20:13:37.973590345Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:37.979603 containerd[1714]: time="2025-02-13T20:13:37.979539670Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:37.980900 containerd[1714]: time="2025-02-13T20:13:37.980737095Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.991939327s" Feb 13 20:13:37.980900 containerd[1714]: time="2025-02-13T20:13:37.980776696Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Feb 13 20:13:37.981925 containerd[1714]: time="2025-02-13T20:13:37.981878219Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Feb 13 20:13:38.558608 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3880638844.mount: Deactivated successfully. Feb 13 20:13:38.579825 containerd[1714]: time="2025-02-13T20:13:38.579767173Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:38.582105 containerd[1714]: time="2025-02-13T20:13:38.582044021Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321146" Feb 13 20:13:38.586059 containerd[1714]: time="2025-02-13T20:13:38.586003404Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:38.595679 containerd[1714]: time="2025-02-13T20:13:38.595599806Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:38.596954 containerd[1714]: time="2025-02-13T20:13:38.596445024Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 614.3629ms" Feb 13 20:13:38.596954 containerd[1714]: time="2025-02-13T20:13:38.596487025Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Feb 13 20:13:38.597120 containerd[1714]: time="2025-02-13T20:13:38.597068037Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Feb 13 20:13:39.251921 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount408283957.mount: Deactivated successfully. Feb 13 20:13:40.334499 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Feb 13 20:13:40.342552 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:40.525368 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:40.536106 (kubelet)[2734]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:13:41.091443 kubelet[2734]: E0213 20:13:41.091384 2734 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:13:41.094824 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:13:41.095036 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:13:41.625773 containerd[1714]: time="2025-02-13T20:13:41.625717168Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:41.628681 containerd[1714]: time="2025-02-13T20:13:41.628603831Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551328" Feb 13 20:13:41.633717 containerd[1714]: time="2025-02-13T20:13:41.633650642Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:41.639600 containerd[1714]: time="2025-02-13T20:13:41.639540671Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:13:41.640960 containerd[1714]: time="2025-02-13T20:13:41.640786398Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 3.043688661s" Feb 13 20:13:41.640960 containerd[1714]: time="2025-02-13T20:13:41.640825899Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Feb 13 20:13:44.735899 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:44.741575 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:44.774091 systemd[1]: Reloading requested from client PID 2772 ('systemctl') (unit session-9.scope)... Feb 13 20:13:44.774323 systemd[1]: Reloading... Feb 13 20:13:44.914373 zram_generator::config[2812]: No configuration found. Feb 13 20:13:45.040465 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:13:45.124513 systemd[1]: Reloading finished in 349 ms. Feb 13 20:13:45.185582 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Feb 13 20:13:45.185890 systemd[1]: kubelet.service: Failed with result 'signal'. Feb 13 20:13:45.186219 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:45.190626 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:45.514694 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:45.527091 (kubelet)[2883]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 20:13:46.195859 kubelet[2883]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:13:46.195859 kubelet[2883]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 20:13:46.195859 kubelet[2883]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:13:46.196394 kubelet[2883]: I0213 20:13:46.195875 2883 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 20:13:46.529620 kubelet[2883]: I0213 20:13:46.529486 2883 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 20:13:46.529620 kubelet[2883]: I0213 20:13:46.529519 2883 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 20:13:46.530152 kubelet[2883]: I0213 20:13:46.530118 2883 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 20:13:46.553062 kubelet[2883]: E0213 20:13:46.553011 2883 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.200.4.48:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:46.554432 kubelet[2883]: I0213 20:13:46.554286 2883 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 20:13:46.567155 kubelet[2883]: E0213 20:13:46.567121 2883 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 20:13:46.567290 kubelet[2883]: I0213 20:13:46.567165 2883 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 20:13:46.570827 kubelet[2883]: I0213 20:13:46.570803 2883 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 20:13:46.571087 kubelet[2883]: I0213 20:13:46.571048 2883 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 20:13:46.571293 kubelet[2883]: I0213 20:13:46.571082 2883 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4152.2.1-a-2754074fca","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 20:13:46.571459 kubelet[2883]: I0213 20:13:46.571298 2883 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 20:13:46.571459 kubelet[2883]: I0213 20:13:46.571313 2883 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 20:13:46.572261 kubelet[2883]: I0213 20:13:46.572236 2883 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:13:46.575266 kubelet[2883]: I0213 20:13:46.575239 2883 kubelet.go:446] "Attempting to sync node with API server" Feb 13 20:13:46.575351 kubelet[2883]: I0213 20:13:46.575274 2883 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 20:13:46.575351 kubelet[2883]: I0213 20:13:46.575301 2883 kubelet.go:352] "Adding apiserver pod source" Feb 13 20:13:46.575351 kubelet[2883]: I0213 20:13:46.575315 2883 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 20:13:46.580877 kubelet[2883]: W0213 20:13:46.580769 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.4.48:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:46.581272 kubelet[2883]: E0213 20:13:46.580848 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.4.48:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:46.581272 kubelet[2883]: I0213 20:13:46.581062 2883 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 20:13:46.581778 kubelet[2883]: I0213 20:13:46.581761 2883 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 20:13:46.582861 kubelet[2883]: W0213 20:13:46.581898 2883 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 20:13:46.584118 kubelet[2883]: I0213 20:13:46.584091 2883 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 20:13:46.584206 kubelet[2883]: I0213 20:13:46.584130 2883 server.go:1287] "Started kubelet" Feb 13 20:13:46.585811 kubelet[2883]: W0213 20:13:46.585280 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.4.48:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4152.2.1-a-2754074fca&limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:46.585811 kubelet[2883]: E0213 20:13:46.585343 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.4.48:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4152.2.1-a-2754074fca&limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:46.585811 kubelet[2883]: I0213 20:13:46.585438 2883 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 20:13:46.588117 kubelet[2883]: I0213 20:13:46.587736 2883 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 20:13:46.589755 kubelet[2883]: I0213 20:13:46.589243 2883 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 20:13:46.589755 kubelet[2883]: I0213 20:13:46.589585 2883 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 20:13:46.591488 kubelet[2883]: E0213 20:13:46.589967 2883 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.4.48:6443/api/v1/namespaces/default/events\": dial tcp 10.200.4.48:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4152.2.1-a-2754074fca.1823ddaf6b4910c6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4152.2.1-a-2754074fca,UID:ci-4152.2.1-a-2754074fca,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4152.2.1-a-2754074fca,},FirstTimestamp:2025-02-13 20:13:46.584109254 +0000 UTC m=+1.052868879,LastTimestamp:2025-02-13 20:13:46.584109254 +0000 UTC m=+1.052868879,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4152.2.1-a-2754074fca,}" Feb 13 20:13:46.592032 kubelet[2883]: I0213 20:13:46.592007 2883 server.go:490] "Adding debug handlers to kubelet server" Feb 13 20:13:46.593644 kubelet[2883]: I0213 20:13:46.593615 2883 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 20:13:46.595392 kubelet[2883]: I0213 20:13:46.595376 2883 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 20:13:46.595724 kubelet[2883]: E0213 20:13:46.595703 2883 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4152.2.1-a-2754074fca\" not found" Feb 13 20:13:46.597092 kubelet[2883]: E0213 20:13:46.597050 2883 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.48:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4152.2.1-a-2754074fca?timeout=10s\": dial tcp 10.200.4.48:6443: connect: connection refused" interval="200ms" Feb 13 20:13:46.597427 kubelet[2883]: I0213 20:13:46.597405 2883 factory.go:221] Registration of the systemd container factory successfully Feb 13 20:13:46.597499 kubelet[2883]: I0213 20:13:46.597487 2883 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 20:13:46.598438 kubelet[2883]: I0213 20:13:46.598417 2883 reconciler.go:26] "Reconciler: start to sync state" Feb 13 20:13:46.598517 kubelet[2883]: I0213 20:13:46.598462 2883 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 20:13:46.599207 kubelet[2883]: W0213 20:13:46.598778 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.4.48:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:46.599207 kubelet[2883]: E0213 20:13:46.598834 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.4.48:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:46.600470 kubelet[2883]: I0213 20:13:46.600446 2883 factory.go:221] Registration of the containerd container factory successfully Feb 13 20:13:46.605428 kubelet[2883]: E0213 20:13:46.605407 2883 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 20:13:46.632112 kubelet[2883]: I0213 20:13:46.632078 2883 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 20:13:46.634133 kubelet[2883]: I0213 20:13:46.633990 2883 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 20:13:46.634133 kubelet[2883]: I0213 20:13:46.634015 2883 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 20:13:46.634133 kubelet[2883]: I0213 20:13:46.634036 2883 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 20:13:46.634133 kubelet[2883]: I0213 20:13:46.634046 2883 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 20:13:46.634133 kubelet[2883]: E0213 20:13:46.634097 2883 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 20:13:46.637614 kubelet[2883]: W0213 20:13:46.637459 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.4.48:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:46.637982 kubelet[2883]: E0213 20:13:46.637916 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.4.48:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:46.643018 kubelet[2883]: I0213 20:13:46.642986 2883 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 20:13:46.643095 kubelet[2883]: I0213 20:13:46.643059 2883 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 20:13:46.643095 kubelet[2883]: I0213 20:13:46.643080 2883 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:13:46.647010 kubelet[2883]: I0213 20:13:46.646989 2883 policy_none.go:49] "None policy: Start" Feb 13 20:13:46.647010 kubelet[2883]: I0213 20:13:46.647010 2883 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 20:13:46.647116 kubelet[2883]: I0213 20:13:46.647025 2883 state_mem.go:35] "Initializing new in-memory state store" Feb 13 20:13:46.655880 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 20:13:46.666209 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 20:13:46.678888 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 20:13:46.680280 kubelet[2883]: I0213 20:13:46.680119 2883 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 20:13:46.680442 kubelet[2883]: I0213 20:13:46.680429 2883 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 20:13:46.680565 kubelet[2883]: I0213 20:13:46.680526 2883 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 20:13:46.681113 kubelet[2883]: I0213 20:13:46.681019 2883 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 20:13:46.682653 kubelet[2883]: E0213 20:13:46.682628 2883 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 20:13:46.682731 kubelet[2883]: E0213 20:13:46.682691 2883 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4152.2.1-a-2754074fca\" not found" Feb 13 20:13:46.748112 systemd[1]: Created slice kubepods-burstable-podbc72a1cbc9282424ad6161167f2dcd7c.slice - libcontainer container kubepods-burstable-podbc72a1cbc9282424ad6161167f2dcd7c.slice. Feb 13 20:13:46.764115 kubelet[2883]: E0213 20:13:46.764076 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.767484 systemd[1]: Created slice kubepods-burstable-pod91c34995111fc0d93acda1be437bb358.slice - libcontainer container kubepods-burstable-pod91c34995111fc0d93acda1be437bb358.slice. Feb 13 20:13:46.775910 kubelet[2883]: E0213 20:13:46.775862 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.778766 systemd[1]: Created slice kubepods-burstable-podf9be579e866b7a058777d411d45d7b9f.slice - libcontainer container kubepods-burstable-podf9be579e866b7a058777d411d45d7b9f.slice. Feb 13 20:13:46.781066 kubelet[2883]: E0213 20:13:46.780754 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.783347 kubelet[2883]: I0213 20:13:46.783294 2883 kubelet_node_status.go:76] "Attempting to register node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.784017 kubelet[2883]: E0213 20:13:46.783988 2883 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.200.4.48:6443/api/v1/nodes\": dial tcp 10.200.4.48:6443: connect: connection refused" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.798594 kubelet[2883]: E0213 20:13:46.798543 2883 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.48:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4152.2.1-a-2754074fca?timeout=10s\": dial tcp 10.200.4.48:6443: connect: connection refused" interval="400ms" Feb 13 20:13:46.799806 kubelet[2883]: I0213 20:13:46.799567 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f9be579e866b7a058777d411d45d7b9f-k8s-certs\") pod \"kube-apiserver-ci-4152.2.1-a-2754074fca\" (UID: \"f9be579e866b7a058777d411d45d7b9f\") " pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.799806 kubelet[2883]: I0213 20:13:46.799600 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-ca-certs\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.799806 kubelet[2883]: I0213 20:13:46.799621 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-flexvolume-dir\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.799806 kubelet[2883]: I0213 20:13:46.799637 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-k8s-certs\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.799806 kubelet[2883]: I0213 20:13:46.799654 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-kubeconfig\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.800009 kubelet[2883]: I0213 20:13:46.799674 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.800009 kubelet[2883]: I0213 20:13:46.799692 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/91c34995111fc0d93acda1be437bb358-kubeconfig\") pod \"kube-scheduler-ci-4152.2.1-a-2754074fca\" (UID: \"91c34995111fc0d93acda1be437bb358\") " pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.800009 kubelet[2883]: I0213 20:13:46.799706 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f9be579e866b7a058777d411d45d7b9f-ca-certs\") pod \"kube-apiserver-ci-4152.2.1-a-2754074fca\" (UID: \"f9be579e866b7a058777d411d45d7b9f\") " pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.800009 kubelet[2883]: I0213 20:13:46.799720 2883 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f9be579e866b7a058777d411d45d7b9f-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4152.2.1-a-2754074fca\" (UID: \"f9be579e866b7a058777d411d45d7b9f\") " pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.986727 kubelet[2883]: I0213 20:13:46.986685 2883 kubelet_node_status.go:76] "Attempting to register node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:46.987183 kubelet[2883]: E0213 20:13:46.987141 2883 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.200.4.48:6443/api/v1/nodes\": dial tcp 10.200.4.48:6443: connect: connection refused" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:47.066393 containerd[1714]: time="2025-02-13T20:13:47.066064218Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4152.2.1-a-2754074fca,Uid:bc72a1cbc9282424ad6161167f2dcd7c,Namespace:kube-system,Attempt:0,}" Feb 13 20:13:47.077689 containerd[1714]: time="2025-02-13T20:13:47.077647772Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4152.2.1-a-2754074fca,Uid:91c34995111fc0d93acda1be437bb358,Namespace:kube-system,Attempt:0,}" Feb 13 20:13:47.082928 containerd[1714]: time="2025-02-13T20:13:47.082577080Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4152.2.1-a-2754074fca,Uid:f9be579e866b7a058777d411d45d7b9f,Namespace:kube-system,Attempt:0,}" Feb 13 20:13:47.199562 kubelet[2883]: E0213 20:13:47.199507 2883 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.48:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4152.2.1-a-2754074fca?timeout=10s\": dial tcp 10.200.4.48:6443: connect: connection refused" interval="800ms" Feb 13 20:13:47.388974 kubelet[2883]: I0213 20:13:47.388936 2883 kubelet_node_status.go:76] "Attempting to register node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:47.389386 kubelet[2883]: E0213 20:13:47.389350 2883 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.200.4.48:6443/api/v1/nodes\": dial tcp 10.200.4.48:6443: connect: connection refused" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:47.676669 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2342483984.mount: Deactivated successfully. Feb 13 20:13:47.700629 containerd[1714]: time="2025-02-13T20:13:47.700569827Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:13:47.709531 containerd[1714]: time="2025-02-13T20:13:47.709478622Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:13:47.718983 containerd[1714]: time="2025-02-13T20:13:47.718925629Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312064" Feb 13 20:13:47.721997 containerd[1714]: time="2025-02-13T20:13:47.721944195Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 20:13:47.736027 containerd[1714]: time="2025-02-13T20:13:47.735962802Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:13:47.746767 containerd[1714]: time="2025-02-13T20:13:47.746690338Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:13:47.751297 containerd[1714]: time="2025-02-13T20:13:47.750616124Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 20:13:47.756703 containerd[1714]: time="2025-02-13T20:13:47.756658456Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:13:47.757577 containerd[1714]: time="2025-02-13T20:13:47.757544975Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 691.359154ms" Feb 13 20:13:47.759701 containerd[1714]: time="2025-02-13T20:13:47.759670322Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 676.99964ms" Feb 13 20:13:47.767880 containerd[1714]: time="2025-02-13T20:13:47.767849801Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 690.101527ms" Feb 13 20:13:47.834542 kubelet[2883]: W0213 20:13:47.834153 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.4.48:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:47.834795 kubelet[2883]: E0213 20:13:47.834662 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.4.48:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:47.932696 kubelet[2883]: W0213 20:13:47.932529 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.4.48:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:47.932696 kubelet[2883]: E0213 20:13:47.932607 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.4.48:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:48.001919 kubelet[2883]: E0213 20:13:48.000756 2883 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.48:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4152.2.1-a-2754074fca?timeout=10s\": dial tcp 10.200.4.48:6443: connect: connection refused" interval="1.6s" Feb 13 20:13:48.133686 kubelet[2883]: W0213 20:13:48.133558 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.4.48:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:48.133686 kubelet[2883]: E0213 20:13:48.133653 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.4.48:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:48.133936 kubelet[2883]: W0213 20:13:48.133889 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.4.48:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4152.2.1-a-2754074fca&limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:48.133993 kubelet[2883]: E0213 20:13:48.133961 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.4.48:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4152.2.1-a-2754074fca&limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:48.193019 kubelet[2883]: I0213 20:13:48.192368 2883 kubelet_node_status.go:76] "Attempting to register node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:48.193019 kubelet[2883]: E0213 20:13:48.192816 2883 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.200.4.48:6443/api/v1/nodes\": dial tcp 10.200.4.48:6443: connect: connection refused" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:48.616288 containerd[1714]: time="2025-02-13T20:13:48.609405050Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:13:48.616288 containerd[1714]: time="2025-02-13T20:13:48.609464351Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:13:48.616288 containerd[1714]: time="2025-02-13T20:13:48.609483352Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:13:48.616288 containerd[1714]: time="2025-02-13T20:13:48.609589154Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:13:48.617344 containerd[1714]: time="2025-02-13T20:13:48.609225146Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:13:48.617344 containerd[1714]: time="2025-02-13T20:13:48.609410650Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:13:48.617344 containerd[1714]: time="2025-02-13T20:13:48.609441951Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:13:48.617344 containerd[1714]: time="2025-02-13T20:13:48.610390972Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:13:48.617344 containerd[1714]: time="2025-02-13T20:13:48.609124344Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:13:48.617344 containerd[1714]: time="2025-02-13T20:13:48.614009652Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:13:48.617344 containerd[1714]: time="2025-02-13T20:13:48.614035453Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:13:48.619818 containerd[1714]: time="2025-02-13T20:13:48.619727278Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:13:48.654486 systemd[1]: Started cri-containerd-2585c06e4136aa92258e3c02dbbddf1b1158605f0fc9c007424b2c9f1c9730e3.scope - libcontainer container 2585c06e4136aa92258e3c02dbbddf1b1158605f0fc9c007424b2c9f1c9730e3. Feb 13 20:13:48.657974 systemd[1]: Started cri-containerd-f896e8f175b9b3452ad5ea9c1b71454057b3602a7329d5999f39722d5722e244.scope - libcontainer container f896e8f175b9b3452ad5ea9c1b71454057b3602a7329d5999f39722d5722e244. Feb 13 20:13:48.663290 systemd[1]: Started cri-containerd-5e9a6ec2b34686af17eabff012560f4c5f772d17cecbb7bb6888ad23d703326a.scope - libcontainer container 5e9a6ec2b34686af17eabff012560f4c5f772d17cecbb7bb6888ad23d703326a. Feb 13 20:13:48.733456 kubelet[2883]: E0213 20:13:48.733412 2883 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.200.4.48:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:48.765827 containerd[1714]: time="2025-02-13T20:13:48.765765808Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4152.2.1-a-2754074fca,Uid:f9be579e866b7a058777d411d45d7b9f,Namespace:kube-system,Attempt:0,} returns sandbox id \"2585c06e4136aa92258e3c02dbbddf1b1158605f0fc9c007424b2c9f1c9730e3\"" Feb 13 20:13:48.770914 containerd[1714]: time="2025-02-13T20:13:48.770876521Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4152.2.1-a-2754074fca,Uid:bc72a1cbc9282424ad6161167f2dcd7c,Namespace:kube-system,Attempt:0,} returns sandbox id \"5e9a6ec2b34686af17eabff012560f4c5f772d17cecbb7bb6888ad23d703326a\"" Feb 13 20:13:48.772234 containerd[1714]: time="2025-02-13T20:13:48.772112848Z" level=info msg="CreateContainer within sandbox \"2585c06e4136aa92258e3c02dbbddf1b1158605f0fc9c007424b2c9f1c9730e3\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 20:13:48.774517 containerd[1714]: time="2025-02-13T20:13:48.774491901Z" level=info msg="CreateContainer within sandbox \"5e9a6ec2b34686af17eabff012560f4c5f772d17cecbb7bb6888ad23d703326a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 20:13:48.776319 containerd[1714]: time="2025-02-13T20:13:48.776220739Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4152.2.1-a-2754074fca,Uid:91c34995111fc0d93acda1be437bb358,Namespace:kube-system,Attempt:0,} returns sandbox id \"f896e8f175b9b3452ad5ea9c1b71454057b3602a7329d5999f39722d5722e244\"" Feb 13 20:13:48.778747 containerd[1714]: time="2025-02-13T20:13:48.778722794Z" level=info msg="CreateContainer within sandbox \"f896e8f175b9b3452ad5ea9c1b71454057b3602a7329d5999f39722d5722e244\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 20:13:48.898860 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2994274302.mount: Deactivated successfully. Feb 13 20:13:49.604583 kubelet[2883]: E0213 20:13:49.604522 2883 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.48:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4152.2.1-a-2754074fca?timeout=10s\": dial tcp 10.200.4.48:6443: connect: connection refused" interval="3.2s" Feb 13 20:13:49.796030 kubelet[2883]: I0213 20:13:49.795996 2883 kubelet_node_status.go:76] "Attempting to register node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:49.796626 kubelet[2883]: E0213 20:13:49.796586 2883 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.200.4.48:6443/api/v1/nodes\": dial tcp 10.200.4.48:6443: connect: connection refused" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:50.107387 kubelet[2883]: W0213 20:13:50.107335 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.4.48:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:50.107593 kubelet[2883]: E0213 20:13:50.107395 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.4.48:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:50.831191 kubelet[2883]: W0213 20:13:50.831143 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.4.48:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:50.873837 kubelet[2883]: E0213 20:13:50.831206 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.4.48:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:50.897124 kubelet[2883]: W0213 20:13:50.897079 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.4.48:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4152.2.1-a-2754074fca&limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:50.897124 kubelet[2883]: E0213 20:13:50.897130 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.4.48:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4152.2.1-a-2754074fca&limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:50.969349 kubelet[2883]: E0213 20:13:50.969208 2883 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.4.48:6443/api/v1/namespaces/default/events\": dial tcp 10.200.4.48:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4152.2.1-a-2754074fca.1823ddaf6b4910c6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4152.2.1-a-2754074fca,UID:ci-4152.2.1-a-2754074fca,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4152.2.1-a-2754074fca,},FirstTimestamp:2025-02-13 20:13:46.584109254 +0000 UTC m=+1.052868879,LastTimestamp:2025-02-13 20:13:46.584109254 +0000 UTC m=+1.052868879,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4152.2.1-a-2754074fca,}" Feb 13 20:13:51.137822 kubelet[2883]: W0213 20:13:51.137772 2883 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.4.48:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.4.48:6443: connect: connection refused Feb 13 20:13:51.138001 kubelet[2883]: E0213 20:13:51.137833 2883 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.4.48:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.4.48:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:13:51.676072 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2742929144.mount: Deactivated successfully. Feb 13 20:13:52.635101 containerd[1714]: time="2025-02-13T20:13:52.635044266Z" level=info msg="CreateContainer within sandbox \"2585c06e4136aa92258e3c02dbbddf1b1158605f0fc9c007424b2c9f1c9730e3\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0c8634624836aaaa7e628808f655b6cd3e94ecc571c605fece2a59440e6d7a89\"" Feb 13 20:13:52.637097 containerd[1714]: time="2025-02-13T20:13:52.635837883Z" level=info msg="StartContainer for \"0c8634624836aaaa7e628808f655b6cd3e94ecc571c605fece2a59440e6d7a89\"" Feb 13 20:13:52.654832 containerd[1714]: time="2025-02-13T20:13:52.654728901Z" level=info msg="CreateContainer within sandbox \"5e9a6ec2b34686af17eabff012560f4c5f772d17cecbb7bb6888ad23d703326a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"9afad47f4691cc86ba936c47c734e6fe4fa1caad88c71937b5c0a569fb326b8a\"" Feb 13 20:13:52.657104 containerd[1714]: time="2025-02-13T20:13:52.657069853Z" level=info msg="CreateContainer within sandbox \"f896e8f175b9b3452ad5ea9c1b71454057b3602a7329d5999f39722d5722e244\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"7a509b3abd7361d9481c73b3a0d2e21b1da5078423e0c9233a97acbf68962b28\"" Feb 13 20:13:52.657456 containerd[1714]: time="2025-02-13T20:13:52.657430061Z" level=info msg="StartContainer for \"9afad47f4691cc86ba936c47c734e6fe4fa1caad88c71937b5c0a569fb326b8a\"" Feb 13 20:13:52.659430 containerd[1714]: time="2025-02-13T20:13:52.659382804Z" level=info msg="StartContainer for \"7a509b3abd7361d9481c73b3a0d2e21b1da5078423e0c9233a97acbf68962b28\"" Feb 13 20:13:52.721437 systemd[1]: Started cri-containerd-0c8634624836aaaa7e628808f655b6cd3e94ecc571c605fece2a59440e6d7a89.scope - libcontainer container 0c8634624836aaaa7e628808f655b6cd3e94ecc571c605fece2a59440e6d7a89. Feb 13 20:13:52.731464 systemd[1]: Started cri-containerd-7a509b3abd7361d9481c73b3a0d2e21b1da5078423e0c9233a97acbf68962b28.scope - libcontainer container 7a509b3abd7361d9481c73b3a0d2e21b1da5078423e0c9233a97acbf68962b28. Feb 13 20:13:52.732988 systemd[1]: Started cri-containerd-9afad47f4691cc86ba936c47c734e6fe4fa1caad88c71937b5c0a569fb326b8a.scope - libcontainer container 9afad47f4691cc86ba936c47c734e6fe4fa1caad88c71937b5c0a569fb326b8a. Feb 13 20:13:52.806870 kubelet[2883]: E0213 20:13:52.805353 2883 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.4.48:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4152.2.1-a-2754074fca?timeout=10s\": dial tcp 10.200.4.48:6443: connect: connection refused" interval="6.4s" Feb 13 20:13:52.808810 containerd[1714]: time="2025-02-13T20:13:52.808644004Z" level=info msg="StartContainer for \"0c8634624836aaaa7e628808f655b6cd3e94ecc571c605fece2a59440e6d7a89\" returns successfully" Feb 13 20:13:52.838404 containerd[1714]: time="2025-02-13T20:13:52.838348161Z" level=info msg="StartContainer for \"9afad47f4691cc86ba936c47c734e6fe4fa1caad88c71937b5c0a569fb326b8a\" returns successfully" Feb 13 20:13:52.903071 containerd[1714]: time="2025-02-13T20:13:52.902826887Z" level=info msg="StartContainer for \"7a509b3abd7361d9481c73b3a0d2e21b1da5078423e0c9233a97acbf68962b28\" returns successfully" Feb 13 20:13:53.000401 kubelet[2883]: I0213 20:13:53.000261 2883 kubelet_node_status.go:76] "Attempting to register node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:53.674665 kubelet[2883]: E0213 20:13:53.674624 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:53.682774 kubelet[2883]: E0213 20:13:53.682735 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:53.683777 kubelet[2883]: E0213 20:13:53.683753 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:54.685329 kubelet[2883]: E0213 20:13:54.685296 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:54.685794 kubelet[2883]: E0213 20:13:54.685710 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:54.687493 kubelet[2883]: E0213 20:13:54.687469 2883 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4152.2.1-a-2754074fca\" not found" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.133889 kubelet[2883]: I0213 20:13:55.133847 2883 kubelet_node_status.go:79] "Successfully registered node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.133889 kubelet[2883]: E0213 20:13:55.133894 2883 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"ci-4152.2.1-a-2754074fca\": node \"ci-4152.2.1-a-2754074fca\" not found" Feb 13 20:13:55.196776 kubelet[2883]: I0213 20:13:55.196720 2883 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.303878 kubelet[2883]: E0213 20:13:55.303654 2883 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.303878 kubelet[2883]: I0213 20:13:55.303699 2883 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.305786 kubelet[2883]: E0213 20:13:55.305756 2883 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4152.2.1-a-2754074fca\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.305786 kubelet[2883]: I0213 20:13:55.305782 2883 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.307306 kubelet[2883]: E0213 20:13:55.307281 2883 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4152.2.1-a-2754074fca\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.582400 kubelet[2883]: I0213 20:13:55.582269 2883 apiserver.go:52] "Watching apiserver" Feb 13 20:13:55.599240 kubelet[2883]: I0213 20:13:55.599204 2883 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 20:13:55.683342 kubelet[2883]: I0213 20:13:55.682976 2883 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.683342 kubelet[2883]: I0213 20:13:55.683020 2883 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.685814 kubelet[2883]: E0213 20:13:55.685563 2883 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4152.2.1-a-2754074fca\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:55.687719 kubelet[2883]: E0213 20:13:55.687505 2883 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4152.2.1-a-2754074fca\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:57.157979 systemd[1]: Reloading requested from client PID 3157 ('systemctl') (unit session-9.scope)... Feb 13 20:13:57.158009 systemd[1]: Reloading... Feb 13 20:13:57.307287 zram_generator::config[3197]: No configuration found. Feb 13 20:13:57.451647 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:13:57.557525 systemd[1]: Reloading finished in 398 ms. Feb 13 20:13:57.601584 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:57.622777 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 20:13:57.623034 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:57.629695 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:13:57.873786 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:13:57.881662 (kubelet)[3264]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 20:13:57.926226 kubelet[3264]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:13:57.926226 kubelet[3264]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 20:13:57.926226 kubelet[3264]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:13:58.333598 kubelet[3264]: I0213 20:13:57.926351 3264 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 20:13:58.333598 kubelet[3264]: I0213 20:13:57.931945 3264 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 20:13:58.333598 kubelet[3264]: I0213 20:13:57.931964 3264 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 20:13:58.333598 kubelet[3264]: I0213 20:13:57.932181 3264 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 20:13:58.334869 kubelet[3264]: I0213 20:13:58.334719 3264 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 20:13:58.343036 kubelet[3264]: I0213 20:13:58.342930 3264 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 20:13:58.349792 kubelet[3264]: E0213 20:13:58.349749 3264 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 20:13:58.349792 kubelet[3264]: I0213 20:13:58.349784 3264 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 20:13:58.355199 kubelet[3264]: I0213 20:13:58.355171 3264 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 20:13:58.355534 kubelet[3264]: I0213 20:13:58.355472 3264 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 20:13:58.357381 kubelet[3264]: I0213 20:13:58.355511 3264 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4152.2.1-a-2754074fca","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 20:13:58.357381 kubelet[3264]: I0213 20:13:58.355888 3264 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 20:13:58.357381 kubelet[3264]: I0213 20:13:58.355904 3264 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 20:13:58.357381 kubelet[3264]: I0213 20:13:58.355961 3264 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:13:58.357381 kubelet[3264]: I0213 20:13:58.356128 3264 kubelet.go:446] "Attempting to sync node with API server" Feb 13 20:13:58.357697 kubelet[3264]: I0213 20:13:58.356143 3264 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 20:13:58.357697 kubelet[3264]: I0213 20:13:58.356171 3264 kubelet.go:352] "Adding apiserver pod source" Feb 13 20:13:58.357697 kubelet[3264]: I0213 20:13:58.356183 3264 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 20:13:58.368281 kubelet[3264]: I0213 20:13:58.367540 3264 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 20:13:58.368588 kubelet[3264]: I0213 20:13:58.368571 3264 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 20:13:58.369760 kubelet[3264]: I0213 20:13:58.369738 3264 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 20:13:58.370088 kubelet[3264]: I0213 20:13:58.369778 3264 server.go:1287] "Started kubelet" Feb 13 20:13:58.371687 kubelet[3264]: I0213 20:13:58.371631 3264 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 20:13:58.381873 kubelet[3264]: I0213 20:13:58.381355 3264 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 20:13:58.382654 kubelet[3264]: I0213 20:13:58.382631 3264 server.go:490] "Adding debug handlers to kubelet server" Feb 13 20:13:58.387218 kubelet[3264]: I0213 20:13:58.387158 3264 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 20:13:58.387563 kubelet[3264]: I0213 20:13:58.387434 3264 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 20:13:58.388286 kubelet[3264]: I0213 20:13:58.388039 3264 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 20:13:58.393292 kubelet[3264]: I0213 20:13:58.393231 3264 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 20:13:58.394358 kubelet[3264]: E0213 20:13:58.394325 3264 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4152.2.1-a-2754074fca\" not found" Feb 13 20:13:58.394849 kubelet[3264]: I0213 20:13:58.394822 3264 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 20:13:58.394985 kubelet[3264]: I0213 20:13:58.394968 3264 reconciler.go:26] "Reconciler: start to sync state" Feb 13 20:13:58.410099 kubelet[3264]: I0213 20:13:58.410062 3264 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 20:13:58.413012 kubelet[3264]: I0213 20:13:58.412978 3264 factory.go:221] Registration of the containerd container factory successfully Feb 13 20:13:58.413012 kubelet[3264]: I0213 20:13:58.412996 3264 factory.go:221] Registration of the systemd container factory successfully Feb 13 20:13:58.417713 kubelet[3264]: I0213 20:13:58.417316 3264 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 20:13:58.422227 kubelet[3264]: I0213 20:13:58.422199 3264 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 20:13:58.422340 kubelet[3264]: I0213 20:13:58.422263 3264 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 20:13:58.422787 kubelet[3264]: I0213 20:13:58.422299 3264 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 20:13:58.422787 kubelet[3264]: I0213 20:13:58.422539 3264 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 20:13:58.422787 kubelet[3264]: E0213 20:13:58.422621 3264 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 20:13:58.460216 sudo[3295]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Feb 13 20:13:58.460691 sudo[3295]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Feb 13 20:13:58.509910 kubelet[3264]: I0213 20:13:58.509873 3264 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 20:13:58.509910 kubelet[3264]: I0213 20:13:58.509902 3264 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 20:13:58.509910 kubelet[3264]: I0213 20:13:58.509931 3264 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:13:58.510286 kubelet[3264]: I0213 20:13:58.510199 3264 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 20:13:58.511336 kubelet[3264]: I0213 20:13:58.510240 3264 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 20:13:58.511336 kubelet[3264]: I0213 20:13:58.510454 3264 policy_none.go:49] "None policy: Start" Feb 13 20:13:58.511336 kubelet[3264]: I0213 20:13:58.510468 3264 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 20:13:58.511336 kubelet[3264]: I0213 20:13:58.510486 3264 state_mem.go:35] "Initializing new in-memory state store" Feb 13 20:13:58.511336 kubelet[3264]: I0213 20:13:58.510644 3264 state_mem.go:75] "Updated machine memory state" Feb 13 20:13:58.518460 kubelet[3264]: I0213 20:13:58.518430 3264 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 20:13:58.518899 kubelet[3264]: I0213 20:13:58.518638 3264 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 20:13:58.518899 kubelet[3264]: I0213 20:13:58.518657 3264 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 20:13:58.520980 kubelet[3264]: I0213 20:13:58.520952 3264 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 20:13:58.524005 kubelet[3264]: E0213 20:13:58.523978 3264 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 20:13:58.525766 kubelet[3264]: I0213 20:13:58.525722 3264 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.531018 kubelet[3264]: I0213 20:13:58.528720 3264 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.531764 kubelet[3264]: I0213 20:13:58.531714 3264 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.552280 kubelet[3264]: W0213 20:13:58.548481 3264 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Feb 13 20:13:58.555834 kubelet[3264]: W0213 20:13:58.555797 3264 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Feb 13 20:13:58.558365 kubelet[3264]: W0213 20:13:58.558329 3264 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Feb 13 20:13:58.650302 kubelet[3264]: I0213 20:13:58.646517 3264 kubelet_node_status.go:76] "Attempting to register node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.663740 kubelet[3264]: I0213 20:13:58.663688 3264 kubelet_node_status.go:125] "Node was previously registered" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.663973 kubelet[3264]: I0213 20:13:58.663816 3264 kubelet_node_status.go:79] "Successfully registered node" node="ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700398 kubelet[3264]: I0213 20:13:58.700063 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-kubeconfig\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700398 kubelet[3264]: I0213 20:13:58.700116 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/91c34995111fc0d93acda1be437bb358-kubeconfig\") pod \"kube-scheduler-ci-4152.2.1-a-2754074fca\" (UID: \"91c34995111fc0d93acda1be437bb358\") " pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700398 kubelet[3264]: I0213 20:13:58.700143 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f9be579e866b7a058777d411d45d7b9f-ca-certs\") pod \"kube-apiserver-ci-4152.2.1-a-2754074fca\" (UID: \"f9be579e866b7a058777d411d45d7b9f\") " pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700398 kubelet[3264]: I0213 20:13:58.700161 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-flexvolume-dir\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700398 kubelet[3264]: I0213 20:13:58.700180 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-ca-certs\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700687 kubelet[3264]: I0213 20:13:58.700197 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-k8s-certs\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700687 kubelet[3264]: I0213 20:13:58.700216 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bc72a1cbc9282424ad6161167f2dcd7c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4152.2.1-a-2754074fca\" (UID: \"bc72a1cbc9282424ad6161167f2dcd7c\") " pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700687 kubelet[3264]: I0213 20:13:58.700237 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f9be579e866b7a058777d411d45d7b9f-k8s-certs\") pod \"kube-apiserver-ci-4152.2.1-a-2754074fca\" (UID: \"f9be579e866b7a058777d411d45d7b9f\") " pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:58.700687 kubelet[3264]: I0213 20:13:58.700277 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f9be579e866b7a058777d411d45d7b9f-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4152.2.1-a-2754074fca\" (UID: \"f9be579e866b7a058777d411d45d7b9f\") " pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" Feb 13 20:13:59.077688 sudo[3295]: pam_unix(sudo:session): session closed for user root Feb 13 20:13:59.361706 kubelet[3264]: I0213 20:13:59.361387 3264 apiserver.go:52] "Watching apiserver" Feb 13 20:13:59.395078 kubelet[3264]: I0213 20:13:59.394979 3264 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 20:13:59.448699 kubelet[3264]: I0213 20:13:59.448653 3264 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:59.462959 kubelet[3264]: W0213 20:13:59.462916 3264 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Feb 13 20:13:59.469504 kubelet[3264]: E0213 20:13:59.468878 3264 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4152.2.1-a-2754074fca\" already exists" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" Feb 13 20:13:59.500146 kubelet[3264]: I0213 20:13:59.500063 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4152.2.1-a-2754074fca" podStartSLOduration=1.500040785 podStartE2EDuration="1.500040785s" podCreationTimestamp="2025-02-13 20:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:13:59.49978498 +0000 UTC m=+1.613207596" watchObservedRunningTime="2025-02-13 20:13:59.500040785 +0000 UTC m=+1.613463301" Feb 13 20:13:59.513181 kubelet[3264]: I0213 20:13:59.512967 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4152.2.1-a-2754074fca" podStartSLOduration=1.512944147 podStartE2EDuration="1.512944147s" podCreationTimestamp="2025-02-13 20:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:13:59.511863526 +0000 UTC m=+1.625286042" watchObservedRunningTime="2025-02-13 20:13:59.512944147 +0000 UTC m=+1.626366663" Feb 13 20:13:59.544153 kubelet[3264]: I0213 20:13:59.543974 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4152.2.1-a-2754074fca" podStartSLOduration=1.543951078 podStartE2EDuration="1.543951078s" podCreationTimestamp="2025-02-13 20:13:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:13:59.528776369 +0000 UTC m=+1.642198885" watchObservedRunningTime="2025-02-13 20:13:59.543951078 +0000 UTC m=+1.657373594" Feb 13 20:14:00.572381 sudo[2270]: pam_unix(sudo:session): session closed for user root Feb 13 20:14:00.676784 sshd[2269]: Connection closed by 10.200.16.10 port 42434 Feb 13 20:14:00.677547 sshd-session[2267]: pam_unix(sshd:session): session closed for user core Feb 13 20:14:00.681764 systemd[1]: sshd@6-10.200.4.48:22-10.200.16.10:42434.service: Deactivated successfully. Feb 13 20:14:00.684010 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 20:14:00.684211 systemd[1]: session-9.scope: Consumed 5.180s CPU time, 154.1M memory peak, 0B memory swap peak. Feb 13 20:14:00.685157 systemd-logind[1685]: Session 9 logged out. Waiting for processes to exit. Feb 13 20:14:00.686423 systemd-logind[1685]: Removed session 9. Feb 13 20:14:01.707019 kubelet[3264]: I0213 20:14:01.706979 3264 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 20:14:01.707614 containerd[1714]: time="2025-02-13T20:14:01.707409059Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 20:14:01.708612 kubelet[3264]: I0213 20:14:01.707619 3264 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 20:14:02.482021 systemd[1]: Created slice kubepods-besteffort-podb6759fa7_e3ca_4031_9619_f5a016789850.slice - libcontainer container kubepods-besteffort-podb6759fa7_e3ca_4031_9619_f5a016789850.slice. Feb 13 20:14:02.520906 systemd[1]: Created slice kubepods-burstable-podb0f8857d_4ed0_46a7_92d5_9c4a58d29dd3.slice - libcontainer container kubepods-burstable-podb0f8857d_4ed0_46a7_92d5_9c4a58d29dd3.slice. Feb 13 20:14:02.524970 kubelet[3264]: I0213 20:14:02.524922 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/b6759fa7-e3ca-4031-9619-f5a016789850-kube-proxy\") pod \"kube-proxy-w4bpj\" (UID: \"b6759fa7-e3ca-4031-9619-f5a016789850\") " pod="kube-system/kube-proxy-w4bpj" Feb 13 20:14:02.524970 kubelet[3264]: I0213 20:14:02.524974 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-clustermesh-secrets\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525217 kubelet[3264]: I0213 20:14:02.525001 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b6759fa7-e3ca-4031-9619-f5a016789850-lib-modules\") pod \"kube-proxy-w4bpj\" (UID: \"b6759fa7-e3ca-4031-9619-f5a016789850\") " pod="kube-system/kube-proxy-w4bpj" Feb 13 20:14:02.525217 kubelet[3264]: I0213 20:14:02.525025 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-bpf-maps\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525217 kubelet[3264]: I0213 20:14:02.525047 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cni-path\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525217 kubelet[3264]: I0213 20:14:02.525070 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-lib-modules\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525217 kubelet[3264]: I0213 20:14:02.525092 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-config-path\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525217 kubelet[3264]: I0213 20:14:02.525113 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-net\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525469 kubelet[3264]: I0213 20:14:02.525132 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-run\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525469 kubelet[3264]: I0213 20:14:02.525151 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hostproc\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525469 kubelet[3264]: I0213 20:14:02.525172 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6v8f\" (UniqueName: \"kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-kube-api-access-s6v8f\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.525469 kubelet[3264]: I0213 20:14:02.525198 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b6759fa7-e3ca-4031-9619-f5a016789850-xtables-lock\") pod \"kube-proxy-w4bpj\" (UID: \"b6759fa7-e3ca-4031-9619-f5a016789850\") " pod="kube-system/kube-proxy-w4bpj" Feb 13 20:14:02.525469 kubelet[3264]: I0213 20:14:02.525220 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-cgroup\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.529646 kubelet[3264]: I0213 20:14:02.525245 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-xtables-lock\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.529780 kubelet[3264]: I0213 20:14:02.529715 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-kernel\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.530021 kubelet[3264]: I0213 20:14:02.529993 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6zjg\" (UniqueName: \"kubernetes.io/projected/b6759fa7-e3ca-4031-9619-f5a016789850-kube-api-access-p6zjg\") pod \"kube-proxy-w4bpj\" (UID: \"b6759fa7-e3ca-4031-9619-f5a016789850\") " pod="kube-system/kube-proxy-w4bpj" Feb 13 20:14:02.530177 kubelet[3264]: I0213 20:14:02.530153 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-etc-cni-netd\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.530326 kubelet[3264]: I0213 20:14:02.530304 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hubble-tls\") pod \"cilium-msm75\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " pod="kube-system/cilium-msm75" Feb 13 20:14:02.792159 containerd[1714]: time="2025-02-13T20:14:02.792028854Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-w4bpj,Uid:b6759fa7-e3ca-4031-9619-f5a016789850,Namespace:kube-system,Attempt:0,}" Feb 13 20:14:02.831913 containerd[1714]: time="2025-02-13T20:14:02.831857321Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-msm75,Uid:b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3,Namespace:kube-system,Attempt:0,}" Feb 13 20:14:02.907221 systemd[1]: Created slice kubepods-besteffort-pod6ba2f865_bcae_4078_92fd_ff4c68c1d9f1.slice - libcontainer container kubepods-besteffort-pod6ba2f865_bcae_4078_92fd_ff4c68c1d9f1.slice. Feb 13 20:14:02.940170 kubelet[3264]: I0213 20:14:02.940098 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-rwzmz\" (UID: \"6ba2f865-bcae-4078-92fd-ff4c68c1d9f1\") " pod="kube-system/cilium-operator-6c4d7847fc-rwzmz" Feb 13 20:14:02.940170 kubelet[3264]: I0213 20:14:02.940175 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t92sm\" (UniqueName: \"kubernetes.io/projected/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-kube-api-access-t92sm\") pod \"cilium-operator-6c4d7847fc-rwzmz\" (UID: \"6ba2f865-bcae-4078-92fd-ff4c68c1d9f1\") " pod="kube-system/cilium-operator-6c4d7847fc-rwzmz" Feb 13 20:14:02.999329 containerd[1714]: time="2025-02-13T20:14:02.999132264Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:14:02.999329 containerd[1714]: time="2025-02-13T20:14:02.999241067Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:14:02.999329 containerd[1714]: time="2025-02-13T20:14:02.999278568Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:03.000656 containerd[1714]: time="2025-02-13T20:14:02.999376970Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:03.001160 containerd[1714]: time="2025-02-13T20:14:03.001061807Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:14:03.001399 containerd[1714]: time="2025-02-13T20:14:03.001345913Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:14:03.001564 containerd[1714]: time="2025-02-13T20:14:03.001522417Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:03.001886 containerd[1714]: time="2025-02-13T20:14:03.001838323Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:03.033571 systemd[1]: Started cri-containerd-459b65a0df4a30c3291ca911f27645886788f99b4c65db856af6c1c11acccdfe.scope - libcontainer container 459b65a0df4a30c3291ca911f27645886788f99b4c65db856af6c1c11acccdfe. Feb 13 20:14:03.039276 systemd[1]: Started cri-containerd-4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170.scope - libcontainer container 4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170. Feb 13 20:14:03.089884 containerd[1714]: time="2025-02-13T20:14:03.089808339Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-w4bpj,Uid:b6759fa7-e3ca-4031-9619-f5a016789850,Namespace:kube-system,Attempt:0,} returns sandbox id \"459b65a0df4a30c3291ca911f27645886788f99b4c65db856af6c1c11acccdfe\"" Feb 13 20:14:03.098091 containerd[1714]: time="2025-02-13T20:14:03.097763213Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-msm75,Uid:b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3,Namespace:kube-system,Attempt:0,} returns sandbox id \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\"" Feb 13 20:14:03.099371 containerd[1714]: time="2025-02-13T20:14:03.099196544Z" level=info msg="CreateContainer within sandbox \"459b65a0df4a30c3291ca911f27645886788f99b4c65db856af6c1c11acccdfe\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 20:14:03.101289 containerd[1714]: time="2025-02-13T20:14:03.100638175Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Feb 13 20:14:03.151545 containerd[1714]: time="2025-02-13T20:14:03.151483783Z" level=info msg="CreateContainer within sandbox \"459b65a0df4a30c3291ca911f27645886788f99b4c65db856af6c1c11acccdfe\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"aca5a90dcc853b21696d88c384688a3ea866fac414cebc10efea18a8504c1219\"" Feb 13 20:14:03.153810 containerd[1714]: time="2025-02-13T20:14:03.152285700Z" level=info msg="StartContainer for \"aca5a90dcc853b21696d88c384688a3ea866fac414cebc10efea18a8504c1219\"" Feb 13 20:14:03.180482 systemd[1]: Started cri-containerd-aca5a90dcc853b21696d88c384688a3ea866fac414cebc10efea18a8504c1219.scope - libcontainer container aca5a90dcc853b21696d88c384688a3ea866fac414cebc10efea18a8504c1219. Feb 13 20:14:03.213184 containerd[1714]: time="2025-02-13T20:14:03.212865620Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-rwzmz,Uid:6ba2f865-bcae-4078-92fd-ff4c68c1d9f1,Namespace:kube-system,Attempt:0,}" Feb 13 20:14:03.220521 containerd[1714]: time="2025-02-13T20:14:03.220474785Z" level=info msg="StartContainer for \"aca5a90dcc853b21696d88c384688a3ea866fac414cebc10efea18a8504c1219\" returns successfully" Feb 13 20:14:03.270219 containerd[1714]: time="2025-02-13T20:14:03.270127667Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:14:03.270403 containerd[1714]: time="2025-02-13T20:14:03.270304171Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:14:03.270403 containerd[1714]: time="2025-02-13T20:14:03.270340972Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:03.270528 containerd[1714]: time="2025-02-13T20:14:03.270438874Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:03.292471 systemd[1]: Started cri-containerd-77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5.scope - libcontainer container 77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5. Feb 13 20:14:03.349297 containerd[1714]: time="2025-02-13T20:14:03.348358171Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-rwzmz,Uid:6ba2f865-bcae-4078-92fd-ff4c68c1d9f1,Namespace:kube-system,Attempt:0,} returns sandbox id \"77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5\"" Feb 13 20:14:03.473152 kubelet[3264]: I0213 20:14:03.472809 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-w4bpj" podStartSLOduration=1.472787681 podStartE2EDuration="1.472787681s" podCreationTimestamp="2025-02-13 20:14:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:14:03.472096166 +0000 UTC m=+5.585518782" watchObservedRunningTime="2025-02-13 20:14:03.472787681 +0000 UTC m=+5.586210197" Feb 13 20:14:08.277459 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2115141504.mount: Deactivated successfully. Feb 13 20:14:10.551643 containerd[1714]: time="2025-02-13T20:14:10.551575002Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:14:10.553932 containerd[1714]: time="2025-02-13T20:14:10.553816452Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Feb 13 20:14:10.557267 containerd[1714]: time="2025-02-13T20:14:10.557192327Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:14:10.558900 containerd[1714]: time="2025-02-13T20:14:10.558726862Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 7.457936083s" Feb 13 20:14:10.558900 containerd[1714]: time="2025-02-13T20:14:10.558769563Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Feb 13 20:14:10.561305 containerd[1714]: time="2025-02-13T20:14:10.560871909Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Feb 13 20:14:10.562328 containerd[1714]: time="2025-02-13T20:14:10.562297741Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Feb 13 20:14:10.602739 containerd[1714]: time="2025-02-13T20:14:10.602691640Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\"" Feb 13 20:14:10.603419 containerd[1714]: time="2025-02-13T20:14:10.603384455Z" level=info msg="StartContainer for \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\"" Feb 13 20:14:10.642461 systemd[1]: Started cri-containerd-4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33.scope - libcontainer container 4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33. Feb 13 20:14:10.673845 containerd[1714]: time="2025-02-13T20:14:10.673407513Z" level=info msg="StartContainer for \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\" returns successfully" Feb 13 20:14:10.687202 systemd[1]: cri-containerd-4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33.scope: Deactivated successfully. Feb 13 20:14:11.586997 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33-rootfs.mount: Deactivated successfully. Feb 13 20:14:14.373885 containerd[1714]: time="2025-02-13T20:14:14.373805302Z" level=info msg="shim disconnected" id=4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33 namespace=k8s.io Feb 13 20:14:14.373885 containerd[1714]: time="2025-02-13T20:14:14.373877504Z" level=warning msg="cleaning up after shim disconnected" id=4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33 namespace=k8s.io Feb 13 20:14:14.373885 containerd[1714]: time="2025-02-13T20:14:14.373888504Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:14:14.390623 containerd[1714]: time="2025-02-13T20:14:14.390540967Z" level=warning msg="cleanup warnings time=\"2025-02-13T20:14:14Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Feb 13 20:14:14.496320 containerd[1714]: time="2025-02-13T20:14:14.496240170Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Feb 13 20:14:14.550783 containerd[1714]: time="2025-02-13T20:14:14.550732657Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\"" Feb 13 20:14:14.551457 containerd[1714]: time="2025-02-13T20:14:14.551393471Z" level=info msg="StartContainer for \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\"" Feb 13 20:14:14.590476 systemd[1]: Started cri-containerd-21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610.scope - libcontainer container 21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610. Feb 13 20:14:14.626452 containerd[1714]: time="2025-02-13T20:14:14.626246302Z" level=info msg="StartContainer for \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\" returns successfully" Feb 13 20:14:14.639016 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 20:14:14.639949 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:14:14.640041 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Feb 13 20:14:14.645579 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 20:14:14.645917 systemd[1]: cri-containerd-21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610.scope: Deactivated successfully. Feb 13 20:14:14.676545 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610-rootfs.mount: Deactivated successfully. Feb 13 20:14:14.678676 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:14:14.694869 containerd[1714]: time="2025-02-13T20:14:14.694769195Z" level=info msg="shim disconnected" id=21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610 namespace=k8s.io Feb 13 20:14:14.694869 containerd[1714]: time="2025-02-13T20:14:14.694840096Z" level=warning msg="cleaning up after shim disconnected" id=21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610 namespace=k8s.io Feb 13 20:14:14.694869 containerd[1714]: time="2025-02-13T20:14:14.694853196Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:14:15.511023 containerd[1714]: time="2025-02-13T20:14:15.510605468Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Feb 13 20:14:15.558830 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1536488029.mount: Deactivated successfully. Feb 13 20:14:15.575788 containerd[1714]: time="2025-02-13T20:14:15.575478782Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\"" Feb 13 20:14:15.578336 containerd[1714]: time="2025-02-13T20:14:15.576883912Z" level=info msg="StartContainer for \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\"" Feb 13 20:14:15.637021 systemd[1]: Started cri-containerd-ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309.scope - libcontainer container ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309. Feb 13 20:14:15.701813 systemd[1]: cri-containerd-ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309.scope: Deactivated successfully. Feb 13 20:14:15.707720 containerd[1714]: time="2025-02-13T20:14:15.707677462Z" level=info msg="StartContainer for \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\" returns successfully" Feb 13 20:14:15.992846 containerd[1714]: time="2025-02-13T20:14:15.992594769Z" level=info msg="shim disconnected" id=ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309 namespace=k8s.io Feb 13 20:14:15.992846 containerd[1714]: time="2025-02-13T20:14:15.992662870Z" level=warning msg="cleaning up after shim disconnected" id=ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309 namespace=k8s.io Feb 13 20:14:15.992846 containerd[1714]: time="2025-02-13T20:14:15.992675570Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:14:16.282830 containerd[1714]: time="2025-02-13T20:14:16.282419583Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:14:16.285838 containerd[1714]: time="2025-02-13T20:14:16.285609352Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Feb 13 20:14:16.290597 containerd[1714]: time="2025-02-13T20:14:16.289220631Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:14:16.290597 containerd[1714]: time="2025-02-13T20:14:16.290441758Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 5.729434145s" Feb 13 20:14:16.290597 containerd[1714]: time="2025-02-13T20:14:16.290479058Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Feb 13 20:14:16.292760 containerd[1714]: time="2025-02-13T20:14:16.292730207Z" level=info msg="CreateContainer within sandbox \"77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Feb 13 20:14:16.323592 containerd[1714]: time="2025-02-13T20:14:16.323546479Z" level=info msg="CreateContainer within sandbox \"77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\"" Feb 13 20:14:16.324137 containerd[1714]: time="2025-02-13T20:14:16.324036089Z" level=info msg="StartContainer for \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\"" Feb 13 20:14:16.350423 systemd[1]: Started cri-containerd-59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650.scope - libcontainer container 59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650. Feb 13 20:14:16.381054 containerd[1714]: time="2025-02-13T20:14:16.381016331Z" level=info msg="StartContainer for \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\" returns successfully" Feb 13 20:14:16.513057 containerd[1714]: time="2025-02-13T20:14:16.512983006Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Feb 13 20:14:16.542667 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309-rootfs.mount: Deactivated successfully. Feb 13 20:14:16.552194 kubelet[3264]: I0213 20:14:16.552093 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-rwzmz" podStartSLOduration=1.6120238150000001 podStartE2EDuration="14.552049357s" podCreationTimestamp="2025-02-13 20:14:02 +0000 UTC" firstStartedPulling="2025-02-13 20:14:03.351413737 +0000 UTC m=+5.464836253" lastFinishedPulling="2025-02-13 20:14:16.291439279 +0000 UTC m=+18.404861795" observedRunningTime="2025-02-13 20:14:16.549967011 +0000 UTC m=+18.663389527" watchObservedRunningTime="2025-02-13 20:14:16.552049357 +0000 UTC m=+18.665471873" Feb 13 20:14:16.558654 containerd[1714]: time="2025-02-13T20:14:16.558592899Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\"" Feb 13 20:14:16.563513 containerd[1714]: time="2025-02-13T20:14:16.563470106Z" level=info msg="StartContainer for \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\"" Feb 13 20:14:16.647350 systemd[1]: run-containerd-runc-k8s.io-24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e-runc.MXFPxz.mount: Deactivated successfully. Feb 13 20:14:16.661472 systemd[1]: Started cri-containerd-24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e.scope - libcontainer container 24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e. Feb 13 20:14:16.758469 systemd[1]: cri-containerd-24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e.scope: Deactivated successfully. Feb 13 20:14:16.761583 containerd[1714]: time="2025-02-13T20:14:16.761213114Z" level=info msg="StartContainer for \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\" returns successfully" Feb 13 20:14:16.987127 containerd[1714]: time="2025-02-13T20:14:16.986844429Z" level=info msg="shim disconnected" id=24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e namespace=k8s.io Feb 13 20:14:16.987127 containerd[1714]: time="2025-02-13T20:14:16.987064934Z" level=warning msg="cleaning up after shim disconnected" id=24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e namespace=k8s.io Feb 13 20:14:16.987127 containerd[1714]: time="2025-02-13T20:14:16.987078234Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:14:17.521050 containerd[1714]: time="2025-02-13T20:14:17.520179648Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Feb 13 20:14:17.540839 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e-rootfs.mount: Deactivated successfully. Feb 13 20:14:17.558884 containerd[1714]: time="2025-02-13T20:14:17.558842091Z" level=info msg="CreateContainer within sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\"" Feb 13 20:14:17.560710 containerd[1714]: time="2025-02-13T20:14:17.559463504Z" level=info msg="StartContainer for \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\"" Feb 13 20:14:17.596416 systemd[1]: Started cri-containerd-d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada.scope - libcontainer container d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada. Feb 13 20:14:17.631682 containerd[1714]: time="2025-02-13T20:14:17.631622676Z" level=info msg="StartContainer for \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\" returns successfully" Feb 13 20:14:17.778328 kubelet[3264]: I0213 20:14:17.778197 3264 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Feb 13 20:14:17.827656 systemd[1]: Created slice kubepods-burstable-pod8a06c51c_7a24_44fe_b831_1154868be4db.slice - libcontainer container kubepods-burstable-pod8a06c51c_7a24_44fe_b831_1154868be4db.slice. Feb 13 20:14:17.838035 systemd[1]: Created slice kubepods-burstable-pod0e0043b7_8129_41b0_9df0_1956609c1c3c.slice - libcontainer container kubepods-burstable-pod0e0043b7_8129_41b0_9df0_1956609c1c3c.slice. Feb 13 20:14:17.853313 kubelet[3264]: I0213 20:14:17.852056 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwtbj\" (UniqueName: \"kubernetes.io/projected/8a06c51c-7a24-44fe-b831-1154868be4db-kube-api-access-cwtbj\") pod \"coredns-668d6bf9bc-hcqz5\" (UID: \"8a06c51c-7a24-44fe-b831-1154868be4db\") " pod="kube-system/coredns-668d6bf9bc-hcqz5" Feb 13 20:14:17.853313 kubelet[3264]: I0213 20:14:17.852111 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb8wd\" (UniqueName: \"kubernetes.io/projected/0e0043b7-8129-41b0-9df0-1956609c1c3c-kube-api-access-wb8wd\") pod \"coredns-668d6bf9bc-wdn5v\" (UID: \"0e0043b7-8129-41b0-9df0-1956609c1c3c\") " pod="kube-system/coredns-668d6bf9bc-wdn5v" Feb 13 20:14:17.853313 kubelet[3264]: I0213 20:14:17.852147 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a06c51c-7a24-44fe-b831-1154868be4db-config-volume\") pod \"coredns-668d6bf9bc-hcqz5\" (UID: \"8a06c51c-7a24-44fe-b831-1154868be4db\") " pod="kube-system/coredns-668d6bf9bc-hcqz5" Feb 13 20:14:17.853313 kubelet[3264]: I0213 20:14:17.852171 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e0043b7-8129-41b0-9df0-1956609c1c3c-config-volume\") pod \"coredns-668d6bf9bc-wdn5v\" (UID: \"0e0043b7-8129-41b0-9df0-1956609c1c3c\") " pod="kube-system/coredns-668d6bf9bc-wdn5v" Feb 13 20:14:18.138470 containerd[1714]: time="2025-02-13T20:14:18.137948007Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hcqz5,Uid:8a06c51c-7a24-44fe-b831-1154868be4db,Namespace:kube-system,Attempt:0,}" Feb 13 20:14:18.144225 containerd[1714]: time="2025-02-13T20:14:18.144181043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wdn5v,Uid:0e0043b7-8129-41b0-9df0-1956609c1c3c,Namespace:kube-system,Attempt:0,}" Feb 13 20:14:18.565993 kubelet[3264]: I0213 20:14:18.565838 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-msm75" podStartSLOduration=9.105686396 podStartE2EDuration="16.565815928s" podCreationTimestamp="2025-02-13 20:14:02 +0000 UTC" firstStartedPulling="2025-02-13 20:14:03.099900259 +0000 UTC m=+5.213322775" lastFinishedPulling="2025-02-13 20:14:10.560029791 +0000 UTC m=+12.673452307" observedRunningTime="2025-02-13 20:14:18.562768862 +0000 UTC m=+20.676191478" watchObservedRunningTime="2025-02-13 20:14:18.565815928 +0000 UTC m=+20.679238444" Feb 13 20:14:19.925813 systemd-networkd[1325]: cilium_host: Link UP Feb 13 20:14:19.926000 systemd-networkd[1325]: cilium_net: Link UP Feb 13 20:14:19.926195 systemd-networkd[1325]: cilium_net: Gained carrier Feb 13 20:14:19.928620 systemd-networkd[1325]: cilium_host: Gained carrier Feb 13 20:14:19.992401 systemd-networkd[1325]: cilium_host: Gained IPv6LL Feb 13 20:14:20.126167 systemd-networkd[1325]: cilium_vxlan: Link UP Feb 13 20:14:20.126178 systemd-networkd[1325]: cilium_vxlan: Gained carrier Feb 13 20:14:20.475152 kernel: NET: Registered PF_ALG protocol family Feb 13 20:14:20.547427 systemd-networkd[1325]: cilium_net: Gained IPv6LL Feb 13 20:14:21.209491 systemd-networkd[1325]: lxc_health: Link UP Feb 13 20:14:21.228658 systemd-networkd[1325]: lxc_health: Gained carrier Feb 13 20:14:21.632493 systemd-networkd[1325]: cilium_vxlan: Gained IPv6LL Feb 13 20:14:21.722878 systemd-networkd[1325]: lxc5e12a4eefc1f: Link UP Feb 13 20:14:21.726288 kernel: eth0: renamed from tmp65c76 Feb 13 20:14:21.735540 systemd-networkd[1325]: lxc5e12a4eefc1f: Gained carrier Feb 13 20:14:21.743819 systemd-networkd[1325]: lxcf2eb7e80e160: Link UP Feb 13 20:14:21.757107 kernel: eth0: renamed from tmpce5ee Feb 13 20:14:21.769475 systemd-networkd[1325]: lxcf2eb7e80e160: Gained carrier Feb 13 20:14:22.848469 systemd-networkd[1325]: lxc_health: Gained IPv6LL Feb 13 20:14:23.424486 systemd-networkd[1325]: lxc5e12a4eefc1f: Gained IPv6LL Feb 13 20:14:23.744537 systemd-networkd[1325]: lxcf2eb7e80e160: Gained IPv6LL Feb 13 20:14:25.566755 containerd[1714]: time="2025-02-13T20:14:25.566460707Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:14:25.566755 containerd[1714]: time="2025-02-13T20:14:25.566715112Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:14:25.568516 containerd[1714]: time="2025-02-13T20:14:25.566910016Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:25.568516 containerd[1714]: time="2025-02-13T20:14:25.568464449Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:25.602582 containerd[1714]: time="2025-02-13T20:14:25.602457368Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:14:25.602860 containerd[1714]: time="2025-02-13T20:14:25.602582870Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:14:25.602860 containerd[1714]: time="2025-02-13T20:14:25.602602171Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:25.603146 containerd[1714]: time="2025-02-13T20:14:25.603004379Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:14:25.638475 systemd[1]: Started cri-containerd-65c76bf4056e2f241dc78e2b6a76c0e0e5fa0a8763ee963b41c25ea0f31a20b4.scope - libcontainer container 65c76bf4056e2f241dc78e2b6a76c0e0e5fa0a8763ee963b41c25ea0f31a20b4. Feb 13 20:14:25.656530 systemd[1]: run-containerd-runc-k8s.io-ce5ee2df7147962b661b453f9e3542995c552c7642d7f6bbe34faf862f4fdc22-runc.PfepBa.mount: Deactivated successfully. Feb 13 20:14:25.673457 systemd[1]: Started cri-containerd-ce5ee2df7147962b661b453f9e3542995c552c7642d7f6bbe34faf862f4fdc22.scope - libcontainer container ce5ee2df7147962b661b453f9e3542995c552c7642d7f6bbe34faf862f4fdc22. Feb 13 20:14:25.767123 containerd[1714]: time="2025-02-13T20:14:25.767061947Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hcqz5,Uid:8a06c51c-7a24-44fe-b831-1154868be4db,Namespace:kube-system,Attempt:0,} returns sandbox id \"65c76bf4056e2f241dc78e2b6a76c0e0e5fa0a8763ee963b41c25ea0f31a20b4\"" Feb 13 20:14:25.776282 containerd[1714]: time="2025-02-13T20:14:25.774840611Z" level=info msg="CreateContainer within sandbox \"65c76bf4056e2f241dc78e2b6a76c0e0e5fa0a8763ee963b41c25ea0f31a20b4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 20:14:25.794210 containerd[1714]: time="2025-02-13T20:14:25.791233058Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-wdn5v,Uid:0e0043b7-8129-41b0-9df0-1956609c1c3c,Namespace:kube-system,Attempt:0,} returns sandbox id \"ce5ee2df7147962b661b453f9e3542995c552c7642d7f6bbe34faf862f4fdc22\"" Feb 13 20:14:25.798335 containerd[1714]: time="2025-02-13T20:14:25.795756553Z" level=info msg="CreateContainer within sandbox \"ce5ee2df7147962b661b453f9e3542995c552c7642d7f6bbe34faf862f4fdc22\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 20:14:25.845930 containerd[1714]: time="2025-02-13T20:14:25.845872912Z" level=info msg="CreateContainer within sandbox \"65c76bf4056e2f241dc78e2b6a76c0e0e5fa0a8763ee963b41c25ea0f31a20b4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"62583dbd0c07ba4954545dc207c473ddc173512f5a1d8cd860e47a883c9f5e04\"" Feb 13 20:14:25.847899 containerd[1714]: time="2025-02-13T20:14:25.846752031Z" level=info msg="StartContainer for \"62583dbd0c07ba4954545dc207c473ddc173512f5a1d8cd860e47a883c9f5e04\"" Feb 13 20:14:25.851530 containerd[1714]: time="2025-02-13T20:14:25.851483531Z" level=info msg="CreateContainer within sandbox \"ce5ee2df7147962b661b453f9e3542995c552c7642d7f6bbe34faf862f4fdc22\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4cbb3ce4056c46ca6c843b4aecb7c3a5593f1f11bddc52f964133d62ce9ec907\"" Feb 13 20:14:25.854640 containerd[1714]: time="2025-02-13T20:14:25.854382592Z" level=info msg="StartContainer for \"4cbb3ce4056c46ca6c843b4aecb7c3a5593f1f11bddc52f964133d62ce9ec907\"" Feb 13 20:14:25.883503 systemd[1]: Started cri-containerd-62583dbd0c07ba4954545dc207c473ddc173512f5a1d8cd860e47a883c9f5e04.scope - libcontainer container 62583dbd0c07ba4954545dc207c473ddc173512f5a1d8cd860e47a883c9f5e04. Feb 13 20:14:25.899456 systemd[1]: Started cri-containerd-4cbb3ce4056c46ca6c843b4aecb7c3a5593f1f11bddc52f964133d62ce9ec907.scope - libcontainer container 4cbb3ce4056c46ca6c843b4aecb7c3a5593f1f11bddc52f964133d62ce9ec907. Feb 13 20:14:25.950596 containerd[1714]: time="2025-02-13T20:14:25.950543625Z" level=info msg="StartContainer for \"62583dbd0c07ba4954545dc207c473ddc173512f5a1d8cd860e47a883c9f5e04\" returns successfully" Feb 13 20:14:25.958850 containerd[1714]: time="2025-02-13T20:14:25.958698597Z" level=info msg="StartContainer for \"4cbb3ce4056c46ca6c843b4aecb7c3a5593f1f11bddc52f964133d62ce9ec907\" returns successfully" Feb 13 20:14:26.559280 kubelet[3264]: I0213 20:14:26.558643 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-wdn5v" podStartSLOduration=24.558617777 podStartE2EDuration="24.558617777s" podCreationTimestamp="2025-02-13 20:14:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:14:26.557069544 +0000 UTC m=+28.670492160" watchObservedRunningTime="2025-02-13 20:14:26.558617777 +0000 UTC m=+28.672040393" Feb 13 20:14:26.611290 kubelet[3264]: I0213 20:14:26.610423 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-hcqz5" podStartSLOduration=24.610388671 podStartE2EDuration="24.610388671s" podCreationTimestamp="2025-02-13 20:14:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:14:26.578949706 +0000 UTC m=+28.692372322" watchObservedRunningTime="2025-02-13 20:14:26.610388671 +0000 UTC m=+28.723811187" Feb 13 20:16:24.541788 systemd[1]: Started sshd@7-10.200.4.48:22-10.200.16.10:40414.service - OpenSSH per-connection server daemon (10.200.16.10:40414). Feb 13 20:16:25.151704 sshd[4651]: Accepted publickey for core from 10.200.16.10 port 40414 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:25.153360 sshd-session[4651]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:25.160207 systemd-logind[1685]: New session 10 of user core. Feb 13 20:16:25.169441 systemd[1]: Started session-10.scope - Session 10 of User core. Feb 13 20:16:25.659824 sshd[4653]: Connection closed by 10.200.16.10 port 40414 Feb 13 20:16:25.661562 sshd-session[4651]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:25.666392 systemd-logind[1685]: Session 10 logged out. Waiting for processes to exit. Feb 13 20:16:25.667481 systemd[1]: sshd@7-10.200.4.48:22-10.200.16.10:40414.service: Deactivated successfully. Feb 13 20:16:25.670038 systemd[1]: session-10.scope: Deactivated successfully. Feb 13 20:16:25.671223 systemd-logind[1685]: Removed session 10. Feb 13 20:16:30.769624 systemd[1]: Started sshd@8-10.200.4.48:22-10.200.16.10:42336.service - OpenSSH per-connection server daemon (10.200.16.10:42336). Feb 13 20:16:31.362444 sshd[4665]: Accepted publickey for core from 10.200.16.10 port 42336 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:31.363892 sshd-session[4665]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:31.368871 systemd-logind[1685]: New session 11 of user core. Feb 13 20:16:31.375411 systemd[1]: Started session-11.scope - Session 11 of User core. Feb 13 20:16:31.842548 sshd[4667]: Connection closed by 10.200.16.10 port 42336 Feb 13 20:16:31.844300 sshd-session[4665]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:31.847946 systemd[1]: sshd@8-10.200.4.48:22-10.200.16.10:42336.service: Deactivated successfully. Feb 13 20:16:31.851497 systemd[1]: session-11.scope: Deactivated successfully. Feb 13 20:16:31.852384 systemd-logind[1685]: Session 11 logged out. Waiting for processes to exit. Feb 13 20:16:31.853401 systemd-logind[1685]: Removed session 11. Feb 13 20:16:36.953671 systemd[1]: Started sshd@9-10.200.4.48:22-10.200.16.10:42344.service - OpenSSH per-connection server daemon (10.200.16.10:42344). Feb 13 20:16:37.546496 sshd[4681]: Accepted publickey for core from 10.200.16.10 port 42344 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:37.548098 sshd-session[4681]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:37.553396 systemd-logind[1685]: New session 12 of user core. Feb 13 20:16:37.562500 systemd[1]: Started session-12.scope - Session 12 of User core. Feb 13 20:16:38.047596 sshd[4683]: Connection closed by 10.200.16.10 port 42344 Feb 13 20:16:38.049532 sshd-session[4681]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:38.053554 systemd[1]: sshd@9-10.200.4.48:22-10.200.16.10:42344.service: Deactivated successfully. Feb 13 20:16:38.056129 systemd[1]: session-12.scope: Deactivated successfully. Feb 13 20:16:38.057630 systemd-logind[1685]: Session 12 logged out. Waiting for processes to exit. Feb 13 20:16:38.058650 systemd-logind[1685]: Removed session 12. Feb 13 20:16:43.154687 systemd[1]: Started sshd@10-10.200.4.48:22-10.200.16.10:47534.service - OpenSSH per-connection server daemon (10.200.16.10:47534). Feb 13 20:16:43.751075 sshd[4696]: Accepted publickey for core from 10.200.16.10 port 47534 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:43.751769 sshd-session[4696]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:43.756920 systemd-logind[1685]: New session 13 of user core. Feb 13 20:16:43.762433 systemd[1]: Started session-13.scope - Session 13 of User core. Feb 13 20:16:44.274067 sshd[4698]: Connection closed by 10.200.16.10 port 47534 Feb 13 20:16:44.274950 sshd-session[4696]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:44.280102 systemd[1]: sshd@10-10.200.4.48:22-10.200.16.10:47534.service: Deactivated successfully. Feb 13 20:16:44.283015 systemd[1]: session-13.scope: Deactivated successfully. Feb 13 20:16:44.283879 systemd-logind[1685]: Session 13 logged out. Waiting for processes to exit. Feb 13 20:16:44.284897 systemd-logind[1685]: Removed session 13. Feb 13 20:16:49.384026 systemd[1]: Started sshd@11-10.200.4.48:22-10.200.16.10:44244.service - OpenSSH per-connection server daemon (10.200.16.10:44244). Feb 13 20:16:49.980121 sshd[4710]: Accepted publickey for core from 10.200.16.10 port 44244 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:49.981679 sshd-session[4710]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:49.987337 systemd-logind[1685]: New session 14 of user core. Feb 13 20:16:49.990433 systemd[1]: Started session-14.scope - Session 14 of User core. Feb 13 20:16:50.462626 sshd[4712]: Connection closed by 10.200.16.10 port 44244 Feb 13 20:16:50.464531 sshd-session[4710]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:50.470523 systemd-logind[1685]: Session 14 logged out. Waiting for processes to exit. Feb 13 20:16:50.472001 systemd[1]: sshd@11-10.200.4.48:22-10.200.16.10:44244.service: Deactivated successfully. Feb 13 20:16:50.474764 systemd[1]: session-14.scope: Deactivated successfully. Feb 13 20:16:50.477099 systemd-logind[1685]: Removed session 14. Feb 13 20:16:50.578069 systemd[1]: Started sshd@12-10.200.4.48:22-10.200.16.10:44254.service - OpenSSH per-connection server daemon (10.200.16.10:44254). Feb 13 20:16:51.170163 sshd[4724]: Accepted publickey for core from 10.200.16.10 port 44254 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:51.170920 sshd-session[4724]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:51.176056 systemd-logind[1685]: New session 15 of user core. Feb 13 20:16:51.183514 systemd[1]: Started session-15.scope - Session 15 of User core. Feb 13 20:16:51.689315 sshd[4726]: Connection closed by 10.200.16.10 port 44254 Feb 13 20:16:51.690715 sshd-session[4724]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:51.694997 systemd[1]: sshd@12-10.200.4.48:22-10.200.16.10:44254.service: Deactivated successfully. Feb 13 20:16:51.699078 systemd[1]: session-15.scope: Deactivated successfully. Feb 13 20:16:51.700142 systemd-logind[1685]: Session 15 logged out. Waiting for processes to exit. Feb 13 20:16:51.701991 systemd-logind[1685]: Removed session 15. Feb 13 20:16:51.796958 systemd[1]: Started sshd@13-10.200.4.48:22-10.200.16.10:44258.service - OpenSSH per-connection server daemon (10.200.16.10:44258). Feb 13 20:16:52.383079 sshd[4735]: Accepted publickey for core from 10.200.16.10 port 44258 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:52.383756 sshd-session[4735]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:52.389212 systemd-logind[1685]: New session 16 of user core. Feb 13 20:16:52.391469 systemd[1]: Started session-16.scope - Session 16 of User core. Feb 13 20:16:52.889898 sshd[4737]: Connection closed by 10.200.16.10 port 44258 Feb 13 20:16:52.891498 sshd-session[4735]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:52.894193 systemd[1]: sshd@13-10.200.4.48:22-10.200.16.10:44258.service: Deactivated successfully. Feb 13 20:16:52.896564 systemd[1]: session-16.scope: Deactivated successfully. Feb 13 20:16:52.899619 systemd-logind[1685]: Session 16 logged out. Waiting for processes to exit. Feb 13 20:16:52.900796 systemd-logind[1685]: Removed session 16. Feb 13 20:16:57.996751 systemd[1]: Started sshd@14-10.200.4.48:22-10.200.16.10:44274.service - OpenSSH per-connection server daemon (10.200.16.10:44274). Feb 13 20:16:58.597875 sshd[4748]: Accepted publickey for core from 10.200.16.10 port 44274 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:58.599485 sshd-session[4748]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:58.604458 systemd-logind[1685]: New session 17 of user core. Feb 13 20:16:58.615545 systemd[1]: Started session-17.scope - Session 17 of User core. Feb 13 20:16:59.075737 sshd[4752]: Connection closed by 10.200.16.10 port 44274 Feb 13 20:16:59.077604 sshd-session[4748]: pam_unix(sshd:session): session closed for user core Feb 13 20:16:59.081895 systemd[1]: sshd@14-10.200.4.48:22-10.200.16.10:44274.service: Deactivated successfully. Feb 13 20:16:59.085217 systemd[1]: session-17.scope: Deactivated successfully. Feb 13 20:16:59.086355 systemd-logind[1685]: Session 17 logged out. Waiting for processes to exit. Feb 13 20:16:59.087487 systemd-logind[1685]: Removed session 17. Feb 13 20:16:59.186432 systemd[1]: Started sshd@15-10.200.4.48:22-10.200.16.10:58676.service - OpenSSH per-connection server daemon (10.200.16.10:58676). Feb 13 20:16:59.780300 sshd[4763]: Accepted publickey for core from 10.200.16.10 port 58676 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:16:59.782575 sshd-session[4763]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:16:59.788152 systemd-logind[1685]: New session 18 of user core. Feb 13 20:16:59.799617 systemd[1]: Started session-18.scope - Session 18 of User core. Feb 13 20:17:00.305080 sshd[4765]: Connection closed by 10.200.16.10 port 58676 Feb 13 20:17:00.306051 sshd-session[4763]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:00.309420 systemd[1]: sshd@15-10.200.4.48:22-10.200.16.10:58676.service: Deactivated successfully. Feb 13 20:17:00.311867 systemd[1]: session-18.scope: Deactivated successfully. Feb 13 20:17:00.313910 systemd-logind[1685]: Session 18 logged out. Waiting for processes to exit. Feb 13 20:17:00.315232 systemd-logind[1685]: Removed session 18. Feb 13 20:17:00.421682 systemd[1]: Started sshd@16-10.200.4.48:22-10.200.16.10:58678.service - OpenSSH per-connection server daemon (10.200.16.10:58678). Feb 13 20:17:01.010788 sshd[4774]: Accepted publickey for core from 10.200.16.10 port 58678 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:01.012359 sshd-session[4774]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:01.017870 systemd-logind[1685]: New session 19 of user core. Feb 13 20:17:01.026553 systemd[1]: Started session-19.scope - Session 19 of User core. Feb 13 20:17:02.404790 sshd[4776]: Connection closed by 10.200.16.10 port 58678 Feb 13 20:17:02.405733 sshd-session[4774]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:02.410422 systemd[1]: sshd@16-10.200.4.48:22-10.200.16.10:58678.service: Deactivated successfully. Feb 13 20:17:02.413886 systemd[1]: session-19.scope: Deactivated successfully. Feb 13 20:17:02.417157 systemd-logind[1685]: Session 19 logged out. Waiting for processes to exit. Feb 13 20:17:02.419058 systemd-logind[1685]: Removed session 19. Feb 13 20:17:02.513502 systemd[1]: Started sshd@17-10.200.4.48:22-10.200.16.10:58690.service - OpenSSH per-connection server daemon (10.200.16.10:58690). Feb 13 20:17:03.110814 sshd[4792]: Accepted publickey for core from 10.200.16.10 port 58690 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:03.111700 sshd-session[4792]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:03.117780 systemd-logind[1685]: New session 20 of user core. Feb 13 20:17:03.123432 systemd[1]: Started session-20.scope - Session 20 of User core. Feb 13 20:17:03.714809 sshd[4794]: Connection closed by 10.200.16.10 port 58690 Feb 13 20:17:03.716516 sshd-session[4792]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:03.720571 systemd[1]: sshd@17-10.200.4.48:22-10.200.16.10:58690.service: Deactivated successfully. Feb 13 20:17:03.722992 systemd[1]: session-20.scope: Deactivated successfully. Feb 13 20:17:03.723944 systemd-logind[1685]: Session 20 logged out. Waiting for processes to exit. Feb 13 20:17:03.725451 systemd-logind[1685]: Removed session 20. Feb 13 20:17:03.832672 systemd[1]: Started sshd@18-10.200.4.48:22-10.200.16.10:58700.service - OpenSSH per-connection server daemon (10.200.16.10:58700). Feb 13 20:17:04.418139 sshd[4804]: Accepted publickey for core from 10.200.16.10 port 58700 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:04.419947 sshd-session[4804]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:04.435173 systemd-logind[1685]: New session 21 of user core. Feb 13 20:17:04.443418 systemd[1]: Started session-21.scope - Session 21 of User core. Feb 13 20:17:04.918326 sshd[4806]: Connection closed by 10.200.16.10 port 58700 Feb 13 20:17:04.917621 sshd-session[4804]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:04.921910 systemd-logind[1685]: Session 21 logged out. Waiting for processes to exit. Feb 13 20:17:04.924831 systemd[1]: sshd@18-10.200.4.48:22-10.200.16.10:58700.service: Deactivated successfully. Feb 13 20:17:04.928855 systemd[1]: session-21.scope: Deactivated successfully. Feb 13 20:17:04.929797 systemd-logind[1685]: Removed session 21. Feb 13 20:17:10.025644 systemd[1]: Started sshd@19-10.200.4.48:22-10.200.16.10:52914.service - OpenSSH per-connection server daemon (10.200.16.10:52914). Feb 13 20:17:10.629147 sshd[4819]: Accepted publickey for core from 10.200.16.10 port 52914 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:10.630946 sshd-session[4819]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:10.637154 systemd-logind[1685]: New session 22 of user core. Feb 13 20:17:10.644458 systemd[1]: Started session-22.scope - Session 22 of User core. Feb 13 20:17:11.113891 sshd[4821]: Connection closed by 10.200.16.10 port 52914 Feb 13 20:17:11.114599 sshd-session[4819]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:11.119571 systemd[1]: sshd@19-10.200.4.48:22-10.200.16.10:52914.service: Deactivated successfully. Feb 13 20:17:11.122138 systemd[1]: session-22.scope: Deactivated successfully. Feb 13 20:17:11.123479 systemd-logind[1685]: Session 22 logged out. Waiting for processes to exit. Feb 13 20:17:11.124563 systemd-logind[1685]: Removed session 22. Feb 13 20:17:16.232585 systemd[1]: Started sshd@20-10.200.4.48:22-10.200.16.10:52918.service - OpenSSH per-connection server daemon (10.200.16.10:52918). Feb 13 20:17:16.834208 sshd[4832]: Accepted publickey for core from 10.200.16.10 port 52918 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:16.836006 sshd-session[4832]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:16.842514 systemd-logind[1685]: New session 23 of user core. Feb 13 20:17:16.846484 systemd[1]: Started session-23.scope - Session 23 of User core. Feb 13 20:17:17.321117 sshd[4834]: Connection closed by 10.200.16.10 port 52918 Feb 13 20:17:17.322413 sshd-session[4832]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:17.325405 systemd[1]: sshd@20-10.200.4.48:22-10.200.16.10:52918.service: Deactivated successfully. Feb 13 20:17:17.327862 systemd[1]: session-23.scope: Deactivated successfully. Feb 13 20:17:17.330178 systemd-logind[1685]: Session 23 logged out. Waiting for processes to exit. Feb 13 20:17:17.331317 systemd-logind[1685]: Removed session 23. Feb 13 20:17:22.431572 systemd[1]: Started sshd@21-10.200.4.48:22-10.200.16.10:34636.service - OpenSSH per-connection server daemon (10.200.16.10:34636). Feb 13 20:17:23.033298 sshd[4846]: Accepted publickey for core from 10.200.16.10 port 34636 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:23.035215 sshd-session[4846]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:23.040718 systemd-logind[1685]: New session 24 of user core. Feb 13 20:17:23.048499 systemd[1]: Started session-24.scope - Session 24 of User core. Feb 13 20:17:23.527494 sshd[4848]: Connection closed by 10.200.16.10 port 34636 Feb 13 20:17:23.528315 sshd-session[4846]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:23.532204 systemd[1]: sshd@21-10.200.4.48:22-10.200.16.10:34636.service: Deactivated successfully. Feb 13 20:17:23.534570 systemd[1]: session-24.scope: Deactivated successfully. Feb 13 20:17:23.535475 systemd-logind[1685]: Session 24 logged out. Waiting for processes to exit. Feb 13 20:17:23.536831 systemd-logind[1685]: Removed session 24. Feb 13 20:17:23.636572 systemd[1]: Started sshd@22-10.200.4.48:22-10.200.16.10:34648.service - OpenSSH per-connection server daemon (10.200.16.10:34648). Feb 13 20:17:24.232276 sshd[4859]: Accepted publickey for core from 10.200.16.10 port 34648 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:24.234099 sshd-session[4859]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:24.239617 systemd-logind[1685]: New session 25 of user core. Feb 13 20:17:24.243465 systemd[1]: Started session-25.scope - Session 25 of User core. Feb 13 20:17:25.866542 containerd[1714]: time="2025-02-13T20:17:25.866476470Z" level=info msg="StopContainer for \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\" with timeout 30 (s)" Feb 13 20:17:25.873094 containerd[1714]: time="2025-02-13T20:17:25.872513797Z" level=info msg="Stop container \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\" with signal terminated" Feb 13 20:17:25.881290 systemd[1]: run-containerd-runc-k8s.io-d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada-runc.y9E301.mount: Deactivated successfully. Feb 13 20:17:25.898556 containerd[1714]: time="2025-02-13T20:17:25.898490244Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 20:17:25.901759 systemd[1]: cri-containerd-59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650.scope: Deactivated successfully. Feb 13 20:17:25.915955 containerd[1714]: time="2025-02-13T20:17:25.915901711Z" level=info msg="StopContainer for \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\" with timeout 2 (s)" Feb 13 20:17:25.917270 containerd[1714]: time="2025-02-13T20:17:25.916671228Z" level=info msg="Stop container \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\" with signal terminated" Feb 13 20:17:25.933365 systemd-networkd[1325]: lxc_health: Link DOWN Feb 13 20:17:25.933376 systemd-networkd[1325]: lxc_health: Lost carrier Feb 13 20:17:25.953633 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650-rootfs.mount: Deactivated successfully. Feb 13 20:17:25.955582 systemd[1]: cri-containerd-d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada.scope: Deactivated successfully. Feb 13 20:17:25.956513 systemd[1]: cri-containerd-d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada.scope: Consumed 7.548s CPU time. Feb 13 20:17:25.993151 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada-rootfs.mount: Deactivated successfully. Feb 13 20:17:26.040813 containerd[1714]: time="2025-02-13T20:17:26.040724442Z" level=info msg="shim disconnected" id=d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada namespace=k8s.io Feb 13 20:17:26.040813 containerd[1714]: time="2025-02-13T20:17:26.040821244Z" level=warning msg="cleaning up after shim disconnected" id=d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada namespace=k8s.io Feb 13 20:17:26.040813 containerd[1714]: time="2025-02-13T20:17:26.040834044Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:26.041808 containerd[1714]: time="2025-02-13T20:17:26.041399756Z" level=info msg="shim disconnected" id=59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650 namespace=k8s.io Feb 13 20:17:26.041808 containerd[1714]: time="2025-02-13T20:17:26.041451157Z" level=warning msg="cleaning up after shim disconnected" id=59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650 namespace=k8s.io Feb 13 20:17:26.041808 containerd[1714]: time="2025-02-13T20:17:26.041462358Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:26.066024 containerd[1714]: time="2025-02-13T20:17:26.065975674Z" level=info msg="StopContainer for \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\" returns successfully" Feb 13 20:17:26.066842 containerd[1714]: time="2025-02-13T20:17:26.066803992Z" level=info msg="StopPodSandbox for \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\"" Feb 13 20:17:26.066966 containerd[1714]: time="2025-02-13T20:17:26.066855993Z" level=info msg="Container to stop \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 13 20:17:26.066966 containerd[1714]: time="2025-02-13T20:17:26.066903894Z" level=info msg="Container to stop \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 13 20:17:26.066966 containerd[1714]: time="2025-02-13T20:17:26.066917094Z" level=info msg="Container to stop \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 13 20:17:26.066966 containerd[1714]: time="2025-02-13T20:17:26.066929094Z" level=info msg="Container to stop \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 13 20:17:26.066966 containerd[1714]: time="2025-02-13T20:17:26.066940795Z" level=info msg="Container to stop \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 13 20:17:26.070944 containerd[1714]: time="2025-02-13T20:17:26.069685952Z" level=info msg="StopContainer for \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\" returns successfully" Feb 13 20:17:26.070015 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170-shm.mount: Deactivated successfully. Feb 13 20:17:26.072182 containerd[1714]: time="2025-02-13T20:17:26.071768996Z" level=info msg="StopPodSandbox for \"77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5\"" Feb 13 20:17:26.072182 containerd[1714]: time="2025-02-13T20:17:26.071937600Z" level=info msg="Container to stop \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 13 20:17:26.081127 systemd[1]: cri-containerd-4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170.scope: Deactivated successfully. Feb 13 20:17:26.094499 systemd[1]: cri-containerd-77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5.scope: Deactivated successfully. Feb 13 20:17:26.127399 containerd[1714]: time="2025-02-13T20:17:26.125962938Z" level=info msg="shim disconnected" id=4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170 namespace=k8s.io Feb 13 20:17:26.127399 containerd[1714]: time="2025-02-13T20:17:26.126029440Z" level=warning msg="cleaning up after shim disconnected" id=4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170 namespace=k8s.io Feb 13 20:17:26.127399 containerd[1714]: time="2025-02-13T20:17:26.126040240Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:26.131538 containerd[1714]: time="2025-02-13T20:17:26.131188949Z" level=info msg="shim disconnected" id=77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5 namespace=k8s.io Feb 13 20:17:26.131538 containerd[1714]: time="2025-02-13T20:17:26.131344452Z" level=warning msg="cleaning up after shim disconnected" id=77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5 namespace=k8s.io Feb 13 20:17:26.131538 containerd[1714]: time="2025-02-13T20:17:26.131467254Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:26.149614 containerd[1714]: time="2025-02-13T20:17:26.149150827Z" level=warning msg="cleanup warnings time=\"2025-02-13T20:17:26Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Feb 13 20:17:26.152695 containerd[1714]: time="2025-02-13T20:17:26.151996787Z" level=info msg="TearDown network for sandbox \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" successfully" Feb 13 20:17:26.152695 containerd[1714]: time="2025-02-13T20:17:26.152041688Z" level=info msg="StopPodSandbox for \"4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170\" returns successfully" Feb 13 20:17:26.156300 containerd[1714]: time="2025-02-13T20:17:26.155560762Z" level=warning msg="cleanup warnings time=\"2025-02-13T20:17:26Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Feb 13 20:17:26.157206 containerd[1714]: time="2025-02-13T20:17:26.157173396Z" level=info msg="TearDown network for sandbox \"77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5\" successfully" Feb 13 20:17:26.157206 containerd[1714]: time="2025-02-13T20:17:26.157206097Z" level=info msg="StopPodSandbox for \"77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5\" returns successfully" Feb 13 20:17:26.202927 kubelet[3264]: I0213 20:17:26.202467 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6v8f\" (UniqueName: \"kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-kube-api-access-s6v8f\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.204084 kubelet[3264]: I0213 20:17:26.204045 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-kernel\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.204236 kubelet[3264]: I0213 20:17:26.204089 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hubble-tls\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.204236 kubelet[3264]: I0213 20:17:26.204118 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-cilium-config-path\") pod \"6ba2f865-bcae-4078-92fd-ff4c68c1d9f1\" (UID: \"6ba2f865-bcae-4078-92fd-ff4c68c1d9f1\") " Feb 13 20:17:26.204236 kubelet[3264]: I0213 20:17:26.204144 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-clustermesh-secrets\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.204236 kubelet[3264]: I0213 20:17:26.204168 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-net\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.204236 kubelet[3264]: I0213 20:17:26.204188 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-xtables-lock\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.204236 kubelet[3264]: I0213 20:17:26.204208 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-bpf-maps\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.204526 kubelet[3264]: I0213 20:17:26.204225 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hostproc\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.207730 kubelet[3264]: I0213 20:17:26.204250 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cni-path\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.207867 kubelet[3264]: I0213 20:17:26.207764 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-lib-modules\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.207867 kubelet[3264]: I0213 20:17:26.207790 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-etc-cni-netd\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.207867 kubelet[3264]: I0213 20:17:26.207820 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t92sm\" (UniqueName: \"kubernetes.io/projected/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-kube-api-access-t92sm\") pod \"6ba2f865-bcae-4078-92fd-ff4c68c1d9f1\" (UID: \"6ba2f865-bcae-4078-92fd-ff4c68c1d9f1\") " Feb 13 20:17:26.207867 kubelet[3264]: I0213 20:17:26.207850 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-config-path\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.208024 kubelet[3264]: I0213 20:17:26.207869 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-run\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.208024 kubelet[3264]: I0213 20:17:26.207893 3264 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-cgroup\") pod \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\" (UID: \"b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3\") " Feb 13 20:17:26.208024 kubelet[3264]: I0213 20:17:26.207987 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.208142 kubelet[3264]: I0213 20:17:26.208032 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.210285 kubelet[3264]: I0213 20:17:26.208338 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hostproc" (OuterVolumeSpecName: "hostproc") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.212262 kubelet[3264]: I0213 20:17:26.211809 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cni-path" (OuterVolumeSpecName: "cni-path") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.212262 kubelet[3264]: I0213 20:17:26.211860 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.212262 kubelet[3264]: I0213 20:17:26.211886 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.213560 kubelet[3264]: I0213 20:17:26.213528 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.213649 kubelet[3264]: I0213 20:17:26.213584 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.213649 kubelet[3264]: I0213 20:17:26.213607 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.213894 kubelet[3264]: I0213 20:17:26.213684 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-kube-api-access-s6v8f" (OuterVolumeSpecName: "kube-api-access-s6v8f") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "kube-api-access-s6v8f". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 13 20:17:26.213894 kubelet[3264]: I0213 20:17:26.213827 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "6ba2f865-bcae-4078-92fd-ff4c68c1d9f1" (UID: "6ba2f865-bcae-4078-92fd-ff4c68c1d9f1"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 13 20:17:26.213894 kubelet[3264]: I0213 20:17:26.213869 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Feb 13 20:17:26.217407 kubelet[3264]: I0213 20:17:26.217346 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 13 20:17:26.218322 kubelet[3264]: I0213 20:17:26.217545 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 13 20:17:26.224241 kubelet[3264]: I0213 20:17:26.224175 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-kube-api-access-t92sm" (OuterVolumeSpecName: "kube-api-access-t92sm") pod "6ba2f865-bcae-4078-92fd-ff4c68c1d9f1" (UID: "6ba2f865-bcae-4078-92fd-ff4c68c1d9f1"). InnerVolumeSpecName "kube-api-access-t92sm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 13 20:17:26.224892 kubelet[3264]: I0213 20:17:26.224861 3264 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" (UID: "b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 13 20:17:26.308749 kubelet[3264]: I0213 20:17:26.308689 3264 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-clustermesh-secrets\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.308749 kubelet[3264]: I0213 20:17:26.308738 3264 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-s6v8f\" (UniqueName: \"kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-kube-api-access-s6v8f\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.308749 kubelet[3264]: I0213 20:17:26.308755 3264 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-kernel\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308774 3264 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hubble-tls\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308789 3264 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-cilium-config-path\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308808 3264 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-bpf-maps\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308824 3264 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-host-proc-sys-net\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308839 3264 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-xtables-lock\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308852 3264 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cni-path\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308864 3264 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-hostproc\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309060 kubelet[3264]: I0213 20:17:26.308876 3264 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-config-path\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309352 kubelet[3264]: I0213 20:17:26.308889 3264 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-lib-modules\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309352 kubelet[3264]: I0213 20:17:26.308909 3264 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-etc-cni-netd\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309352 kubelet[3264]: I0213 20:17:26.308925 3264 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-t92sm\" (UniqueName: \"kubernetes.io/projected/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1-kube-api-access-t92sm\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309352 kubelet[3264]: I0213 20:17:26.308938 3264 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-run\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.309352 kubelet[3264]: I0213 20:17:26.308952 3264 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3-cilium-cgroup\") on node \"ci-4152.2.1-a-2754074fca\" DevicePath \"\"" Feb 13 20:17:26.431057 systemd[1]: Removed slice kubepods-burstable-podb0f8857d_4ed0_46a7_92d5_9c4a58d29dd3.slice - libcontainer container kubepods-burstable-podb0f8857d_4ed0_46a7_92d5_9c4a58d29dd3.slice. Feb 13 20:17:26.431398 systemd[1]: kubepods-burstable-podb0f8857d_4ed0_46a7_92d5_9c4a58d29dd3.slice: Consumed 7.653s CPU time. Feb 13 20:17:26.433333 systemd[1]: Removed slice kubepods-besteffort-pod6ba2f865_bcae_4078_92fd_ff4c68c1d9f1.slice - libcontainer container kubepods-besteffort-pod6ba2f865_bcae_4078_92fd_ff4c68c1d9f1.slice. Feb 13 20:17:26.872367 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5-rootfs.mount: Deactivated successfully. Feb 13 20:17:26.872512 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-77ad46289d28f9ee8d82050256fe570858be42f3e62630fc331fa1af8652c9c5-shm.mount: Deactivated successfully. Feb 13 20:17:26.872646 systemd[1]: var-lib-kubelet-pods-6ba2f865\x2dbcae\x2d4078\x2d92fd\x2dff4c68c1d9f1-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dt92sm.mount: Deactivated successfully. Feb 13 20:17:26.872752 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4da8b94f8e90bce97fb15a817c9283bd5224d1fc99d571ecd267add247427170-rootfs.mount: Deactivated successfully. Feb 13 20:17:26.872862 systemd[1]: var-lib-kubelet-pods-b0f8857d\x2d4ed0\x2d46a7\x2d92d5\x2d9c4a58d29dd3-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2ds6v8f.mount: Deactivated successfully. Feb 13 20:17:26.872967 systemd[1]: var-lib-kubelet-pods-b0f8857d\x2d4ed0\x2d46a7\x2d92d5\x2d9c4a58d29dd3-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Feb 13 20:17:26.873069 systemd[1]: var-lib-kubelet-pods-b0f8857d\x2d4ed0\x2d46a7\x2d92d5\x2d9c4a58d29dd3-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Feb 13 20:17:26.953672 kubelet[3264]: I0213 20:17:26.953074 3264 scope.go:117] "RemoveContainer" containerID="59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650" Feb 13 20:17:26.957671 containerd[1714]: time="2025-02-13T20:17:26.956737347Z" level=info msg="RemoveContainer for \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\"" Feb 13 20:17:26.970666 containerd[1714]: time="2025-02-13T20:17:26.970604539Z" level=info msg="RemoveContainer for \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\" returns successfully" Feb 13 20:17:26.974780 kubelet[3264]: I0213 20:17:26.974755 3264 scope.go:117] "RemoveContainer" containerID="59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650" Feb 13 20:17:26.978669 containerd[1714]: time="2025-02-13T20:17:26.978605508Z" level=error msg="ContainerStatus for \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\": not found" Feb 13 20:17:26.979365 kubelet[3264]: E0213 20:17:26.979328 3264 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\": not found" containerID="59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650" Feb 13 20:17:26.979717 kubelet[3264]: I0213 20:17:26.979508 3264 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650"} err="failed to get container status \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\": rpc error: code = NotFound desc = an error occurred when try to find container \"59a023e47c29e71ded7e83068966d4ce9fcf5edf5d8029ee5bdd737dba58d650\": not found" Feb 13 20:17:26.979717 kubelet[3264]: I0213 20:17:26.979613 3264 scope.go:117] "RemoveContainer" containerID="d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada" Feb 13 20:17:26.985137 containerd[1714]: time="2025-02-13T20:17:26.985087344Z" level=info msg="RemoveContainer for \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\"" Feb 13 20:17:26.992737 containerd[1714]: time="2025-02-13T20:17:26.992643303Z" level=info msg="RemoveContainer for \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\" returns successfully" Feb 13 20:17:26.992838 kubelet[3264]: I0213 20:17:26.992818 3264 scope.go:117] "RemoveContainer" containerID="24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e" Feb 13 20:17:26.998039 containerd[1714]: time="2025-02-13T20:17:26.997735111Z" level=info msg="RemoveContainer for \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\"" Feb 13 20:17:27.004732 containerd[1714]: time="2025-02-13T20:17:27.004701158Z" level=info msg="RemoveContainer for \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\" returns successfully" Feb 13 20:17:27.004920 kubelet[3264]: I0213 20:17:27.004889 3264 scope.go:117] "RemoveContainer" containerID="ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309" Feb 13 20:17:27.005795 containerd[1714]: time="2025-02-13T20:17:27.005770880Z" level=info msg="RemoveContainer for \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\"" Feb 13 20:17:27.017812 containerd[1714]: time="2025-02-13T20:17:27.017778033Z" level=info msg="RemoveContainer for \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\" returns successfully" Feb 13 20:17:27.017974 kubelet[3264]: I0213 20:17:27.017943 3264 scope.go:117] "RemoveContainer" containerID="21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610" Feb 13 20:17:27.018907 containerd[1714]: time="2025-02-13T20:17:27.018869256Z" level=info msg="RemoveContainer for \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\"" Feb 13 20:17:27.038973 containerd[1714]: time="2025-02-13T20:17:27.038936279Z" level=info msg="RemoveContainer for \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\" returns successfully" Feb 13 20:17:27.039181 kubelet[3264]: I0213 20:17:27.039156 3264 scope.go:117] "RemoveContainer" containerID="4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33" Feb 13 20:17:27.040247 containerd[1714]: time="2025-02-13T20:17:27.040214806Z" level=info msg="RemoveContainer for \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\"" Feb 13 20:17:27.062845 containerd[1714]: time="2025-02-13T20:17:27.062808982Z" level=info msg="RemoveContainer for \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\" returns successfully" Feb 13 20:17:27.063095 kubelet[3264]: I0213 20:17:27.063068 3264 scope.go:117] "RemoveContainer" containerID="d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada" Feb 13 20:17:27.063379 containerd[1714]: time="2025-02-13T20:17:27.063347094Z" level=error msg="ContainerStatus for \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\": not found" Feb 13 20:17:27.063514 kubelet[3264]: E0213 20:17:27.063486 3264 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\": not found" containerID="d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada" Feb 13 20:17:27.063600 kubelet[3264]: I0213 20:17:27.063525 3264 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada"} err="failed to get container status \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\": rpc error: code = NotFound desc = an error occurred when try to find container \"d1ced0215d0fcac4cb223300d28758823126dfb649796e0106075e4a5351aada\": not found" Feb 13 20:17:27.063600 kubelet[3264]: I0213 20:17:27.063550 3264 scope.go:117] "RemoveContainer" containerID="24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e" Feb 13 20:17:27.063765 containerd[1714]: time="2025-02-13T20:17:27.063731502Z" level=error msg="ContainerStatus for \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\": not found" Feb 13 20:17:27.063952 kubelet[3264]: E0213 20:17:27.063876 3264 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\": not found" containerID="24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e" Feb 13 20:17:27.063952 kubelet[3264]: I0213 20:17:27.063906 3264 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e"} err="failed to get container status \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\": rpc error: code = NotFound desc = an error occurred when try to find container \"24f834acfbd3bb573b18602a330aaf09864a315920f9a35363809298265d302e\": not found" Feb 13 20:17:27.063952 kubelet[3264]: I0213 20:17:27.063926 3264 scope.go:117] "RemoveContainer" containerID="ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309" Feb 13 20:17:27.064345 containerd[1714]: time="2025-02-13T20:17:27.064315714Z" level=error msg="ContainerStatus for \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\": not found" Feb 13 20:17:27.064464 kubelet[3264]: E0213 20:17:27.064439 3264 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\": not found" containerID="ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309" Feb 13 20:17:27.064548 kubelet[3264]: I0213 20:17:27.064469 3264 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309"} err="failed to get container status \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\": rpc error: code = NotFound desc = an error occurred when try to find container \"ce33fa3cc357ec908a5098ae535ec00968f11b8342099cb98fd3a1e310961309\": not found" Feb 13 20:17:27.064548 kubelet[3264]: I0213 20:17:27.064490 3264 scope.go:117] "RemoveContainer" containerID="21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610" Feb 13 20:17:27.064680 containerd[1714]: time="2025-02-13T20:17:27.064651221Z" level=error msg="ContainerStatus for \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\": not found" Feb 13 20:17:27.064786 kubelet[3264]: E0213 20:17:27.064762 3264 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\": not found" containerID="21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610" Feb 13 20:17:27.064870 kubelet[3264]: I0213 20:17:27.064791 3264 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610"} err="failed to get container status \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\": rpc error: code = NotFound desc = an error occurred when try to find container \"21a30d091bc819ae66c0e7764562f5ff5974d9c06d2b5923781c179addc5a610\": not found" Feb 13 20:17:27.064870 kubelet[3264]: I0213 20:17:27.064808 3264 scope.go:117] "RemoveContainer" containerID="4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33" Feb 13 20:17:27.065042 containerd[1714]: time="2025-02-13T20:17:27.065010429Z" level=error msg="ContainerStatus for \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\": not found" Feb 13 20:17:27.065245 kubelet[3264]: E0213 20:17:27.065205 3264 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\": not found" containerID="4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33" Feb 13 20:17:27.065327 kubelet[3264]: I0213 20:17:27.065265 3264 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33"} err="failed to get container status \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\": rpc error: code = NotFound desc = an error occurred when try to find container \"4ae66300b897c7715b2cc29aa30478b1c429a41c871534cb90af2ae84d723a33\": not found" Feb 13 20:17:27.905845 sshd[4861]: Connection closed by 10.200.16.10 port 34648 Feb 13 20:17:27.907493 sshd-session[4859]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:27.911502 systemd[1]: sshd@22-10.200.4.48:22-10.200.16.10:34648.service: Deactivated successfully. Feb 13 20:17:27.919577 systemd[1]: session-25.scope: Deactivated successfully. Feb 13 20:17:27.922413 systemd-logind[1685]: Session 25 logged out. Waiting for processes to exit. Feb 13 20:17:27.925091 systemd-logind[1685]: Removed session 25. Feb 13 20:17:28.019624 systemd[1]: Started sshd@23-10.200.4.48:22-10.200.16.10:34662.service - OpenSSH per-connection server daemon (10.200.16.10:34662). Feb 13 20:17:28.425497 kubelet[3264]: I0213 20:17:28.425451 3264 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba2f865-bcae-4078-92fd-ff4c68c1d9f1" path="/var/lib/kubelet/pods/6ba2f865-bcae-4078-92fd-ff4c68c1d9f1/volumes" Feb 13 20:17:28.426020 kubelet[3264]: I0213 20:17:28.425997 3264 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" path="/var/lib/kubelet/pods/b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3/volumes" Feb 13 20:17:28.580802 kubelet[3264]: E0213 20:17:28.580729 3264 kubelet.go:3008] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Feb 13 20:17:28.605353 sshd[5027]: Accepted publickey for core from 10.200.16.10 port 34662 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:28.606866 sshd-session[5027]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:28.615994 systemd-logind[1685]: New session 26 of user core. Feb 13 20:17:28.623426 systemd[1]: Started session-26.scope - Session 26 of User core. Feb 13 20:17:29.393132 kubelet[3264]: I0213 20:17:29.392508 3264 memory_manager.go:355] "RemoveStaleState removing state" podUID="b0f8857d-4ed0-46a7-92d5-9c4a58d29dd3" containerName="cilium-agent" Feb 13 20:17:29.393132 kubelet[3264]: I0213 20:17:29.392540 3264 memory_manager.go:355] "RemoveStaleState removing state" podUID="6ba2f865-bcae-4078-92fd-ff4c68c1d9f1" containerName="cilium-operator" Feb 13 20:17:29.405745 systemd[1]: Created slice kubepods-burstable-podb712e214_d881_4b7e_89cf_88dd39cc7593.slice - libcontainer container kubepods-burstable-podb712e214_d881_4b7e_89cf_88dd39cc7593.slice. Feb 13 20:17:29.428275 kubelet[3264]: I0213 20:17:29.427023 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/b712e214-d881-4b7e-89cf-88dd39cc7593-hubble-tls\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429195 kubelet[3264]: I0213 20:17:29.428753 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cmdq\" (UniqueName: \"kubernetes.io/projected/b712e214-d881-4b7e-89cf-88dd39cc7593-kube-api-access-6cmdq\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429195 kubelet[3264]: I0213 20:17:29.428794 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-etc-cni-netd\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429195 kubelet[3264]: I0213 20:17:29.428816 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-bpf-maps\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429195 kubelet[3264]: I0213 20:17:29.428836 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-xtables-lock\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429195 kubelet[3264]: I0213 20:17:29.428862 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-hostproc\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429195 kubelet[3264]: I0213 20:17:29.428882 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-cni-path\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429573 kubelet[3264]: I0213 20:17:29.428907 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/b712e214-d881-4b7e-89cf-88dd39cc7593-clustermesh-secrets\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429573 kubelet[3264]: I0213 20:17:29.428943 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/b712e214-d881-4b7e-89cf-88dd39cc7593-cilium-ipsec-secrets\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429573 kubelet[3264]: I0213 20:17:29.428968 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-host-proc-sys-net\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429573 kubelet[3264]: I0213 20:17:29.428993 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-host-proc-sys-kernel\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429573 kubelet[3264]: I0213 20:17:29.429015 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-cilium-cgroup\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429795 kubelet[3264]: I0213 20:17:29.429037 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-lib-modules\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429795 kubelet[3264]: I0213 20:17:29.429058 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/b712e214-d881-4b7e-89cf-88dd39cc7593-cilium-run\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.429795 kubelet[3264]: I0213 20:17:29.429081 3264 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/b712e214-d881-4b7e-89cf-88dd39cc7593-cilium-config-path\") pod \"cilium-2x9zx\" (UID: \"b712e214-d881-4b7e-89cf-88dd39cc7593\") " pod="kube-system/cilium-2x9zx" Feb 13 20:17:29.468416 sshd[5031]: Connection closed by 10.200.16.10 port 34662 Feb 13 20:17:29.469208 sshd-session[5027]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:29.474915 systemd[1]: sshd@23-10.200.4.48:22-10.200.16.10:34662.service: Deactivated successfully. Feb 13 20:17:29.480247 systemd[1]: session-26.scope: Deactivated successfully. Feb 13 20:17:29.483181 systemd-logind[1685]: Session 26 logged out. Waiting for processes to exit. Feb 13 20:17:29.485740 systemd-logind[1685]: Removed session 26. Feb 13 20:17:29.576424 systemd[1]: Started sshd@24-10.200.4.48:22-10.200.16.10:52684.service - OpenSSH per-connection server daemon (10.200.16.10:52684). Feb 13 20:17:29.712567 containerd[1714]: time="2025-02-13T20:17:29.712424522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-2x9zx,Uid:b712e214-d881-4b7e-89cf-88dd39cc7593,Namespace:kube-system,Attempt:0,}" Feb 13 20:17:29.753965 containerd[1714]: time="2025-02-13T20:17:29.753870896Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:17:29.753965 containerd[1714]: time="2025-02-13T20:17:29.753975298Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:17:29.754250 containerd[1714]: time="2025-02-13T20:17:29.754006698Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:17:29.754250 containerd[1714]: time="2025-02-13T20:17:29.754097200Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:17:29.773465 systemd[1]: Started cri-containerd-1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72.scope - libcontainer container 1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72. Feb 13 20:17:29.803150 containerd[1714]: time="2025-02-13T20:17:29.802791427Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-2x9zx,Uid:b712e214-d881-4b7e-89cf-88dd39cc7593,Namespace:kube-system,Attempt:0,} returns sandbox id \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\"" Feb 13 20:17:29.806643 containerd[1714]: time="2025-02-13T20:17:29.806610407Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Feb 13 20:17:29.844150 containerd[1714]: time="2025-02-13T20:17:29.844101897Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594\"" Feb 13 20:17:29.845175 containerd[1714]: time="2025-02-13T20:17:29.844762311Z" level=info msg="StartContainer for \"5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594\"" Feb 13 20:17:29.874443 systemd[1]: Started cri-containerd-5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594.scope - libcontainer container 5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594. Feb 13 20:17:29.908248 containerd[1714]: time="2025-02-13T20:17:29.907412231Z" level=info msg="StartContainer for \"5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594\" returns successfully" Feb 13 20:17:29.916136 systemd[1]: cri-containerd-5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594.scope: Deactivated successfully. Feb 13 20:17:30.003386 containerd[1714]: time="2025-02-13T20:17:30.003225151Z" level=info msg="shim disconnected" id=5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594 namespace=k8s.io Feb 13 20:17:30.003386 containerd[1714]: time="2025-02-13T20:17:30.003292652Z" level=warning msg="cleaning up after shim disconnected" id=5365a1ab79d2cab3b2f9160da043fa7f47735e94ea70976c479b50611c6f0594 namespace=k8s.io Feb 13 20:17:30.003386 containerd[1714]: time="2025-02-13T20:17:30.003303752Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:30.016868 containerd[1714]: time="2025-02-13T20:17:30.016819037Z" level=warning msg="cleanup warnings time=\"2025-02-13T20:17:30Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Feb 13 20:17:30.171935 sshd[5044]: Accepted publickey for core from 10.200.16.10 port 52684 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:30.173666 sshd-session[5044]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:30.178843 systemd-logind[1685]: New session 27 of user core. Feb 13 20:17:30.188432 systemd[1]: Started session-27.scope - Session 27 of User core. Feb 13 20:17:30.585930 sshd[5149]: Connection closed by 10.200.16.10 port 52684 Feb 13 20:17:30.586760 sshd-session[5044]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:30.590817 systemd-logind[1685]: Session 27 logged out. Waiting for processes to exit. Feb 13 20:17:30.591410 systemd[1]: sshd@24-10.200.4.48:22-10.200.16.10:52684.service: Deactivated successfully. Feb 13 20:17:30.593869 systemd[1]: session-27.scope: Deactivated successfully. Feb 13 20:17:30.594822 systemd-logind[1685]: Removed session 27. Feb 13 20:17:30.690694 systemd[1]: Started sshd@25-10.200.4.48:22-10.200.16.10:52694.service - OpenSSH per-connection server daemon (10.200.16.10:52694). Feb 13 20:17:30.980169 containerd[1714]: time="2025-02-13T20:17:30.979794532Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Feb 13 20:17:31.011418 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1973296982.mount: Deactivated successfully. Feb 13 20:17:31.020732 containerd[1714]: time="2025-02-13T20:17:31.020682993Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e\"" Feb 13 20:17:31.021913 containerd[1714]: time="2025-02-13T20:17:31.021647114Z" level=info msg="StartContainer for \"11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e\"" Feb 13 20:17:31.072471 systemd[1]: Started cri-containerd-11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e.scope - libcontainer container 11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e. Feb 13 20:17:31.111859 containerd[1714]: time="2025-02-13T20:17:31.111589909Z" level=info msg="StartContainer for \"11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e\" returns successfully" Feb 13 20:17:31.114865 systemd[1]: cri-containerd-11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e.scope: Deactivated successfully. Feb 13 20:17:31.152497 containerd[1714]: time="2025-02-13T20:17:31.152354468Z" level=info msg="shim disconnected" id=11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e namespace=k8s.io Feb 13 20:17:31.152497 containerd[1714]: time="2025-02-13T20:17:31.152495971Z" level=warning msg="cleaning up after shim disconnected" id=11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e namespace=k8s.io Feb 13 20:17:31.152497 containerd[1714]: time="2025-02-13T20:17:31.152509872Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:31.285794 sshd[5155]: Accepted publickey for core from 10.200.16.10 port 52694 ssh2: RSA SHA256:Nrus+ZqhYVn5ywEyfPWwt1mSwJTlPgyWKWwtj0nVtyc Feb 13 20:17:31.287522 sshd-session[5155]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:17:31.292374 systemd-logind[1685]: New session 28 of user core. Feb 13 20:17:31.301455 systemd[1]: Started session-28.scope - Session 28 of User core. Feb 13 20:17:31.537812 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-11378f9864199e4ebf1d0712e5cbb83de9c32047f2fca1a4ae41df47576c031e-rootfs.mount: Deactivated successfully. Feb 13 20:17:31.983690 containerd[1714]: time="2025-02-13T20:17:31.983372482Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Feb 13 20:17:32.026492 containerd[1714]: time="2025-02-13T20:17:32.026435889Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57\"" Feb 13 20:17:32.028606 containerd[1714]: time="2025-02-13T20:17:32.027270207Z" level=info msg="StartContainer for \"2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57\"" Feb 13 20:17:32.082480 systemd[1]: Started cri-containerd-2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57.scope - libcontainer container 2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57. Feb 13 20:17:32.128114 systemd[1]: cri-containerd-2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57.scope: Deactivated successfully. Feb 13 20:17:32.129760 containerd[1714]: time="2025-02-13T20:17:32.129718766Z" level=info msg="StartContainer for \"2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57\" returns successfully" Feb 13 20:17:32.176833 containerd[1714]: time="2025-02-13T20:17:32.176727857Z" level=info msg="shim disconnected" id=2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57 namespace=k8s.io Feb 13 20:17:32.176833 containerd[1714]: time="2025-02-13T20:17:32.176823659Z" level=warning msg="cleaning up after shim disconnected" id=2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57 namespace=k8s.io Feb 13 20:17:32.177475 containerd[1714]: time="2025-02-13T20:17:32.176852059Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:32.537565 systemd[1]: run-containerd-runc-k8s.io-2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57-runc.AMKNTH.mount: Deactivated successfully. Feb 13 20:17:32.537714 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2392b40c3b9a86932e7ce30171e8677a18531afbf129af9fdf06982b5bcb2c57-rootfs.mount: Deactivated successfully. Feb 13 20:17:32.988391 containerd[1714]: time="2025-02-13T20:17:32.988331759Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Feb 13 20:17:33.022617 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1401572333.mount: Deactivated successfully. Feb 13 20:17:33.027989 kubelet[3264]: I0213 20:17:33.027613 3264 setters.go:602] "Node became not ready" node="ci-4152.2.1-a-2754074fca" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-02-13T20:17:33Z","lastTransitionTime":"2025-02-13T20:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Feb 13 20:17:33.036583 containerd[1714]: time="2025-02-13T20:17:33.036538274Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047\"" Feb 13 20:17:33.037204 containerd[1714]: time="2025-02-13T20:17:33.037170688Z" level=info msg="StartContainer for \"d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047\"" Feb 13 20:17:33.083477 systemd[1]: Started cri-containerd-d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047.scope - libcontainer container d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047. Feb 13 20:17:33.126135 systemd[1]: cri-containerd-d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047.scope: Deactivated successfully. Feb 13 20:17:33.133822 containerd[1714]: time="2025-02-13T20:17:33.133763623Z" level=info msg="StartContainer for \"d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047\" returns successfully" Feb 13 20:17:33.165391 containerd[1714]: time="2025-02-13T20:17:33.165312988Z" level=info msg="shim disconnected" id=d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047 namespace=k8s.io Feb 13 20:17:33.165391 containerd[1714]: time="2025-02-13T20:17:33.165379089Z" level=warning msg="cleaning up after shim disconnected" id=d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047 namespace=k8s.io Feb 13 20:17:33.165391 containerd[1714]: time="2025-02-13T20:17:33.165391290Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:17:33.538423 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d3c0c65642c7a4a27906df476978efa90faa761e47cee5bedb21d98f272ed047-rootfs.mount: Deactivated successfully. Feb 13 20:17:33.582844 kubelet[3264]: E0213 20:17:33.582781 3264 kubelet.go:3008] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Feb 13 20:17:33.992272 containerd[1714]: time="2025-02-13T20:17:33.992204211Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Feb 13 20:17:34.028098 containerd[1714]: time="2025-02-13T20:17:34.028048966Z" level=info msg="CreateContainer within sandbox \"1139e3897a2b5fca4b4a62e5b82ff42a5d4041fc73c7b90d1c2d6ac603ae8b72\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062\"" Feb 13 20:17:34.028715 containerd[1714]: time="2025-02-13T20:17:34.028673380Z" level=info msg="StartContainer for \"1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062\"" Feb 13 20:17:34.064453 systemd[1]: Started cri-containerd-1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062.scope - libcontainer container 1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062. Feb 13 20:17:34.108223 containerd[1714]: time="2025-02-13T20:17:34.108165655Z" level=info msg="StartContainer for \"1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062\" returns successfully" Feb 13 20:17:34.553980 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Feb 13 20:17:37.452127 systemd-networkd[1325]: lxc_health: Link UP Feb 13 20:17:37.462923 systemd-networkd[1325]: lxc_health: Gained carrier Feb 13 20:17:37.750019 kubelet[3264]: I0213 20:17:37.749825 3264 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-2x9zx" podStartSLOduration=8.749800887 podStartE2EDuration="8.749800887s" podCreationTimestamp="2025-02-13 20:17:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:17:35.013696635 +0000 UTC m=+217.127119251" watchObservedRunningTime="2025-02-13 20:17:37.749800887 +0000 UTC m=+219.863223503" Feb 13 20:17:38.033211 systemd[1]: run-containerd-runc-k8s.io-1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062-runc.tSHyTe.mount: Deactivated successfully. Feb 13 20:17:38.880483 systemd-networkd[1325]: lxc_health: Gained IPv6LL Feb 13 20:17:40.294034 kubelet[3264]: E0213 20:17:40.293784 3264 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 127.0.0.1:32866->127.0.0.1:35795: read tcp 127.0.0.1:32866->127.0.0.1:35795: read: connection reset by peer Feb 13 20:17:42.384189 systemd[1]: run-containerd-runc-k8s.io-1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062-runc.scTJoN.mount: Deactivated successfully. Feb 13 20:17:44.523591 systemd[1]: run-containerd-runc-k8s.io-1c061392214dab638576089bee405ed6d4850566e8c79b7df5e5d93468b3c062-runc.zkoCvq.mount: Deactivated successfully. Feb 13 20:17:44.669315 sshd[5219]: Connection closed by 10.200.16.10 port 52694 Feb 13 20:17:44.670297 sshd-session[5155]: pam_unix(sshd:session): session closed for user core Feb 13 20:17:44.675111 systemd[1]: sshd@25-10.200.4.48:22-10.200.16.10:52694.service: Deactivated successfully. Feb 13 20:17:44.677811 systemd[1]: session-28.scope: Deactivated successfully. Feb 13 20:17:44.678907 systemd-logind[1685]: Session 28 logged out. Waiting for processes to exit. Feb 13 20:17:44.680375 systemd-logind[1685]: Removed session 28.