Mar 17 19:05:35.061652 kernel: Linux version 5.15.179-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Mon Mar 17 17:12:34 -00 2025 Mar 17 19:05:35.061677 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=249ccd113f901380672c0d31e18f792e8e0344094c0e39eedc449f039418b31a Mar 17 19:05:35.061687 kernel: BIOS-provided physical RAM map: Mar 17 19:05:35.061695 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 17 19:05:35.061702 kernel: BIOS-e820: [mem 0x00000000000c0000-0x00000000000fffff] reserved Mar 17 19:05:35.061707 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000003ff40fff] usable Mar 17 19:05:35.061717 kernel: BIOS-e820: [mem 0x000000003ff41000-0x000000003ffc8fff] reserved Mar 17 19:05:35.061725 kernel: BIOS-e820: [mem 0x000000003ffc9000-0x000000003fffafff] ACPI data Mar 17 19:05:35.061732 kernel: BIOS-e820: [mem 0x000000003fffb000-0x000000003fffefff] ACPI NVS Mar 17 19:05:35.061739 kernel: BIOS-e820: [mem 0x000000003ffff000-0x000000003fffffff] usable Mar 17 19:05:35.061745 kernel: BIOS-e820: [mem 0x0000000100000000-0x00000002bfffffff] usable Mar 17 19:05:35.061752 kernel: printk: bootconsole [earlyser0] enabled Mar 17 19:05:35.061759 kernel: NX (Execute Disable) protection: active Mar 17 19:05:35.061766 kernel: efi: EFI v2.70 by Microsoft Mar 17 19:05:35.061778 kernel: efi: ACPI=0x3fffa000 ACPI 2.0=0x3fffa014 SMBIOS=0x3ff85000 SMBIOS 3.0=0x3ff83000 MEMATTR=0x3f5c7a98 RNG=0x3ffd1018 Mar 17 19:05:35.061784 kernel: random: crng init done Mar 17 19:05:35.061793 kernel: SMBIOS 3.1.0 present. Mar 17 19:05:35.061800 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 03/08/2024 Mar 17 19:05:35.061809 kernel: Hypervisor detected: Microsoft Hyper-V Mar 17 19:05:35.061815 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0x64e24, misc 0xbed7b2 Mar 17 19:05:35.061822 kernel: Hyper-V Host Build:20348-10.0-1-0.1799 Mar 17 19:05:35.061830 kernel: Hyper-V: Nested features: 0x1e0101 Mar 17 19:05:35.061840 kernel: Hyper-V: LAPIC Timer Frequency: 0x30d40 Mar 17 19:05:35.061848 kernel: Hyper-V: Using hypercall for remote TLB flush Mar 17 19:05:35.061854 kernel: clocksource: hyperv_clocksource_tsc_page: mask: 0xffffffffffffffff max_cycles: 0x24e6a1710, max_idle_ns: 440795202120 ns Mar 17 19:05:35.061862 kernel: tsc: Marking TSC unstable due to running on Hyper-V Mar 17 19:05:35.061870 kernel: tsc: Detected 2593.906 MHz processor Mar 17 19:05:35.061880 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 17 19:05:35.061887 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 17 19:05:35.061894 kernel: last_pfn = 0x2c0000 max_arch_pfn = 0x400000000 Mar 17 19:05:35.061903 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 17 19:05:35.061910 kernel: e820: update [mem 0x40000000-0xffffffff] usable ==> reserved Mar 17 19:05:35.061921 kernel: last_pfn = 0x40000 max_arch_pfn = 0x400000000 Mar 17 19:05:35.061927 kernel: Using GB pages for direct mapping Mar 17 19:05:35.061936 kernel: Secure boot disabled Mar 17 19:05:35.061943 kernel: ACPI: Early table checksum verification disabled Mar 17 19:05:35.061953 kernel: ACPI: RSDP 0x000000003FFFA014 000024 (v02 VRTUAL) Mar 17 19:05:35.061959 kernel: ACPI: XSDT 0x000000003FFF90E8 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.061966 kernel: ACPI: FACP 0x000000003FFF8000 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.061976 kernel: ACPI: DSDT 0x000000003FFD6000 01E184 (v02 MSFTVM DSDT01 00000001 MSFT 05000000) Mar 17 19:05:35.061990 kernel: ACPI: FACS 0x000000003FFFE000 000040 Mar 17 19:05:35.061997 kernel: ACPI: OEM0 0x000000003FFF7000 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.062005 kernel: ACPI: SPCR 0x000000003FFF6000 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.062014 kernel: ACPI: WAET 0x000000003FFF5000 000028 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.062023 kernel: ACPI: APIC 0x000000003FFD5000 000058 (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.062031 kernel: ACPI: SRAT 0x000000003FFD4000 0002D0 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.062040 kernel: ACPI: BGRT 0x000000003FFD3000 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.062050 kernel: ACPI: FPDT 0x000000003FFD2000 000034 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 17 19:05:35.062058 kernel: ACPI: Reserving FACP table memory at [mem 0x3fff8000-0x3fff8113] Mar 17 19:05:35.062081 kernel: ACPI: Reserving DSDT table memory at [mem 0x3ffd6000-0x3fff4183] Mar 17 19:05:35.062089 kernel: ACPI: Reserving FACS table memory at [mem 0x3fffe000-0x3fffe03f] Mar 17 19:05:35.062099 kernel: ACPI: Reserving OEM0 table memory at [mem 0x3fff7000-0x3fff7063] Mar 17 19:05:35.062106 kernel: ACPI: Reserving SPCR table memory at [mem 0x3fff6000-0x3fff604f] Mar 17 19:05:35.062114 kernel: ACPI: Reserving WAET table memory at [mem 0x3fff5000-0x3fff5027] Mar 17 19:05:35.062126 kernel: ACPI: Reserving APIC table memory at [mem 0x3ffd5000-0x3ffd5057] Mar 17 19:05:35.062136 kernel: ACPI: Reserving SRAT table memory at [mem 0x3ffd4000-0x3ffd42cf] Mar 17 19:05:35.062142 kernel: ACPI: Reserving BGRT table memory at [mem 0x3ffd3000-0x3ffd3037] Mar 17 19:05:35.062151 kernel: ACPI: Reserving FPDT table memory at [mem 0x3ffd2000-0x3ffd2033] Mar 17 19:05:35.062159 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Mar 17 19:05:35.062168 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Mar 17 19:05:35.062176 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Mar 17 19:05:35.062183 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x2bfffffff] hotplug Mar 17 19:05:35.062193 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2c0000000-0xfdfffffff] hotplug Mar 17 19:05:35.062202 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Mar 17 19:05:35.062211 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Mar 17 19:05:35.062218 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Mar 17 19:05:35.062227 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Mar 17 19:05:35.062235 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Mar 17 19:05:35.062246 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Mar 17 19:05:35.062252 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Mar 17 19:05:35.062260 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Mar 17 19:05:35.062269 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Mar 17 19:05:35.062281 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000000-0x1ffffffffffff] hotplug Mar 17 19:05:35.062289 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x2000000000000-0x3ffffffffffff] hotplug Mar 17 19:05:35.062298 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x4000000000000-0x7ffffffffffff] hotplug Mar 17 19:05:35.062307 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x8000000000000-0xfffffffffffff] hotplug Mar 17 19:05:35.062315 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x2bfffffff] -> [mem 0x00000000-0x2bfffffff] Mar 17 19:05:35.062322 kernel: NODE_DATA(0) allocated [mem 0x2bfffa000-0x2bfffffff] Mar 17 19:05:35.062329 kernel: Zone ranges: Mar 17 19:05:35.062336 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 17 19:05:35.062343 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Mar 17 19:05:35.062355 kernel: Normal [mem 0x0000000100000000-0x00000002bfffffff] Mar 17 19:05:35.062362 kernel: Movable zone start for each node Mar 17 19:05:35.062370 kernel: Early memory node ranges Mar 17 19:05:35.062379 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Mar 17 19:05:35.062387 kernel: node 0: [mem 0x0000000000100000-0x000000003ff40fff] Mar 17 19:05:35.062396 kernel: node 0: [mem 0x000000003ffff000-0x000000003fffffff] Mar 17 19:05:35.062403 kernel: node 0: [mem 0x0000000100000000-0x00000002bfffffff] Mar 17 19:05:35.062413 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x00000002bfffffff] Mar 17 19:05:35.062420 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 17 19:05:35.062431 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Mar 17 19:05:35.062438 kernel: On node 0, zone DMA32: 190 pages in unavailable ranges Mar 17 19:05:35.062447 kernel: ACPI: PM-Timer IO Port: 0x408 Mar 17 19:05:35.062455 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] dfl dfl lint[0x1]) Mar 17 19:05:35.062465 kernel: IOAPIC[0]: apic_id 2, version 17, address 0xfec00000, GSI 0-23 Mar 17 19:05:35.062472 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 17 19:05:35.062479 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 17 19:05:35.062489 kernel: ACPI: SPCR: console: uart,io,0x3f8,115200 Mar 17 19:05:35.062497 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Mar 17 19:05:35.062508 kernel: [mem 0x40000000-0xffffffff] available for PCI devices Mar 17 19:05:35.062515 kernel: Booting paravirtualized kernel on Hyper-V Mar 17 19:05:35.062525 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 17 19:05:35.062533 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 Mar 17 19:05:35.062542 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u1048576 Mar 17 19:05:35.062549 kernel: pcpu-alloc: s188696 r8192 d32488 u1048576 alloc=1*2097152 Mar 17 19:05:35.062559 kernel: pcpu-alloc: [0] 0 1 Mar 17 19:05:35.062566 kernel: Hyper-V: PV spinlocks enabled Mar 17 19:05:35.062576 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 17 19:05:35.062584 kernel: Built 1 zonelists, mobility grouping on. Total pages: 2062618 Mar 17 19:05:35.062594 kernel: Policy zone: Normal Mar 17 19:05:35.062603 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=249ccd113f901380672c0d31e18f792e8e0344094c0e39eedc449f039418b31a Mar 17 19:05:35.062612 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Mar 17 19:05:35.062619 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Mar 17 19:05:35.062628 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 17 19:05:35.062636 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 17 19:05:35.062646 kernel: Memory: 8071680K/8387460K available (12294K kernel code, 2278K rwdata, 13724K rodata, 47472K init, 4108K bss, 315520K reserved, 0K cma-reserved) Mar 17 19:05:35.062655 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 17 19:05:35.062664 kernel: ftrace: allocating 34580 entries in 136 pages Mar 17 19:05:35.062680 kernel: ftrace: allocated 136 pages with 2 groups Mar 17 19:05:35.062690 kernel: rcu: Hierarchical RCU implementation. Mar 17 19:05:35.062699 kernel: rcu: RCU event tracing is enabled. Mar 17 19:05:35.062708 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 17 19:05:35.062719 kernel: Rude variant of Tasks RCU enabled. Mar 17 19:05:35.062726 kernel: Tracing variant of Tasks RCU enabled. Mar 17 19:05:35.062734 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 17 19:05:35.062744 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 17 19:05:35.062753 kernel: Using NULL legacy PIC Mar 17 19:05:35.062764 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 0 Mar 17 19:05:35.062773 kernel: Console: colour dummy device 80x25 Mar 17 19:05:35.062782 kernel: printk: console [tty1] enabled Mar 17 19:05:35.062792 kernel: printk: console [ttyS0] enabled Mar 17 19:05:35.062800 kernel: printk: bootconsole [earlyser0] disabled Mar 17 19:05:35.062809 kernel: ACPI: Core revision 20210730 Mar 17 19:05:35.062816 kernel: Failed to register legacy timer interrupt Mar 17 19:05:35.062826 kernel: APIC: Switch to symmetric I/O mode setup Mar 17 19:05:35.062834 kernel: Hyper-V: Using IPI hypercalls Mar 17 19:05:35.062844 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 5187.81 BogoMIPS (lpj=2593906) Mar 17 19:05:35.062851 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Mar 17 19:05:35.062862 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Mar 17 19:05:35.062870 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 17 19:05:35.062880 kernel: Spectre V2 : Mitigation: Retpolines Mar 17 19:05:35.062887 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Mar 17 19:05:35.062899 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Mar 17 19:05:35.062911 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Mar 17 19:05:35.062919 kernel: RETBleed: Vulnerable Mar 17 19:05:35.062927 kernel: Speculative Store Bypass: Vulnerable Mar 17 19:05:35.062937 kernel: TAA: Vulnerable: Clear CPU buffers attempted, no microcode Mar 17 19:05:35.062946 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Mar 17 19:05:35.062954 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 17 19:05:35.062962 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 17 19:05:35.062972 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 17 19:05:35.062980 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Mar 17 19:05:35.062991 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Mar 17 19:05:35.062998 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Mar 17 19:05:35.063009 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 17 19:05:35.063016 kernel: x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 Mar 17 19:05:35.063026 kernel: x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 Mar 17 19:05:35.063033 kernel: x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 Mar 17 19:05:35.063043 kernel: x86/fpu: Enabled xstate features 0xe7, context size is 2432 bytes, using 'compacted' format. Mar 17 19:05:35.063051 kernel: Freeing SMP alternatives memory: 32K Mar 17 19:05:35.063081 kernel: pid_max: default: 32768 minimum: 301 Mar 17 19:05:35.063090 kernel: LSM: Security Framework initializing Mar 17 19:05:35.063099 kernel: SELinux: Initializing. Mar 17 19:05:35.063106 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Mar 17 19:05:35.063119 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Mar 17 19:05:35.063128 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8272CL CPU @ 2.60GHz (family: 0x6, model: 0x55, stepping: 0x7) Mar 17 19:05:35.063137 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Mar 17 19:05:35.063144 kernel: signal: max sigframe size: 3632 Mar 17 19:05:35.063154 kernel: rcu: Hierarchical SRCU implementation. Mar 17 19:05:35.063162 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Mar 17 19:05:35.063172 kernel: smp: Bringing up secondary CPUs ... Mar 17 19:05:35.063179 kernel: x86: Booting SMP configuration: Mar 17 19:05:35.063188 kernel: .... node #0, CPUs: #1 Mar 17 19:05:35.063197 kernel: TAA CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/tsx_async_abort.html for more details. Mar 17 19:05:35.063210 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Mar 17 19:05:35.063217 kernel: smp: Brought up 1 node, 2 CPUs Mar 17 19:05:35.063226 kernel: smpboot: Max logical packages: 1 Mar 17 19:05:35.063235 kernel: smpboot: Total of 2 processors activated (10375.62 BogoMIPS) Mar 17 19:05:35.063245 kernel: devtmpfs: initialized Mar 17 19:05:35.063253 kernel: x86/mm: Memory block size: 128MB Mar 17 19:05:35.063260 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x3fffb000-0x3fffefff] (16384 bytes) Mar 17 19:05:35.063270 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 17 19:05:35.063281 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 17 19:05:35.063290 kernel: pinctrl core: initialized pinctrl subsystem Mar 17 19:05:35.063297 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 17 19:05:35.063307 kernel: audit: initializing netlink subsys (disabled) Mar 17 19:05:35.063316 kernel: audit: type=2000 audit(1742238333.024:1): state=initialized audit_enabled=0 res=1 Mar 17 19:05:35.063326 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 17 19:05:35.063333 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 17 19:05:35.063342 kernel: cpuidle: using governor menu Mar 17 19:05:35.063350 kernel: ACPI: bus type PCI registered Mar 17 19:05:35.063362 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 17 19:05:35.063369 kernel: dca service started, version 1.12.1 Mar 17 19:05:35.063379 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 17 19:05:35.063387 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Mar 17 19:05:35.063397 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Mar 17 19:05:35.063405 kernel: ACPI: Added _OSI(Module Device) Mar 17 19:05:35.063413 kernel: ACPI: Added _OSI(Processor Device) Mar 17 19:05:35.063423 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Mar 17 19:05:35.063433 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 17 19:05:35.063442 kernel: ACPI: Added _OSI(Linux-Dell-Video) Mar 17 19:05:35.063450 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Mar 17 19:05:35.063460 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Mar 17 19:05:35.063470 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 17 19:05:35.063478 kernel: ACPI: Interpreter enabled Mar 17 19:05:35.063485 kernel: ACPI: PM: (supports S0 S5) Mar 17 19:05:35.063497 kernel: ACPI: Using IOAPIC for interrupt routing Mar 17 19:05:35.063511 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 17 19:05:35.063525 kernel: ACPI: Enabled 1 GPEs in block 00 to 0F Mar 17 19:05:35.063544 kernel: iommu: Default domain type: Translated Mar 17 19:05:35.063558 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 17 19:05:35.063573 kernel: vgaarb: loaded Mar 17 19:05:35.063587 kernel: pps_core: LinuxPPS API ver. 1 registered Mar 17 19:05:35.063602 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Mar 17 19:05:35.063618 kernel: PTP clock support registered Mar 17 19:05:35.063632 kernel: Registered efivars operations Mar 17 19:05:35.063646 kernel: PCI: Using ACPI for IRQ routing Mar 17 19:05:35.063661 kernel: PCI: System does not support PCI Mar 17 19:05:35.063678 kernel: clocksource: Switched to clocksource hyperv_clocksource_tsc_page Mar 17 19:05:35.063692 kernel: VFS: Disk quotas dquot_6.6.0 Mar 17 19:05:35.063706 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 17 19:05:35.063722 kernel: pnp: PnP ACPI init Mar 17 19:05:35.063736 kernel: pnp: PnP ACPI: found 3 devices Mar 17 19:05:35.063751 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 17 19:05:35.063765 kernel: NET: Registered PF_INET protocol family Mar 17 19:05:35.063779 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Mar 17 19:05:35.063793 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Mar 17 19:05:35.063811 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 17 19:05:35.063824 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 17 19:05:35.063838 kernel: TCP bind hash table entries: 65536 (order: 8, 1048576 bytes, linear) Mar 17 19:05:35.063852 kernel: TCP: Hash tables configured (established 65536 bind 65536) Mar 17 19:05:35.063865 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Mar 17 19:05:35.063880 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Mar 17 19:05:35.063893 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 17 19:05:35.063908 kernel: NET: Registered PF_XDP protocol family Mar 17 19:05:35.063922 kernel: PCI: CLS 0 bytes, default 64 Mar 17 19:05:35.063942 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Mar 17 19:05:35.063957 kernel: software IO TLB: mapped [mem 0x000000003a8ad000-0x000000003e8ad000] (64MB) Mar 17 19:05:35.063972 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Mar 17 19:05:35.063987 kernel: Initialise system trusted keyrings Mar 17 19:05:35.064001 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Mar 17 19:05:35.064016 kernel: Key type asymmetric registered Mar 17 19:05:35.064031 kernel: Asymmetric key parser 'x509' registered Mar 17 19:05:35.064045 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Mar 17 19:05:35.064069 kernel: io scheduler mq-deadline registered Mar 17 19:05:35.064085 kernel: io scheduler kyber registered Mar 17 19:05:35.064099 kernel: io scheduler bfq registered Mar 17 19:05:35.064114 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 17 19:05:35.064128 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 17 19:05:35.064142 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 17 19:05:35.064157 kernel: 00:01: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Mar 17 19:05:35.064173 kernel: i8042: PNP: No PS/2 controller found. Mar 17 19:05:35.064362 kernel: rtc_cmos 00:02: registered as rtc0 Mar 17 19:05:35.064477 kernel: rtc_cmos 00:02: setting system clock to 2025-03-17T19:05:34 UTC (1742238334) Mar 17 19:05:35.064579 kernel: rtc_cmos 00:02: alarms up to one month, 114 bytes nvram Mar 17 19:05:35.064595 kernel: intel_pstate: CPU model not supported Mar 17 19:05:35.064607 kernel: efifb: probing for efifb Mar 17 19:05:35.064621 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Mar 17 19:05:35.064633 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Mar 17 19:05:35.064646 kernel: efifb: scrolling: redraw Mar 17 19:05:35.064659 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 17 19:05:35.064673 kernel: Console: switching to colour frame buffer device 128x48 Mar 17 19:05:35.064691 kernel: fb0: EFI VGA frame buffer device Mar 17 19:05:35.064704 kernel: pstore: Registered efi as persistent store backend Mar 17 19:05:35.064717 kernel: NET: Registered PF_INET6 protocol family Mar 17 19:05:35.064731 kernel: Segment Routing with IPv6 Mar 17 19:05:35.064745 kernel: In-situ OAM (IOAM) with IPv6 Mar 17 19:05:35.064758 kernel: NET: Registered PF_PACKET protocol family Mar 17 19:05:35.064772 kernel: Key type dns_resolver registered Mar 17 19:05:35.064785 kernel: IPI shorthand broadcast: enabled Mar 17 19:05:35.064798 kernel: sched_clock: Marking stable (749194400, 21575200)->(964983600, -194214000) Mar 17 19:05:35.064815 kernel: registered taskstats version 1 Mar 17 19:05:35.064828 kernel: Loading compiled-in X.509 certificates Mar 17 19:05:35.064841 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.179-flatcar: d5b956bbabb2d386c0246a969032c0de9eaa8220' Mar 17 19:05:35.064854 kernel: Key type .fscrypt registered Mar 17 19:05:35.064867 kernel: Key type fscrypt-provisioning registered Mar 17 19:05:35.064881 kernel: pstore: Using crash dump compression: deflate Mar 17 19:05:35.064895 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 17 19:05:35.064908 kernel: ima: Allocated hash algorithm: sha1 Mar 17 19:05:35.064925 kernel: ima: No architecture policies found Mar 17 19:05:35.064938 kernel: clk: Disabling unused clocks Mar 17 19:05:35.064951 kernel: Freeing unused kernel image (initmem) memory: 47472K Mar 17 19:05:35.064965 kernel: Write protecting the kernel read-only data: 28672k Mar 17 19:05:35.064979 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Mar 17 19:05:35.064993 kernel: Freeing unused kernel image (rodata/data gap) memory: 612K Mar 17 19:05:35.065006 kernel: Run /init as init process Mar 17 19:05:35.065020 kernel: with arguments: Mar 17 19:05:35.065034 kernel: /init Mar 17 19:05:35.065049 kernel: with environment: Mar 17 19:05:35.065098 kernel: HOME=/ Mar 17 19:05:35.065112 kernel: TERM=linux Mar 17 19:05:35.065125 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Mar 17 19:05:35.065142 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Mar 17 19:05:35.065159 systemd[1]: Detected virtualization microsoft. Mar 17 19:05:35.065173 systemd[1]: Detected architecture x86-64. Mar 17 19:05:35.065187 systemd[1]: Running in initrd. Mar 17 19:05:35.065204 systemd[1]: No hostname configured, using default hostname. Mar 17 19:05:35.065218 systemd[1]: Hostname set to . Mar 17 19:05:35.065234 systemd[1]: Initializing machine ID from random generator. Mar 17 19:05:35.065248 systemd[1]: Queued start job for default target initrd.target. Mar 17 19:05:35.065263 systemd[1]: Started systemd-ask-password-console.path. Mar 17 19:05:35.065277 systemd[1]: Reached target cryptsetup.target. Mar 17 19:05:35.065291 systemd[1]: Reached target paths.target. Mar 17 19:05:35.065305 systemd[1]: Reached target slices.target. Mar 17 19:05:35.065319 systemd[1]: Reached target swap.target. Mar 17 19:05:35.065336 systemd[1]: Reached target timers.target. Mar 17 19:05:35.065351 systemd[1]: Listening on iscsid.socket. Mar 17 19:05:35.065366 systemd[1]: Listening on iscsiuio.socket. Mar 17 19:05:35.065381 systemd[1]: Listening on systemd-journald-audit.socket. Mar 17 19:05:35.065395 systemd[1]: Listening on systemd-journald-dev-log.socket. Mar 17 19:05:35.065409 systemd[1]: Listening on systemd-journald.socket. Mar 17 19:05:35.065424 systemd[1]: Listening on systemd-networkd.socket. Mar 17 19:05:35.065440 systemd[1]: Listening on systemd-udevd-control.socket. Mar 17 19:05:35.065455 systemd[1]: Listening on systemd-udevd-kernel.socket. Mar 17 19:05:35.065469 systemd[1]: Reached target sockets.target. Mar 17 19:05:35.065484 systemd[1]: Starting kmod-static-nodes.service... Mar 17 19:05:35.065499 systemd[1]: Finished network-cleanup.service. Mar 17 19:05:35.065513 systemd[1]: Starting systemd-fsck-usr.service... Mar 17 19:05:35.065527 systemd[1]: Starting systemd-journald.service... Mar 17 19:05:35.065542 systemd[1]: Starting systemd-modules-load.service... Mar 17 19:05:35.065557 systemd[1]: Starting systemd-resolved.service... Mar 17 19:05:35.065574 systemd[1]: Starting systemd-vconsole-setup.service... Mar 17 19:05:35.065588 systemd[1]: Finished kmod-static-nodes.service. Mar 17 19:05:35.065603 kernel: audit: type=1130 audit(1742238335.049:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.065617 systemd[1]: Finished systemd-fsck-usr.service. Mar 17 19:05:35.065636 systemd-journald[183]: Journal started Mar 17 19:05:35.065700 systemd-journald[183]: Runtime Journal (/run/log/journal/d433c672e861452fae60440e3261e1bf) is 8.0M, max 159.0M, 151.0M free. Mar 17 19:05:35.049000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.035456 systemd-modules-load[184]: Inserted module 'overlay' Mar 17 19:05:35.076000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.093141 systemd[1]: Started systemd-journald.service. Mar 17 19:05:35.093184 kernel: audit: type=1130 audit(1742238335.076:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.095885 systemd[1]: Finished systemd-vconsole-setup.service. Mar 17 19:05:35.120418 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 17 19:05:35.120474 kernel: audit: type=1130 audit(1742238335.095:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.095000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.108000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.132250 systemd-resolved[185]: Positive Trust Anchors: Mar 17 19:05:35.148020 kernel: Bridge firewalling registered Mar 17 19:05:35.148043 kernel: audit: type=1130 audit(1742238335.108:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.132259 systemd-resolved[185]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 19:05:35.132292 systemd-resolved[185]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Mar 17 19:05:35.134951 systemd-resolved[185]: Defaulting to hostname 'linux'. Mar 17 19:05:35.135907 systemd[1]: Starting dracut-cmdline-ask.service... Mar 17 19:05:35.144025 systemd-modules-load[184]: Inserted module 'br_netfilter' Mar 17 19:05:35.148908 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Mar 17 19:05:35.175780 systemd[1]: Started systemd-resolved.service. Mar 17 19:05:35.180523 systemd[1]: Finished dracut-cmdline-ask.service. Mar 17 19:05:35.180000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.194850 kernel: audit: type=1130 audit(1742238335.180:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.195800 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Mar 17 19:05:35.233284 kernel: audit: type=1130 audit(1742238335.195:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.233324 kernel: SCSI subsystem initialized Mar 17 19:05:35.233344 kernel: audit: type=1130 audit(1742238335.212:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.195000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.212000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.212667 systemd[1]: Reached target nss-lookup.target. Mar 17 19:05:35.226950 systemd[1]: Starting dracut-cmdline.service... Mar 17 19:05:35.245031 dracut-cmdline[202]: dracut-dracut-053 Mar 17 19:05:35.256539 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 17 19:05:35.256569 kernel: device-mapper: uevent: version 1.0.3 Mar 17 19:05:35.256585 dracut-cmdline[202]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlyprintk=ttyS0,115200 flatcar.first_boot=detected flatcar.oem.id=azure flatcar.autologin verity.usrhash=249ccd113f901380672c0d31e18f792e8e0344094c0e39eedc449f039418b31a Mar 17 19:05:35.275732 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Mar 17 19:05:35.278282 systemd-modules-load[184]: Inserted module 'dm_multipath' Mar 17 19:05:35.280046 systemd[1]: Finished systemd-modules-load.service. Mar 17 19:05:35.285000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.286132 systemd[1]: Starting systemd-sysctl.service... Mar 17 19:05:35.300255 kernel: audit: type=1130 audit(1742238335.285:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.309000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.307658 systemd[1]: Finished systemd-sysctl.service. Mar 17 19:05:35.324594 kernel: audit: type=1130 audit(1742238335.309:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.349083 kernel: Loading iSCSI transport class v2.0-870. Mar 17 19:05:35.368091 kernel: iscsi: registered transport (tcp) Mar 17 19:05:35.394797 kernel: iscsi: registered transport (qla4xxx) Mar 17 19:05:35.394871 kernel: QLogic iSCSI HBA Driver Mar 17 19:05:35.423973 systemd[1]: Finished dracut-cmdline.service. Mar 17 19:05:35.428000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.429203 systemd[1]: Starting dracut-pre-udev.service... Mar 17 19:05:35.479090 kernel: raid6: avx512x4 gen() 18249 MB/s Mar 17 19:05:35.499078 kernel: raid6: avx512x4 xor() 8137 MB/s Mar 17 19:05:35.519074 kernel: raid6: avx512x2 gen() 18361 MB/s Mar 17 19:05:35.539080 kernel: raid6: avx512x2 xor() 29759 MB/s Mar 17 19:05:35.559073 kernel: raid6: avx512x1 gen() 18506 MB/s Mar 17 19:05:35.579073 kernel: raid6: avx512x1 xor() 26929 MB/s Mar 17 19:05:35.599074 kernel: raid6: avx2x4 gen() 18440 MB/s Mar 17 19:05:35.619073 kernel: raid6: avx2x4 xor() 7825 MB/s Mar 17 19:05:35.639071 kernel: raid6: avx2x2 gen() 18455 MB/s Mar 17 19:05:35.659075 kernel: raid6: avx2x2 xor() 21788 MB/s Mar 17 19:05:35.679071 kernel: raid6: avx2x1 gen() 14125 MB/s Mar 17 19:05:35.699071 kernel: raid6: avx2x1 xor() 19396 MB/s Mar 17 19:05:35.719074 kernel: raid6: sse2x4 gen() 11758 MB/s Mar 17 19:05:35.738071 kernel: raid6: sse2x4 xor() 7185 MB/s Mar 17 19:05:35.758071 kernel: raid6: sse2x2 gen() 12848 MB/s Mar 17 19:05:35.778073 kernel: raid6: sse2x2 xor() 7672 MB/s Mar 17 19:05:35.798072 kernel: raid6: sse2x1 gen() 11630 MB/s Mar 17 19:05:35.820787 kernel: raid6: sse2x1 xor() 5968 MB/s Mar 17 19:05:35.820807 kernel: raid6: using algorithm avx512x1 gen() 18506 MB/s Mar 17 19:05:35.820830 kernel: raid6: .... xor() 26929 MB/s, rmw enabled Mar 17 19:05:35.824074 kernel: raid6: using avx512x2 recovery algorithm Mar 17 19:05:35.843085 kernel: xor: automatically using best checksumming function avx Mar 17 19:05:35.939089 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Mar 17 19:05:35.947325 systemd[1]: Finished dracut-pre-udev.service. Mar 17 19:05:35.949000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.951000 audit: BPF prog-id=7 op=LOAD Mar 17 19:05:35.951000 audit: BPF prog-id=8 op=LOAD Mar 17 19:05:35.952221 systemd[1]: Starting systemd-udevd.service... Mar 17 19:05:35.966785 systemd-udevd[385]: Using default interface naming scheme 'v252'. Mar 17 19:05:35.976000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:35.973764 systemd[1]: Started systemd-udevd.service. Mar 17 19:05:35.977332 systemd[1]: Starting dracut-pre-trigger.service... Mar 17 19:05:35.997562 dracut-pre-trigger[399]: rd.md=0: removing MD RAID activation Mar 17 19:05:36.031000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:36.029497 systemd[1]: Finished dracut-pre-trigger.service. Mar 17 19:05:36.034695 systemd[1]: Starting systemd-udev-trigger.service... Mar 17 19:05:36.072000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:36.069376 systemd[1]: Finished systemd-udev-trigger.service. Mar 17 19:05:36.119081 kernel: cryptd: max_cpu_qlen set to 1000 Mar 17 19:05:36.136236 kernel: hv_vmbus: Vmbus version:5.2 Mar 17 19:05:36.144433 kernel: AVX2 version of gcm_enc/dec engaged. Mar 17 19:05:36.144497 kernel: AES CTR mode by8 optimization enabled Mar 17 19:05:36.159550 kernel: hv_vmbus: registering driver hyperv_keyboard Mar 17 19:05:36.171092 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 17 19:05:36.185080 kernel: hv_vmbus: registering driver hv_storvsc Mar 17 19:05:36.195083 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/VMBUS:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Mar 17 19:05:36.209085 kernel: scsi host1: storvsc_host_t Mar 17 19:05:36.209279 kernel: scsi host0: storvsc_host_t Mar 17 19:05:36.216899 kernel: hv_vmbus: registering driver hid_hyperv Mar 17 19:05:36.216960 kernel: hv_vmbus: registering driver hv_netvsc Mar 17 19:05:36.222970 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Mar 17 19:05:36.223001 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Mar 17 19:05:36.235752 kernel: hid 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Mar 17 19:05:36.245257 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Mar 17 19:05:36.277457 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Mar 17 19:05:36.278384 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 17 19:05:36.278402 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Mar 17 19:05:36.281080 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Mar 17 19:05:36.302289 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Mar 17 19:05:36.302471 kernel: sd 0:0:0:0: [sda] Write Protect is off Mar 17 19:05:36.302641 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Mar 17 19:05:36.302797 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Mar 17 19:05:36.302965 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 17 19:05:36.302984 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Mar 17 19:05:36.426811 kernel: hv_netvsc 7c1e52fc-a1da-7c1e-52fc-a1da7c1e52fc eth0: VF slot 1 added Mar 17 19:05:36.436345 kernel: hv_vmbus: registering driver hv_pci Mar 17 19:05:36.442077 kernel: hv_pci 51be81f3-d8d7-459f-91c8-94d8afbab5dc: PCI VMBus probing: Using version 0x10004 Mar 17 19:05:36.522284 kernel: hv_pci 51be81f3-d8d7-459f-91c8-94d8afbab5dc: PCI host bridge to bus d8d7:00 Mar 17 19:05:36.522472 kernel: pci_bus d8d7:00: root bus resource [mem 0xfe0000000-0xfe00fffff window] Mar 17 19:05:36.522637 kernel: pci_bus d8d7:00: No busn resource found for root bus, will use [bus 00-ff] Mar 17 19:05:36.522789 kernel: pci d8d7:00:02.0: [15b3:1016] type 00 class 0x020000 Mar 17 19:05:36.522962 kernel: pci d8d7:00:02.0: reg 0x10: [mem 0xfe0000000-0xfe00fffff 64bit pref] Mar 17 19:05:36.523138 kernel: pci d8d7:00:02.0: enabling Extended Tags Mar 17 19:05:36.523293 kernel: pci d8d7:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at d8d7:00:02.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Mar 17 19:05:36.523446 kernel: pci_bus d8d7:00: busn_res: [bus 00-ff] end is updated to 00 Mar 17 19:05:36.523597 kernel: pci d8d7:00:02.0: BAR 0: assigned [mem 0xfe0000000-0xfe00fffff 64bit pref] Mar 17 19:05:36.620899 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Mar 17 19:05:36.634081 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (452) Mar 17 19:05:36.634137 kernel: mlx5_core d8d7:00:02.0: firmware version: 14.30.5000 Mar 17 19:05:36.919282 kernel: mlx5_core d8d7:00:02.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0) Mar 17 19:05:36.919477 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 17 19:05:36.919497 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 17 19:05:36.919514 kernel: mlx5_core d8d7:00:02.0: Supported tc offload range - chains: 1, prios: 1 Mar 17 19:05:36.919664 kernel: mlx5_core d8d7:00:02.0: mlx5e_tc_post_act_init:40:(pid 188): firmware level support is missing Mar 17 19:05:36.919804 kernel: hv_netvsc 7c1e52fc-a1da-7c1e-52fc-a1da7c1e52fc eth0: VF registering: eth1 Mar 17 19:05:36.919902 kernel: mlx5_core d8d7:00:02.0 eth1: joined to eth0 Mar 17 19:05:36.655194 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Mar 17 19:05:36.788247 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Mar 17 19:05:36.817509 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Mar 17 19:05:36.820076 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Mar 17 19:05:36.823704 systemd[1]: Starting disk-uuid.service... Mar 17 19:05:36.943082 kernel: mlx5_core d8d7:00:02.0 enP55511s1: renamed from eth1 Mar 17 19:05:37.851969 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 17 19:05:37.852399 disk-uuid[564]: The operation has completed successfully. Mar 17 19:05:37.940524 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 17 19:05:37.942000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:37.942000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:37.940658 systemd[1]: Finished disk-uuid.service. Mar 17 19:05:37.951577 systemd[1]: Starting verity-setup.service... Mar 17 19:05:37.984082 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Mar 17 19:05:38.245498 systemd[1]: Found device dev-mapper-usr.device. Mar 17 19:05:38.253564 systemd[1]: Mounting sysusr-usr.mount... Mar 17 19:05:38.257000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.255803 systemd[1]: Finished verity-setup.service. Mar 17 19:05:38.333091 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Mar 17 19:05:38.333227 systemd[1]: Mounted sysusr-usr.mount. Mar 17 19:05:38.336829 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Mar 17 19:05:38.340946 systemd[1]: Starting ignition-setup.service... Mar 17 19:05:38.348570 systemd[1]: Starting parse-ip-for-networkd.service... Mar 17 19:05:38.370092 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 17 19:05:38.370147 kernel: BTRFS info (device sda6): using free space tree Mar 17 19:05:38.370161 kernel: BTRFS info (device sda6): has skinny extents Mar 17 19:05:38.414010 systemd[1]: Finished parse-ip-for-networkd.service. Mar 17 19:05:38.418000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.419000 audit: BPF prog-id=9 op=LOAD Mar 17 19:05:38.420249 systemd[1]: Starting systemd-networkd.service... Mar 17 19:05:38.443971 systemd-networkd[805]: lo: Link UP Mar 17 19:05:38.443983 systemd-networkd[805]: lo: Gained carrier Mar 17 19:05:38.448200 systemd-networkd[805]: Enumeration completed Mar 17 19:05:38.450000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.448292 systemd[1]: Started systemd-networkd.service. Mar 17 19:05:38.450524 systemd[1]: Reached target network.target. Mar 17 19:05:38.450921 systemd-networkd[805]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 19:05:38.461939 systemd[1]: Starting iscsiuio.service... Mar 17 19:05:38.468357 systemd[1]: Started iscsiuio.service. Mar 17 19:05:38.476000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.477933 systemd[1]: Starting iscsid.service... Mar 17 19:05:38.485644 iscsid[815]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Mar 17 19:05:38.485644 iscsid[815]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Mar 17 19:05:38.485644 iscsid[815]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Mar 17 19:05:38.485644 iscsid[815]: If using hardware iscsi like qla4xxx this message can be ignored. Mar 17 19:05:38.485644 iscsid[815]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Mar 17 19:05:38.485644 iscsid[815]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Mar 17 19:05:38.488501 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 17 19:05:38.517187 systemd[1]: Started iscsid.service. Mar 17 19:05:38.518000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.520499 systemd[1]: Starting dracut-initqueue.service... Mar 17 19:05:38.531120 kernel: mlx5_core d8d7:00:02.0 enP55511s1: Link up Mar 17 19:05:38.533802 systemd[1]: Finished dracut-initqueue.service. Mar 17 19:05:38.534000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.535164 systemd[1]: Reached target remote-fs-pre.target. Mar 17 19:05:38.550000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.535429 systemd[1]: Reached target remote-cryptsetup.target. Mar 17 19:05:38.536735 systemd[1]: Reached target remote-fs.target. Mar 17 19:05:38.538464 systemd[1]: Starting dracut-pre-mount.service... Mar 17 19:05:38.547103 systemd[1]: Finished dracut-pre-mount.service. Mar 17 19:05:38.567082 kernel: hv_netvsc 7c1e52fc-a1da-7c1e-52fc-a1da7c1e52fc eth0: Data path switched to VF: enP55511s1 Mar 17 19:05:38.567272 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Mar 17 19:05:38.570699 systemd-networkd[805]: enP55511s1: Link UP Mar 17 19:05:38.570937 systemd-networkd[805]: eth0: Link UP Mar 17 19:05:38.571109 systemd-networkd[805]: eth0: Gained carrier Mar 17 19:05:38.576283 systemd-networkd[805]: enP55511s1: Gained carrier Mar 17 19:05:38.612161 systemd-networkd[805]: eth0: DHCPv4 address 10.200.8.15/24, gateway 10.200.8.1 acquired from 168.63.129.16 Mar 17 19:05:38.619543 systemd[1]: Finished ignition-setup.service. Mar 17 19:05:38.623000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:38.624654 systemd[1]: Starting ignition-fetch-offline.service... Mar 17 19:05:39.671378 systemd-networkd[805]: eth0: Gained IPv6LL Mar 17 19:05:41.270357 ignition[832]: Ignition 2.14.0 Mar 17 19:05:41.270376 ignition[832]: Stage: fetch-offline Mar 17 19:05:41.270464 ignition[832]: reading system config file "/usr/lib/ignition/base.d/base.ign" Mar 17 19:05:41.270514 ignition[832]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Mar 17 19:05:41.416411 ignition[832]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 17 19:05:41.416636 ignition[832]: parsed url from cmdline: "" Mar 17 19:05:41.419300 ignition[832]: no config URL provided Mar 17 19:05:41.419322 ignition[832]: reading system config file "/usr/lib/ignition/user.ign" Mar 17 19:05:41.419338 ignition[832]: no config at "/usr/lib/ignition/user.ign" Mar 17 19:05:41.419347 ignition[832]: failed to fetch config: resource requires networking Mar 17 19:05:41.420563 ignition[832]: Ignition finished successfully Mar 17 19:05:41.430721 systemd[1]: Finished ignition-fetch-offline.service. Mar 17 19:05:41.433000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.434170 systemd[1]: Starting ignition-fetch.service... Mar 17 19:05:41.453765 kernel: kauditd_printk_skb: 18 callbacks suppressed Mar 17 19:05:41.453799 kernel: audit: type=1130 audit(1742238341.433:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.456139 ignition[838]: Ignition 2.14.0 Mar 17 19:05:41.456150 ignition[838]: Stage: fetch Mar 17 19:05:41.456303 ignition[838]: reading system config file "/usr/lib/ignition/base.d/base.ign" Mar 17 19:05:41.456337 ignition[838]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Mar 17 19:05:41.460300 ignition[838]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 17 19:05:41.460494 ignition[838]: parsed url from cmdline: "" Mar 17 19:05:41.460498 ignition[838]: no config URL provided Mar 17 19:05:41.460503 ignition[838]: reading system config file "/usr/lib/ignition/user.ign" Mar 17 19:05:41.460513 ignition[838]: no config at "/usr/lib/ignition/user.ign" Mar 17 19:05:41.460548 ignition[838]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Mar 17 19:05:41.548530 ignition[838]: GET result: OK Mar 17 19:05:41.548560 ignition[838]: failed to retrieve userdata from IMDS, falling back to custom data: not a config (empty) Mar 17 19:05:41.646435 ignition[838]: opening config device: "/dev/sr0" Mar 17 19:05:41.646875 ignition[838]: getting drive status for "/dev/sr0" Mar 17 19:05:41.646923 ignition[838]: drive status: OK Mar 17 19:05:41.646961 ignition[838]: mounting config device Mar 17 19:05:41.646977 ignition[838]: op(1): [started] mounting "/dev/sr0" at "/tmp/ignition-azure1150383048" Mar 17 19:05:41.674034 ignition[838]: op(1): [finished] mounting "/dev/sr0" at "/tmp/ignition-azure1150383048" Mar 17 19:05:41.677303 kernel: UDF-fs: INFO Mounting volume 'UDF Volume', timestamp 2025/03/18 00:00 (1000) Mar 17 19:05:41.676149 systemd[1]: tmp-ignition\x2dazure1150383048.mount: Deactivated successfully. Mar 17 19:05:41.675045 ignition[838]: checking for config drive Mar 17 19:05:41.675402 ignition[838]: reading config Mar 17 19:05:41.675759 ignition[838]: op(2): [started] unmounting "/dev/sr0" at "/tmp/ignition-azure1150383048" Mar 17 19:05:41.677152 ignition[838]: op(2): [finished] unmounting "/dev/sr0" at "/tmp/ignition-azure1150383048" Mar 17 19:05:41.681112 unknown[838]: fetched base config from "system" Mar 17 19:05:41.677171 ignition[838]: config has been read from custom data Mar 17 19:05:41.693000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.681130 unknown[838]: fetched base config from "system" Mar 17 19:05:41.677249 ignition[838]: parsing config with SHA512: 818beb764b4cffefb177ae2f187964f5e47ff0a7a819fd0a7f81b1a132a3cb06d2ba6d0304d4793f6aa57c3f86075acf573684a20d20dc0d84001188ce6cbf79 Mar 17 19:05:41.681136 unknown[838]: fetched user config from "azure" Mar 17 19:05:41.681698 ignition[838]: fetch: fetch complete Mar 17 19:05:41.688052 systemd[1]: Finished ignition-fetch.service. Mar 17 19:05:41.681702 ignition[838]: fetch: fetch passed Mar 17 19:05:41.694257 systemd[1]: Starting ignition-kargs.service... Mar 17 19:05:41.681744 ignition[838]: Ignition finished successfully Mar 17 19:05:41.718863 kernel: audit: type=1130 audit(1742238341.693:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.727811 ignition[846]: Ignition 2.14.0 Mar 17 19:05:41.727821 ignition[846]: Stage: kargs Mar 17 19:05:41.727947 ignition[846]: reading system config file "/usr/lib/ignition/base.d/base.ign" Mar 17 19:05:41.727974 ignition[846]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Mar 17 19:05:41.750213 ignition[846]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 17 19:05:41.753899 ignition[846]: kargs: kargs passed Mar 17 19:05:41.753959 ignition[846]: Ignition finished successfully Mar 17 19:05:41.755915 systemd[1]: Finished ignition-kargs.service. Mar 17 19:05:41.775487 kernel: audit: type=1130 audit(1742238341.759:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.759000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.768469 ignition[852]: Ignition 2.14.0 Mar 17 19:05:41.760502 systemd[1]: Starting ignition-disks.service... Mar 17 19:05:41.768475 ignition[852]: Stage: disks Mar 17 19:05:41.768588 ignition[852]: reading system config file "/usr/lib/ignition/base.d/base.ign" Mar 17 19:05:41.782000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.779661 systemd[1]: Finished ignition-disks.service. Mar 17 19:05:41.804184 kernel: audit: type=1130 audit(1742238341.782:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.768613 ignition[852]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Mar 17 19:05:41.782918 systemd[1]: Reached target initrd-root-device.target. Mar 17 19:05:41.776568 ignition[852]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 17 19:05:41.797724 systemd[1]: Reached target local-fs-pre.target. Mar 17 19:05:41.778822 ignition[852]: disks: disks passed Mar 17 19:05:41.804169 systemd[1]: Reached target local-fs.target. Mar 17 19:05:41.778862 ignition[852]: Ignition finished successfully Mar 17 19:05:41.805859 systemd[1]: Reached target sysinit.target. Mar 17 19:05:41.807822 systemd[1]: Reached target basic.target. Mar 17 19:05:41.810693 systemd[1]: Starting systemd-fsck-root.service... Mar 17 19:05:41.866992 systemd-fsck[860]: ROOT: clean, 623/7326000 files, 481078/7359488 blocks Mar 17 19:05:41.871041 systemd[1]: Finished systemd-fsck-root.service. Mar 17 19:05:41.875000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.876586 systemd[1]: Mounting sysroot.mount... Mar 17 19:05:41.888043 kernel: audit: type=1130 audit(1742238341.875:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:41.904080 kernel: EXT4-fs (sda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Mar 17 19:05:41.904197 systemd[1]: Mounted sysroot.mount. Mar 17 19:05:41.907664 systemd[1]: Reached target initrd-root-fs.target. Mar 17 19:05:41.994241 systemd[1]: Mounting sysroot-usr.mount... Mar 17 19:05:41.999976 systemd[1]: Starting flatcar-metadata-hostname.service... Mar 17 19:05:42.004624 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 17 19:05:42.004665 systemd[1]: Reached target ignition-diskful.target. Mar 17 19:05:42.011846 systemd[1]: Mounted sysroot-usr.mount. Mar 17 19:05:42.071300 systemd[1]: Mounting sysroot-usr-share-oem.mount... Mar 17 19:05:42.077329 systemd[1]: Starting initrd-setup-root.service... Mar 17 19:05:42.093084 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (870) Mar 17 19:05:42.093138 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 17 19:05:42.097661 initrd-setup-root[875]: cut: /sysroot/etc/passwd: No such file or directory Mar 17 19:05:42.107920 kernel: BTRFS info (device sda6): using free space tree Mar 17 19:05:42.107941 kernel: BTRFS info (device sda6): has skinny extents Mar 17 19:05:42.110768 systemd[1]: Mounted sysroot-usr-share-oem.mount. Mar 17 19:05:42.129991 initrd-setup-root[901]: cut: /sysroot/etc/group: No such file or directory Mar 17 19:05:42.145353 initrd-setup-root[909]: cut: /sysroot/etc/shadow: No such file or directory Mar 17 19:05:42.149991 initrd-setup-root[917]: cut: /sysroot/etc/gshadow: No such file or directory Mar 17 19:05:42.584031 systemd[1]: Finished initrd-setup-root.service. Mar 17 19:05:42.600980 kernel: audit: type=1130 audit(1742238342.586:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:42.586000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:42.598259 systemd[1]: Starting ignition-mount.service... Mar 17 19:05:42.606034 systemd[1]: Starting sysroot-boot.service... Mar 17 19:05:42.630635 systemd[1]: Finished sysroot-boot.service. Mar 17 19:05:42.647220 kernel: audit: type=1130 audit(1742238342.632:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:42.632000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:42.647307 ignition[936]: INFO : Ignition 2.14.0 Mar 17 19:05:42.647307 ignition[936]: INFO : Stage: mount Mar 17 19:05:42.647307 ignition[936]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Mar 17 19:05:42.647307 ignition[936]: DEBUG : parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Mar 17 19:05:42.647307 ignition[936]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 17 19:05:42.647307 ignition[936]: INFO : mount: mount passed Mar 17 19:05:42.647307 ignition[936]: INFO : Ignition finished successfully Mar 17 19:05:42.679122 kernel: audit: type=1130 audit(1742238342.649:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:42.649000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:42.647354 systemd[1]: Finished ignition-mount.service. Mar 17 19:05:42.675761 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. Mar 17 19:05:42.675835 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. Mar 17 19:05:43.379089 coreos-metadata[869]: Mar 17 19:05:43.378 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 17 19:05:43.395900 coreos-metadata[869]: Mar 17 19:05:43.395 INFO Fetch successful Mar 17 19:05:43.431635 coreos-metadata[869]: Mar 17 19:05:43.431 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Mar 17 19:05:43.446386 coreos-metadata[869]: Mar 17 19:05:43.446 INFO Fetch successful Mar 17 19:05:43.460683 coreos-metadata[869]: Mar 17 19:05:43.460 INFO wrote hostname ci-3510.3.7-a-a60daf7667 to /sysroot/etc/hostname Mar 17 19:05:43.462587 systemd[1]: Finished flatcar-metadata-hostname.service. Mar 17 19:05:43.467000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:43.479627 systemd[1]: Starting ignition-files.service... Mar 17 19:05:43.485398 kernel: audit: type=1130 audit(1742238343.467:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:43.491233 systemd[1]: Mounting sysroot-usr-share-oem.mount... Mar 17 19:05:43.504081 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (948) Mar 17 19:05:43.514385 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Mar 17 19:05:43.514441 kernel: BTRFS info (device sda6): using free space tree Mar 17 19:05:43.514454 kernel: BTRFS info (device sda6): has skinny extents Mar 17 19:05:43.523008 systemd[1]: Mounted sysroot-usr-share-oem.mount. Mar 17 19:05:43.537038 ignition[967]: INFO : Ignition 2.14.0 Mar 17 19:05:43.537038 ignition[967]: INFO : Stage: files Mar 17 19:05:43.541125 ignition[967]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Mar 17 19:05:43.541125 ignition[967]: DEBUG : parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Mar 17 19:05:43.554090 ignition[967]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 17 19:05:43.575720 ignition[967]: DEBUG : files: compiled without relabeling support, skipping Mar 17 19:05:43.578995 ignition[967]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 17 19:05:43.578995 ignition[967]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 17 19:05:43.662711 ignition[967]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 17 19:05:43.666900 ignition[967]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 17 19:05:43.675372 unknown[967]: wrote ssh authorized keys file for user: core Mar 17 19:05:43.678115 ignition[967]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 17 19:05:43.681906 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Mar 17 19:05:43.688484 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Mar 17 19:05:43.941540 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 17 19:05:44.111675 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Mar 17 19:05:44.116404 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 17 19:05:44.121157 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Mar 17 19:05:44.716861 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 17 19:05:45.019675 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 17 19:05:45.019675 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 17 19:05:45.031467 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 17 19:05:45.031467 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 17 19:05:45.040019 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 17 19:05:45.040019 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 17 19:05:45.048867 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 17 19:05:45.048867 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 17 19:05:45.057659 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 17 19:05:45.061841 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 19:05:45.066295 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 19:05:45.070503 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 17 19:05:45.076846 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 17 19:05:45.095722 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/etc/systemd/system/waagent.service" Mar 17 19:05:45.100535 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): oem config not found in "/usr/share/oem", looking on oem partition Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): op(c): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem178246072" Mar 17 19:05:45.111085 ignition[967]: CRITICAL : files: createFilesystemsFiles: createFiles: op(b): op(c): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem178246072": device or resource busy Mar 17 19:05:45.111085 ignition[967]: ERROR : files: createFilesystemsFiles: createFiles: op(b): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem178246072", trying btrfs: device or resource busy Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): op(d): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem178246072" Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): op(d): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem178246072" Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): op(e): [started] unmounting "/mnt/oem178246072" Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): op(e): [finished] unmounting "/mnt/oem178246072" Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/etc/systemd/system/waagent.service" Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/etc/systemd/system/nvidia.service" Mar 17 19:05:45.111085 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): oem config not found in "/usr/share/oem", looking on oem partition Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): op(10): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1399374092" Mar 17 19:05:45.162369 ignition[967]: CRITICAL : files: createFilesystemsFiles: createFiles: op(f): op(10): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1399374092": device or resource busy Mar 17 19:05:45.162369 ignition[967]: ERROR : files: createFilesystemsFiles: createFiles: op(f): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem1399374092", trying btrfs: device or resource busy Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): op(11): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1399374092" Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): op(11): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1399374092" Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): op(12): [started] unmounting "/mnt/oem1399374092" Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): op(12): [finished] unmounting "/mnt/oem1399374092" Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/etc/systemd/system/nvidia.service" Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(13): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 17 19:05:45.162369 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(13): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Mar 17 19:05:45.113530 systemd[1]: mnt-oem178246072.mount: Deactivated successfully. Mar 17 19:05:45.576539 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(13): GET result: OK Mar 17 19:05:45.935268 ignition[967]: INFO : files: createFilesystemsFiles: createFiles: op(13): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Mar 17 19:05:45.935268 ignition[967]: INFO : files: op(14): [started] processing unit "waagent.service" Mar 17 19:05:45.935268 ignition[967]: INFO : files: op(14): [finished] processing unit "waagent.service" Mar 17 19:05:45.935268 ignition[967]: INFO : files: op(15): [started] processing unit "nvidia.service" Mar 17 19:05:45.935268 ignition[967]: INFO : files: op(15): [finished] processing unit "nvidia.service" Mar 17 19:05:45.935268 ignition[967]: INFO : files: op(16): [started] processing unit "prepare-helm.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(16): op(17): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(16): op(17): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(16): [finished] processing unit "prepare-helm.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(18): [started] setting preset to enabled for "prepare-helm.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(18): [finished] setting preset to enabled for "prepare-helm.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(19): [started] setting preset to enabled for "waagent.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(19): [finished] setting preset to enabled for "waagent.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(1a): [started] setting preset to enabled for "nvidia.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: op(1a): [finished] setting preset to enabled for "nvidia.service" Mar 17 19:05:45.955759 ignition[967]: INFO : files: createResultFile: createFiles: op(1b): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 17 19:05:45.955759 ignition[967]: INFO : files: createResultFile: createFiles: op(1b): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 17 19:05:45.955759 ignition[967]: INFO : files: files passed Mar 17 19:05:45.955759 ignition[967]: INFO : Ignition finished successfully Mar 17 19:05:46.022516 kernel: audit: type=1130 audit(1742238345.962:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:45.962000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.002000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.002000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:45.959932 systemd[1]: Finished ignition-files.service. Mar 17 19:05:46.026000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:45.964029 systemd[1]: Starting initrd-setup-root-after-ignition.service... Mar 17 19:05:46.028111 initrd-setup-root-after-ignition[992]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 17 19:05:45.981312 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Mar 17 19:05:45.982239 systemd[1]: Starting ignition-quench.service... Mar 17 19:05:45.989952 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 17 19:05:45.990049 systemd[1]: Finished ignition-quench.service. Mar 17 19:05:46.022565 systemd[1]: Finished initrd-setup-root-after-ignition.service. Mar 17 19:05:46.027190 systemd[1]: Reached target ignition-complete.target. Mar 17 19:05:46.050123 systemd[1]: Starting initrd-parse-etc.service... Mar 17 19:05:46.063872 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 17 19:05:46.066000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.066000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.063955 systemd[1]: Finished initrd-parse-etc.service. Mar 17 19:05:46.066254 systemd[1]: Reached target initrd-fs.target. Mar 17 19:05:46.070774 systemd[1]: Reached target initrd.target. Mar 17 19:05:46.072718 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Mar 17 19:05:46.073646 systemd[1]: Starting dracut-pre-pivot.service... Mar 17 19:05:46.091010 systemd[1]: Finished dracut-pre-pivot.service. Mar 17 19:05:46.095000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.096007 systemd[1]: Starting initrd-cleanup.service... Mar 17 19:05:46.105886 systemd[1]: Stopped target nss-lookup.target. Mar 17 19:05:46.110053 systemd[1]: Stopped target remote-cryptsetup.target. Mar 17 19:05:46.114353 systemd[1]: Stopped target timers.target. Mar 17 19:05:46.118334 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 17 19:05:46.118469 systemd[1]: Stopped dracut-pre-pivot.service. Mar 17 19:05:46.122000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.122782 systemd[1]: Stopped target initrd.target. Mar 17 19:05:46.126454 systemd[1]: Stopped target basic.target. Mar 17 19:05:46.130154 systemd[1]: Stopped target ignition-complete.target. Mar 17 19:05:46.136246 systemd[1]: Stopped target ignition-diskful.target. Mar 17 19:05:46.140468 systemd[1]: Stopped target initrd-root-device.target. Mar 17 19:05:46.145100 systemd[1]: Stopped target remote-fs.target. Mar 17 19:05:46.149166 systemd[1]: Stopped target remote-fs-pre.target. Mar 17 19:05:46.153011 systemd[1]: Stopped target sysinit.target. Mar 17 19:05:46.156808 systemd[1]: Stopped target local-fs.target. Mar 17 19:05:46.160735 systemd[1]: Stopped target local-fs-pre.target. Mar 17 19:05:46.164826 systemd[1]: Stopped target swap.target. Mar 17 19:05:46.168337 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 17 19:05:46.171000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.168491 systemd[1]: Stopped dracut-pre-mount.service. Mar 17 19:05:46.172197 systemd[1]: Stopped target cryptsetup.target. Mar 17 19:05:46.180000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.184000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.176919 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 17 19:05:46.177082 systemd[1]: Stopped dracut-initqueue.service. Mar 17 19:05:46.189000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.180918 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 17 19:05:46.193000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.181055 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Mar 17 19:05:46.211000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.213000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.220461 ignition[1005]: INFO : Ignition 2.14.0 Mar 17 19:05:46.220461 ignition[1005]: INFO : Stage: umount Mar 17 19:05:46.220461 ignition[1005]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Mar 17 19:05:46.220461 ignition[1005]: DEBUG : parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Mar 17 19:05:46.224000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.230000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.230000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.185002 systemd[1]: ignition-files.service: Deactivated successfully. Mar 17 19:05:46.243000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.243245 ignition[1005]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 17 19:05:46.243245 ignition[1005]: INFO : umount: umount passed Mar 17 19:05:46.243245 ignition[1005]: INFO : Ignition finished successfully Mar 17 19:05:46.248000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.253000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.255000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.262000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.185271 systemd[1]: Stopped ignition-files.service. Mar 17 19:05:46.189310 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 17 19:05:46.189436 systemd[1]: Stopped flatcar-metadata-hostname.service. Mar 17 19:05:46.194467 systemd[1]: Stopping ignition-mount.service... Mar 17 19:05:46.198364 systemd[1]: Stopping iscsiuio.service... Mar 17 19:05:46.201202 systemd[1]: Stopping sysroot-boot.service... Mar 17 19:05:46.204025 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 17 19:05:46.204237 systemd[1]: Stopped systemd-udev-trigger.service. Mar 17 19:05:46.211219 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 17 19:05:46.211370 systemd[1]: Stopped dracut-pre-trigger.service. Mar 17 19:05:46.215659 systemd[1]: iscsiuio.service: Deactivated successfully. Mar 17 19:05:46.215775 systemd[1]: Stopped iscsiuio.service. Mar 17 19:05:46.225935 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 17 19:05:46.226021 systemd[1]: Finished initrd-cleanup.service. Mar 17 19:05:46.230707 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 17 19:05:46.230789 systemd[1]: Stopped ignition-mount.service. Mar 17 19:05:46.244010 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 17 19:05:46.244068 systemd[1]: Stopped ignition-disks.service. Mar 17 19:05:46.248403 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 17 19:05:46.250373 systemd[1]: Stopped ignition-kargs.service. Mar 17 19:05:46.253374 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 17 19:05:46.253424 systemd[1]: Stopped ignition-fetch.service. Mar 17 19:05:46.255552 systemd[1]: Stopped target network.target. Mar 17 19:05:46.257669 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 17 19:05:46.257727 systemd[1]: Stopped ignition-fetch-offline.service. Mar 17 19:05:46.262140 systemd[1]: Stopped target paths.target. Mar 17 19:05:46.265735 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 17 19:05:46.270535 systemd[1]: Stopped systemd-ask-password-console.path. Mar 17 19:05:46.323249 systemd[1]: Stopped target slices.target. Mar 17 19:05:46.327056 systemd[1]: Stopped target sockets.target. Mar 17 19:05:46.331053 systemd[1]: iscsid.socket: Deactivated successfully. Mar 17 19:05:46.331113 systemd[1]: Closed iscsid.socket. Mar 17 19:05:46.336218 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 17 19:05:46.336276 systemd[1]: Closed iscsiuio.socket. Mar 17 19:05:46.341770 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 17 19:05:46.344014 systemd[1]: Stopped ignition-setup.service. Mar 17 19:05:46.347000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.347850 systemd[1]: Stopping systemd-networkd.service... Mar 17 19:05:46.353525 systemd[1]: Stopping systemd-resolved.service... Mar 17 19:05:46.358148 systemd-networkd[805]: eth0: DHCPv6 lease lost Mar 17 19:05:46.359594 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 17 19:05:46.364180 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 17 19:05:46.364294 systemd[1]: Stopped systemd-resolved.service. Mar 17 19:05:46.369000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.371775 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 17 19:05:46.371897 systemd[1]: Stopped systemd-networkd.service. Mar 17 19:05:46.378000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.378000 audit: BPF prog-id=6 op=UNLOAD Mar 17 19:05:46.378562 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 17 19:05:46.378617 systemd[1]: Closed systemd-networkd.socket. Mar 17 19:05:46.383826 systemd[1]: Stopping network-cleanup.service... Mar 17 19:05:46.383000 audit: BPF prog-id=9 op=UNLOAD Mar 17 19:05:46.391000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.387163 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 17 19:05:46.395000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.387217 systemd[1]: Stopped parse-ip-for-networkd.service. Mar 17 19:05:46.399000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.391307 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 17 19:05:46.391359 systemd[1]: Stopped systemd-sysctl.service. Mar 17 19:05:46.395935 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 17 19:05:46.409000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.395986 systemd[1]: Stopped systemd-modules-load.service. Mar 17 19:05:46.399905 systemd[1]: Stopping systemd-udevd.service... Mar 17 19:05:46.406652 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 17 19:05:46.407184 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 17 19:05:46.407318 systemd[1]: Stopped systemd-udevd.service. Mar 17 19:05:46.427000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.429000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.412511 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 17 19:05:46.442113 kernel: kauditd_printk_skb: 33 callbacks suppressed Mar 17 19:05:46.442152 kernel: audit: type=1131 audit(1742238346.434:72): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.434000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.412559 systemd[1]: Closed systemd-udevd-control.socket. Mar 17 19:05:46.419244 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 17 19:05:46.463814 kernel: hv_netvsc 7c1e52fc-a1da-7c1e-52fc-a1da7c1e52fc eth0: Data path switched from VF: enP55511s1 Mar 17 19:05:46.467241 kernel: audit: type=1131 audit(1742238346.466:73): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.466000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.419288 systemd[1]: Closed systemd-udevd-kernel.socket. Mar 17 19:05:46.423626 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 17 19:05:46.423690 systemd[1]: Stopped dracut-pre-udev.service. Mar 17 19:05:46.427505 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 17 19:05:46.427554 systemd[1]: Stopped dracut-cmdline.service. Mar 17 19:05:46.429709 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 17 19:05:46.429760 systemd[1]: Stopped dracut-cmdline-ask.service. Mar 17 19:05:46.434952 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Mar 17 19:05:46.454982 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 17 19:05:46.455055 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Mar 17 19:05:46.480307 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 17 19:05:46.482561 systemd[1]: Stopped kmod-static-nodes.service. Mar 17 19:05:46.504139 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 19:05:46.504212 systemd[1]: Stopped systemd-vconsole-setup.service. Mar 17 19:05:46.503000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.511784 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 17 19:05:46.540111 kernel: audit: type=1131 audit(1742238346.503:74): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.540144 kernel: audit: type=1131 audit(1742238346.510:75): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.510000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.512360 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 17 19:05:46.565634 kernel: audit: type=1130 audit(1742238346.539:76): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.565666 kernel: audit: type=1131 audit(1742238346.540:77): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.539000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.540000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.512447 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Mar 17 19:05:46.569055 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 17 19:05:46.571000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.569172 systemd[1]: Stopped network-cleanup.service. Mar 17 19:05:46.588710 kernel: audit: type=1131 audit(1742238346.571:78): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.969368 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 17 19:05:46.969480 systemd[1]: Stopped sysroot-boot.service. Mar 17 19:05:46.973000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.973933 systemd[1]: Reached target initrd-switch-root.target. Mar 17 19:05:46.990083 kernel: audit: type=1131 audit(1742238346.973:79): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.987750 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 17 19:05:46.987823 systemd[1]: Stopped initrd-setup-root.service. Mar 17 19:05:47.009209 kernel: audit: type=1131 audit(1742238346.994:80): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:46.994000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:47.007141 systemd[1]: Starting initrd-switch-root.service... Mar 17 19:05:47.025113 systemd[1]: Switching root. Mar 17 19:05:47.053793 systemd-journald[183]: Received SIGTERM from PID 1 (systemd). Mar 17 19:05:47.053887 iscsid[815]: iscsid shutting down. Mar 17 19:05:47.056285 systemd-journald[183]: Journal stopped Mar 17 19:06:02.552132 kernel: SELinux: Class mctp_socket not defined in policy. Mar 17 19:06:02.552159 kernel: SELinux: Class anon_inode not defined in policy. Mar 17 19:06:02.552174 kernel: SELinux: the above unknown classes and permissions will be allowed Mar 17 19:06:02.552182 kernel: SELinux: policy capability network_peer_controls=1 Mar 17 19:06:02.552191 kernel: SELinux: policy capability open_perms=1 Mar 17 19:06:02.552202 kernel: SELinux: policy capability extended_socket_class=1 Mar 17 19:06:02.552214 kernel: SELinux: policy capability always_check_network=0 Mar 17 19:06:02.552224 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 17 19:06:02.552236 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 17 19:06:02.552244 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 17 19:06:02.552255 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 17 19:06:02.552263 kernel: audit: type=1403 audit(1742238349.877:81): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 17 19:06:02.552277 systemd[1]: Successfully loaded SELinux policy in 350.800ms. Mar 17 19:06:02.552291 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 21.046ms. Mar 17 19:06:02.552306 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Mar 17 19:06:02.552317 systemd[1]: Detected virtualization microsoft. Mar 17 19:06:02.552330 systemd[1]: Detected architecture x86-64. Mar 17 19:06:02.552339 systemd[1]: Detected first boot. Mar 17 19:06:02.552354 systemd[1]: Hostname set to . Mar 17 19:06:02.552366 systemd[1]: Initializing machine ID from random generator. Mar 17 19:06:02.552375 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Mar 17 19:06:02.552387 kernel: kauditd_printk_skb: 7 callbacks suppressed Mar 17 19:06:02.552399 kernel: audit: type=1400 audit(1742238351.994:89): avc: denied { associate } for pid=1038 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Mar 17 19:06:02.552409 kernel: audit: type=1300 audit(1742238351.994:89): arch=c000003e syscall=188 success=yes exit=0 a0=c0000247fc a1=c00002aae0 a2=c000028d00 a3=32 items=0 ppid=1021 pid=1038 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 19:06:02.552421 kernel: audit: type=1327 audit(1742238351.994:89): proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Mar 17 19:06:02.552435 kernel: audit: type=1400 audit(1742238352.000:90): avc: denied { associate } for pid=1038 comm="torcx-generator" name="lib" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Mar 17 19:06:02.552445 kernel: audit: type=1300 audit(1742238352.000:90): arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c0000248d5 a2=1ed a3=0 items=2 ppid=1021 pid=1038 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 19:06:02.552457 kernel: audit: type=1307 audit(1742238352.000:90): cwd="/" Mar 17 19:06:02.552469 kernel: audit: type=1302 audit(1742238352.000:90): item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:02.552479 kernel: audit: type=1302 audit(1742238352.000:90): item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:02.552490 kernel: audit: type=1327 audit(1742238352.000:90): proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Mar 17 19:06:02.552503 systemd[1]: Populated /etc with preset unit settings. Mar 17 19:06:02.552516 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 19:06:02.552526 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 19:06:02.552540 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 19:06:02.552552 kernel: audit: type=1334 audit(1742238361.924:91): prog-id=12 op=LOAD Mar 17 19:06:02.552561 kernel: audit: type=1334 audit(1742238361.924:92): prog-id=3 op=UNLOAD Mar 17 19:06:02.552572 kernel: audit: type=1334 audit(1742238361.930:93): prog-id=13 op=LOAD Mar 17 19:06:02.552584 kernel: audit: type=1334 audit(1742238361.942:94): prog-id=14 op=LOAD Mar 17 19:06:02.552594 kernel: audit: type=1334 audit(1742238361.942:95): prog-id=4 op=UNLOAD Mar 17 19:06:02.552606 kernel: audit: type=1334 audit(1742238361.942:96): prog-id=5 op=UNLOAD Mar 17 19:06:02.552618 kernel: audit: type=1334 audit(1742238361.952:97): prog-id=15 op=LOAD Mar 17 19:06:02.552630 kernel: audit: type=1334 audit(1742238361.952:98): prog-id=12 op=UNLOAD Mar 17 19:06:02.552639 kernel: audit: type=1334 audit(1742238361.957:99): prog-id=16 op=LOAD Mar 17 19:06:02.552651 kernel: audit: type=1334 audit(1742238361.969:100): prog-id=17 op=LOAD Mar 17 19:06:02.552662 systemd[1]: iscsid.service: Deactivated successfully. Mar 17 19:06:02.552673 systemd[1]: Stopped iscsid.service. Mar 17 19:06:02.552688 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 17 19:06:02.552700 systemd[1]: Stopped initrd-switch-root.service. Mar 17 19:06:02.552711 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 17 19:06:02.552723 systemd[1]: Created slice system-addon\x2dconfig.slice. Mar 17 19:06:02.552735 systemd[1]: Created slice system-addon\x2drun.slice. Mar 17 19:06:02.552746 systemd[1]: Created slice system-getty.slice. Mar 17 19:06:02.552757 systemd[1]: Created slice system-modprobe.slice. Mar 17 19:06:02.552772 systemd[1]: Created slice system-serial\x2dgetty.slice. Mar 17 19:06:02.552784 systemd[1]: Created slice system-system\x2dcloudinit.slice. Mar 17 19:06:02.552796 systemd[1]: Created slice system-systemd\x2dfsck.slice. Mar 17 19:06:02.552807 systemd[1]: Created slice user.slice. Mar 17 19:06:02.552818 systemd[1]: Started systemd-ask-password-console.path. Mar 17 19:06:02.552829 systemd[1]: Started systemd-ask-password-wall.path. Mar 17 19:06:02.552840 systemd[1]: Set up automount boot.automount. Mar 17 19:06:02.552853 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Mar 17 19:06:02.552863 systemd[1]: Stopped target initrd-switch-root.target. Mar 17 19:06:02.552877 systemd[1]: Stopped target initrd-fs.target. Mar 17 19:06:02.552890 systemd[1]: Stopped target initrd-root-fs.target. Mar 17 19:06:02.552899 systemd[1]: Reached target integritysetup.target. Mar 17 19:06:02.552913 systemd[1]: Reached target remote-cryptsetup.target. Mar 17 19:06:02.552926 systemd[1]: Reached target remote-fs.target. Mar 17 19:06:02.552935 systemd[1]: Reached target slices.target. Mar 17 19:06:02.552948 systemd[1]: Reached target swap.target. Mar 17 19:06:02.552961 systemd[1]: Reached target torcx.target. Mar 17 19:06:02.552973 systemd[1]: Reached target veritysetup.target. Mar 17 19:06:02.552985 systemd[1]: Listening on systemd-coredump.socket. Mar 17 19:06:02.553000 systemd[1]: Listening on systemd-initctl.socket. Mar 17 19:06:02.553016 systemd[1]: Listening on systemd-networkd.socket. Mar 17 19:06:02.553036 systemd[1]: Listening on systemd-udevd-control.socket. Mar 17 19:06:02.553057 systemd[1]: Listening on systemd-udevd-kernel.socket. Mar 17 19:06:02.553087 systemd[1]: Listening on systemd-userdbd.socket. Mar 17 19:06:02.553109 systemd[1]: Mounting dev-hugepages.mount... Mar 17 19:06:02.553132 systemd[1]: Mounting dev-mqueue.mount... Mar 17 19:06:02.553150 systemd[1]: Mounting media.mount... Mar 17 19:06:02.553169 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:02.553188 systemd[1]: Mounting sys-kernel-debug.mount... Mar 17 19:06:02.553210 systemd[1]: Mounting sys-kernel-tracing.mount... Mar 17 19:06:02.553230 systemd[1]: Mounting tmp.mount... Mar 17 19:06:02.553252 systemd[1]: Starting flatcar-tmpfiles.service... Mar 17 19:06:02.553274 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 19:06:02.553294 systemd[1]: Starting kmod-static-nodes.service... Mar 17 19:06:02.553312 systemd[1]: Starting modprobe@configfs.service... Mar 17 19:06:02.553331 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 19:06:02.553354 systemd[1]: Starting modprobe@drm.service... Mar 17 19:06:02.553372 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 19:06:02.553391 systemd[1]: Starting modprobe@fuse.service... Mar 17 19:06:02.553409 systemd[1]: Starting modprobe@loop.service... Mar 17 19:06:02.553433 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 17 19:06:02.553451 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 17 19:06:02.553469 systemd[1]: Stopped systemd-fsck-root.service. Mar 17 19:06:02.553486 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 17 19:06:02.553507 systemd[1]: Stopped systemd-fsck-usr.service. Mar 17 19:06:02.553525 systemd[1]: Stopped systemd-journald.service. Mar 17 19:06:02.553544 systemd[1]: Starting systemd-journald.service... Mar 17 19:06:02.553563 kernel: loop: module loaded Mar 17 19:06:02.553583 systemd[1]: Starting systemd-modules-load.service... Mar 17 19:06:02.553601 systemd[1]: Starting systemd-network-generator.service... Mar 17 19:06:02.553621 systemd[1]: Starting systemd-remount-fs.service... Mar 17 19:06:02.553639 systemd[1]: Starting systemd-udev-trigger.service... Mar 17 19:06:02.553658 systemd[1]: verity-setup.service: Deactivated successfully. Mar 17 19:06:02.553678 systemd[1]: Stopped verity-setup.service. Mar 17 19:06:02.553697 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:02.553718 kernel: fuse: init (API version 7.34) Mar 17 19:06:02.553735 systemd[1]: Mounted dev-hugepages.mount. Mar 17 19:06:02.553758 systemd[1]: Mounted dev-mqueue.mount. Mar 17 19:06:02.553777 systemd[1]: Mounted media.mount. Mar 17 19:06:02.553799 systemd[1]: Mounted sys-kernel-debug.mount. Mar 17 19:06:02.553819 systemd[1]: Mounted sys-kernel-tracing.mount. Mar 17 19:06:02.553836 systemd[1]: Mounted tmp.mount. Mar 17 19:06:02.553853 systemd[1]: Finished flatcar-tmpfiles.service. Mar 17 19:06:02.553869 systemd[1]: Finished kmod-static-nodes.service. Mar 17 19:06:02.553898 systemd-journald[1132]: Journal started Mar 17 19:06:02.555680 systemd-journald[1132]: Runtime Journal (/run/log/journal/71675666f05f4068b737dc548a803496) is 8.0M, max 159.0M, 151.0M free. Mar 17 19:05:49.877000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 17 19:05:50.558000 audit[1]: AVC avc: denied { integrity } for pid=1 comm="systemd" lockdown_reason="/dev/mem,kmem,port" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Mar 17 19:05:50.571000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Mar 17 19:05:50.571000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Mar 17 19:05:50.571000 audit: BPF prog-id=10 op=LOAD Mar 17 19:05:50.572000 audit: BPF prog-id=10 op=UNLOAD Mar 17 19:05:50.572000 audit: BPF prog-id=11 op=LOAD Mar 17 19:05:50.572000 audit: BPF prog-id=11 op=UNLOAD Mar 17 19:05:51.994000 audit[1038]: AVC avc: denied { associate } for pid=1038 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Mar 17 19:05:51.994000 audit[1038]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c0000247fc a1=c00002aae0 a2=c000028d00 a3=32 items=0 ppid=1021 pid=1038 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 19:05:51.994000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Mar 17 19:05:52.000000 audit[1038]: AVC avc: denied { associate } for pid=1038 comm="torcx-generator" name="lib" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Mar 17 19:05:52.000000 audit[1038]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c0000248d5 a2=1ed a3=0 items=2 ppid=1021 pid=1038 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 19:05:52.000000 audit: CWD cwd="/" Mar 17 19:05:52.000000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:05:52.000000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:05:52.000000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Mar 17 19:06:01.924000 audit: BPF prog-id=12 op=LOAD Mar 17 19:06:01.924000 audit: BPF prog-id=3 op=UNLOAD Mar 17 19:06:01.930000 audit: BPF prog-id=13 op=LOAD Mar 17 19:06:01.942000 audit: BPF prog-id=14 op=LOAD Mar 17 19:06:01.942000 audit: BPF prog-id=4 op=UNLOAD Mar 17 19:06:01.942000 audit: BPF prog-id=5 op=UNLOAD Mar 17 19:06:01.952000 audit: BPF prog-id=15 op=LOAD Mar 17 19:06:01.952000 audit: BPF prog-id=12 op=UNLOAD Mar 17 19:06:01.957000 audit: BPF prog-id=16 op=LOAD Mar 17 19:06:01.969000 audit: BPF prog-id=17 op=LOAD Mar 17 19:06:01.969000 audit: BPF prog-id=13 op=UNLOAD Mar 17 19:06:01.969000 audit: BPF prog-id=14 op=UNLOAD Mar 17 19:06:01.979000 audit: BPF prog-id=18 op=LOAD Mar 17 19:06:01.979000 audit: BPF prog-id=15 op=UNLOAD Mar 17 19:06:01.980000 audit: BPF prog-id=19 op=LOAD Mar 17 19:06:01.980000 audit: BPF prog-id=20 op=LOAD Mar 17 19:06:01.980000 audit: BPF prog-id=16 op=UNLOAD Mar 17 19:06:01.980000 audit: BPF prog-id=17 op=UNLOAD Mar 17 19:06:01.980000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:01.991000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:01.996000 audit: BPF prog-id=18 op=UNLOAD Mar 17 19:06:02.002000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.002000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.341000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.354000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.360000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.360000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.362000 audit: BPF prog-id=21 op=LOAD Mar 17 19:06:02.362000 audit: BPF prog-id=22 op=LOAD Mar 17 19:06:02.362000 audit: BPF prog-id=23 op=LOAD Mar 17 19:06:02.362000 audit: BPF prog-id=19 op=UNLOAD Mar 17 19:06:02.362000 audit: BPF prog-id=20 op=UNLOAD Mar 17 19:06:02.450000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.544000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Mar 17 19:06:02.544000 audit[1132]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=4 a1=7ffc2d6fb8e0 a2=4000 a3=7ffc2d6fb97c items=0 ppid=1 pid=1132 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 19:06:02.544000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Mar 17 19:06:02.550000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:51.884030 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 19:06:01.923255 systemd[1]: Queued start job for default target multi-user.target. Mar 17 19:05:51.964142 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Mar 17 19:06:01.923267 systemd[1]: Unnecessary job was removed for dev-sda6.device. Mar 17 19:05:51.964185 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Mar 17 19:06:01.980755 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 17 19:05:51.964232 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Mar 17 19:05:51.964247 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="skipped missing lower profile" missing profile=oem Mar 17 19:05:51.964296 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Mar 17 19:05:51.964317 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Mar 17 19:05:51.964557 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Mar 17 19:06:02.556000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:05:51.964605 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Mar 17 19:05:51.964619 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Mar 17 19:05:51.982352 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Mar 17 19:05:51.982412 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Mar 17 19:05:51.982460 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.7: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.7 Mar 17 19:05:51.982485 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Mar 17 19:05:51.982513 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.7: no such file or directory" path=/var/lib/torcx/store/3510.3.7 Mar 17 19:05:51.982528 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:05:51Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Mar 17 19:06:00.395609 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:06:00Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 19:06:00.395837 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:06:00Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 19:06:00.395933 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:06:00Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 19:06:00.396113 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:06:00Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 19:06:00.396159 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:06:00Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Mar 17 19:06:00.396221 /usr/lib/systemd/system-generators/torcx-generator[1038]: time="2025-03-17T19:06:00Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Mar 17 19:06:02.560123 systemd[1]: Started systemd-journald.service. Mar 17 19:06:02.561000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.562608 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 17 19:06:02.562771 systemd[1]: Finished modprobe@configfs.service. Mar 17 19:06:02.565000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.565000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.565980 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 19:06:02.566157 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 19:06:02.570000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.570000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.570477 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 19:06:02.570627 systemd[1]: Finished modprobe@drm.service. Mar 17 19:06:02.574000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.574000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.574953 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 19:06:02.575143 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 19:06:02.580000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.580000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.580424 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 17 19:06:02.580576 systemd[1]: Finished modprobe@fuse.service. Mar 17 19:06:02.585000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.586000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.586400 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 19:06:02.586551 systemd[1]: Finished modprobe@loop.service. Mar 17 19:06:02.590000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.590000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.590762 systemd[1]: Finished systemd-network-generator.service. Mar 17 19:06:02.595000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.596372 systemd[1]: Finished systemd-modules-load.service. Mar 17 19:06:02.600000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.600625 systemd[1]: Finished systemd-remount-fs.service. Mar 17 19:06:02.604000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.605443 systemd[1]: Reached target network-pre.target. Mar 17 19:06:02.611308 systemd[1]: Mounting sys-fs-fuse-connections.mount... Mar 17 19:06:02.618888 systemd[1]: Mounting sys-kernel-config.mount... Mar 17 19:06:02.628226 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 17 19:06:02.644473 systemd[1]: Starting systemd-hwdb-update.service... Mar 17 19:06:02.648244 systemd[1]: Starting systemd-journal-flush.service... Mar 17 19:06:02.650755 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 19:06:02.652386 systemd[1]: Starting systemd-random-seed.service... Mar 17 19:06:02.654767 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 19:06:02.656242 systemd[1]: Starting systemd-sysctl.service... Mar 17 19:06:02.660734 systemd[1]: Starting systemd-sysusers.service... Mar 17 19:06:02.667401 systemd[1]: Mounted sys-fs-fuse-connections.mount. Mar 17 19:06:02.671957 systemd[1]: Mounted sys-kernel-config.mount. Mar 17 19:06:02.676281 systemd[1]: Finished systemd-udev-trigger.service. Mar 17 19:06:02.680000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.687341 systemd[1]: Starting systemd-udev-settle.service... Mar 17 19:06:02.694406 udevadm[1162]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Mar 17 19:06:02.730579 systemd-journald[1132]: Time spent on flushing to /var/log/journal/71675666f05f4068b737dc548a803496 is 19.578ms for 1190 entries. Mar 17 19:06:02.730579 systemd-journald[1132]: System Journal (/var/log/journal/71675666f05f4068b737dc548a803496) is 8.0M, max 2.6G, 2.6G free. Mar 17 19:06:02.823433 systemd-journald[1132]: Received client request to flush runtime journal. Mar 17 19:06:02.773000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.769732 systemd[1]: Finished systemd-random-seed.service. Mar 17 19:06:02.774042 systemd[1]: Reached target first-boot-complete.target. Mar 17 19:06:02.824721 systemd[1]: Finished systemd-journal-flush.service. Mar 17 19:06:02.829000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:02.873081 systemd[1]: Finished systemd-sysctl.service. Mar 17 19:06:02.877000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:03.517126 systemd[1]: Finished systemd-sysusers.service. Mar 17 19:06:03.521000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:03.522580 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Mar 17 19:06:03.937000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:03.932680 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Mar 17 19:06:03.942935 systemd[1]: Finished systemd-hwdb-update.service. Mar 17 19:06:03.945000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:03.945000 audit: BPF prog-id=24 op=LOAD Mar 17 19:06:03.945000 audit: BPF prog-id=25 op=LOAD Mar 17 19:06:03.945000 audit: BPF prog-id=7 op=UNLOAD Mar 17 19:06:03.945000 audit: BPF prog-id=8 op=UNLOAD Mar 17 19:06:03.946678 systemd[1]: Starting systemd-udevd.service... Mar 17 19:06:03.968458 systemd-udevd[1167]: Using default interface naming scheme 'v252'. Mar 17 19:06:04.241702 systemd[1]: Started systemd-udevd.service. Mar 17 19:06:04.246000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:04.250000 audit: BPF prog-id=26 op=LOAD Mar 17 19:06:04.251367 systemd[1]: Starting systemd-networkd.service... Mar 17 19:06:04.285160 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Mar 17 19:06:04.333135 kernel: mousedev: PS/2 mouse device common for all mice Mar 17 19:06:04.350000 audit: BPF prog-id=27 op=LOAD Mar 17 19:06:04.350000 audit: BPF prog-id=28 op=LOAD Mar 17 19:06:04.350000 audit: BPF prog-id=29 op=LOAD Mar 17 19:06:04.351620 systemd[1]: Starting systemd-userdbd.service... Mar 17 19:06:04.362000 audit[1182]: AVC avc: denied { confidentiality } for pid=1182 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Mar 17 19:06:04.369086 kernel: hv_vmbus: registering driver hv_balloon Mar 17 19:06:04.378086 kernel: hv_vmbus: registering driver hyperv_fb Mar 17 19:06:04.398663 kernel: hv_utils: Registering HyperV Utility Driver Mar 17 19:06:04.398758 kernel: hv_vmbus: registering driver hv_utils Mar 17 19:06:04.408131 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Mar 17 19:06:04.410000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:04.408469 systemd[1]: Started systemd-userdbd.service. Mar 17 19:06:04.362000 audit[1182]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=5647df0f7210 a1=f884 a2=7f5e01ac4bc5 a3=5 items=12 ppid=1167 pid=1182 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 19:06:04.362000 audit: CWD cwd="/" Mar 17 19:06:04.362000 audit: PATH item=0 name=(null) inode=1237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=1 name=(null) inode=14247 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=2 name=(null) inode=14247 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=3 name=(null) inode=14248 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=4 name=(null) inode=14247 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=5 name=(null) inode=14249 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=6 name=(null) inode=14247 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=7 name=(null) inode=14250 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=8 name=(null) inode=14247 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=9 name=(null) inode=14251 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=10 name=(null) inode=14247 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PATH item=11 name=(null) inode=14252 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 19:06:04.362000 audit: PROCTITLE proctitle="(udev-worker)" Mar 17 19:06:04.489674 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Mar 17 19:06:04.489775 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Mar 17 19:06:04.494452 kernel: Console: switching to colour dummy device 80x25 Mar 17 19:06:04.495082 kernel: hv_utils: Shutdown IC version 3.2 Mar 17 19:06:04.495155 kernel: hv_utils: Heartbeat IC version 3.0 Mar 17 19:06:04.495180 kernel: hv_utils: TimeSync IC version 4.0 Mar 17 19:06:04.927396 kernel: Console: switching to colour frame buffer device 128x48 Mar 17 19:06:05.075891 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Mar 17 19:06:05.129398 kernel: KVM: vmx: using Hyper-V Enlightened VMCS Mar 17 19:06:05.192786 systemd[1]: Finished systemd-udev-settle.service. Mar 17 19:06:05.193000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:05.196646 systemd[1]: Starting lvm2-activation-early.service... Mar 17 19:06:05.317778 systemd-networkd[1180]: lo: Link UP Mar 17 19:06:05.317789 systemd-networkd[1180]: lo: Gained carrier Mar 17 19:06:05.318438 systemd-networkd[1180]: Enumeration completed Mar 17 19:06:05.318574 systemd[1]: Started systemd-networkd.service. Mar 17 19:06:05.320000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:05.322289 systemd[1]: Starting systemd-networkd-wait-online.service... Mar 17 19:06:05.337207 systemd-networkd[1180]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 19:06:05.388404 kernel: mlx5_core d8d7:00:02.0 enP55511s1: Link up Mar 17 19:06:05.411414 kernel: hv_netvsc 7c1e52fc-a1da-7c1e-52fc-a1da7c1e52fc eth0: Data path switched to VF: enP55511s1 Mar 17 19:06:05.412464 systemd-networkd[1180]: enP55511s1: Link UP Mar 17 19:06:05.412666 systemd-networkd[1180]: eth0: Link UP Mar 17 19:06:05.412672 systemd-networkd[1180]: eth0: Gained carrier Mar 17 19:06:05.416672 systemd-networkd[1180]: enP55511s1: Gained carrier Mar 17 19:06:05.432540 systemd-networkd[1180]: eth0: DHCPv4 address 10.200.8.15/24, gateway 10.200.8.1 acquired from 168.63.129.16 Mar 17 19:06:05.497852 lvm[1243]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 19:06:05.524497 systemd[1]: Finished lvm2-activation-early.service. Mar 17 19:06:05.525000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:05.527086 systemd[1]: Reached target cryptsetup.target. Mar 17 19:06:05.530646 systemd[1]: Starting lvm2-activation.service... Mar 17 19:06:05.534768 lvm[1245]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 19:06:05.553401 systemd[1]: Finished lvm2-activation.service. Mar 17 19:06:05.554000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:05.555641 systemd[1]: Reached target local-fs-pre.target. Mar 17 19:06:05.557650 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 17 19:06:05.557681 systemd[1]: Reached target local-fs.target. Mar 17 19:06:05.559643 systemd[1]: Reached target machines.target. Mar 17 19:06:05.562850 systemd[1]: Starting ldconfig.service... Mar 17 19:06:05.565028 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 19:06:05.565125 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 19:06:05.566313 systemd[1]: Starting systemd-boot-update.service... Mar 17 19:06:05.569481 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Mar 17 19:06:05.573235 systemd[1]: Starting systemd-machine-id-commit.service... Mar 17 19:06:05.576892 systemd[1]: Starting systemd-sysext.service... Mar 17 19:06:05.833229 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1247 (bootctl) Mar 17 19:06:05.834922 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Mar 17 19:06:06.054235 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Mar 17 19:06:06.056000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.152742 systemd[1]: Unmounting usr-share-oem.mount... Mar 17 19:06:06.168252 systemd[1]: usr-share-oem.mount: Deactivated successfully. Mar 17 19:06:06.168508 systemd[1]: Unmounted usr-share-oem.mount. Mar 17 19:06:06.200405 kernel: loop0: detected capacity change from 0 to 205544 Mar 17 19:06:06.215169 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 17 19:06:06.216226 systemd[1]: Finished systemd-machine-id-commit.service. Mar 17 19:06:06.217000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.271563 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 17 19:06:06.287417 kernel: loop1: detected capacity change from 0 to 205544 Mar 17 19:06:06.293120 (sd-sysext)[1259]: Using extensions 'kubernetes'. Mar 17 19:06:06.293610 (sd-sysext)[1259]: Merged extensions into '/usr'. Mar 17 19:06:06.309571 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:06.311294 systemd[1]: Mounting usr-share-oem.mount... Mar 17 19:06:06.313562 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.317370 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 19:06:06.320777 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 19:06:06.324346 systemd[1]: Starting modprobe@loop.service... Mar 17 19:06:06.326351 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.326584 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 19:06:06.326758 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:06.329464 systemd[1]: Mounted usr-share-oem.mount. Mar 17 19:06:06.332050 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 19:06:06.332206 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 19:06:06.333000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.333000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.334875 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 19:06:06.335023 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 19:06:06.336000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.336000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.337950 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 19:06:06.338093 systemd[1]: Finished modprobe@loop.service. Mar 17 19:06:06.339000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.339000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.340898 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 19:06:06.341050 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.342211 systemd[1]: Finished systemd-sysext.service. Mar 17 19:06:06.344000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.345840 systemd[1]: Starting ensure-sysext.service... Mar 17 19:06:06.349118 systemd[1]: Starting systemd-tmpfiles-setup.service... Mar 17 19:06:06.360648 systemd[1]: Reloading. Mar 17 19:06:06.402063 systemd-tmpfiles[1266]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Mar 17 19:06:06.416935 /usr/lib/systemd/system-generators/torcx-generator[1294]: time="2025-03-17T19:06:06Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 19:06:06.417430 /usr/lib/systemd/system-generators/torcx-generator[1294]: time="2025-03-17T19:06:06Z" level=info msg="torcx already run" Mar 17 19:06:06.437967 systemd-tmpfiles[1266]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 17 19:06:06.471344 systemd-tmpfiles[1266]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 17 19:06:06.514657 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 19:06:06.514679 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 19:06:06.531101 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 19:06:06.596000 audit: BPF prog-id=30 op=LOAD Mar 17 19:06:06.596000 audit: BPF prog-id=26 op=UNLOAD Mar 17 19:06:06.599000 audit: BPF prog-id=31 op=LOAD Mar 17 19:06:06.599000 audit: BPF prog-id=32 op=LOAD Mar 17 19:06:06.599000 audit: BPF prog-id=24 op=UNLOAD Mar 17 19:06:06.599000 audit: BPF prog-id=25 op=UNLOAD Mar 17 19:06:06.599000 audit: BPF prog-id=33 op=LOAD Mar 17 19:06:06.599000 audit: BPF prog-id=27 op=UNLOAD Mar 17 19:06:06.600000 audit: BPF prog-id=34 op=LOAD Mar 17 19:06:06.600000 audit: BPF prog-id=35 op=LOAD Mar 17 19:06:06.600000 audit: BPF prog-id=28 op=UNLOAD Mar 17 19:06:06.600000 audit: BPF prog-id=29 op=UNLOAD Mar 17 19:06:06.601000 audit: BPF prog-id=36 op=LOAD Mar 17 19:06:06.601000 audit: BPF prog-id=21 op=UNLOAD Mar 17 19:06:06.601000 audit: BPF prog-id=37 op=LOAD Mar 17 19:06:06.601000 audit: BPF prog-id=38 op=LOAD Mar 17 19:06:06.602000 audit: BPF prog-id=22 op=UNLOAD Mar 17 19:06:06.602000 audit: BPF prog-id=23 op=UNLOAD Mar 17 19:06:06.616307 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:06.616632 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.618051 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 19:06:06.622269 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 19:06:06.624312 systemd[1]: Starting modprobe@loop.service... Mar 17 19:06:06.625310 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.625480 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 19:06:06.625637 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:06.626969 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 19:06:06.627142 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 19:06:06.625000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.625000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.625000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.625000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.628788 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 19:06:06.628912 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 19:06:06.633658 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 19:06:06.633802 systemd[1]: Finished modprobe@loop.service. Mar 17 19:06:06.632000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.632000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.636561 systemd[1]: Finished ensure-sysext.service. Mar 17 19:06:06.635000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.639055 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:06.639294 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.640752 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 19:06:06.644733 systemd[1]: Starting modprobe@drm.service... Mar 17 19:06:06.647013 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 19:06:06.648685 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.648776 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 19:06:06.648931 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 19:06:06.650532 systemd-networkd[1180]: eth0: Gained IPv6LL Mar 17 19:06:06.652497 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 19:06:06.652695 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 19:06:06.651000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.651000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.654352 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 19:06:06.654777 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 19:06:06.654899 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 19:06:06.654000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.654000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.656565 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 19:06:06.658012 systemd[1]: Finished systemd-networkd-wait-online.service. Mar 17 19:06:06.656000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.656000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.656000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:06.659175 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 19:06:06.659283 systemd[1]: Finished modprobe@drm.service. Mar 17 19:06:07.003474 systemd-fsck[1254]: fsck.fat 4.2 (2021-01-31) Mar 17 19:06:07.003474 systemd-fsck[1254]: /dev/sda1: 789 files, 119299/258078 clusters Mar 17 19:06:07.006140 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Mar 17 19:06:07.007000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:07.010793 systemd[1]: Mounting boot.mount... Mar 17 19:06:07.025075 systemd[1]: Mounted boot.mount. Mar 17 19:06:07.041366 systemd[1]: Finished systemd-boot-update.service. Mar 17 19:06:07.043000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.229645 systemd[1]: Finished systemd-tmpfiles-setup.service. Mar 17 19:06:08.231000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.234079 systemd[1]: Starting audit-rules.service... Mar 17 19:06:08.235427 kernel: kauditd_printk_skb: 124 callbacks suppressed Mar 17 19:06:08.235477 kernel: audit: type=1130 audit(1742238368.231:208): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.252339 systemd[1]: Starting clean-ca-certificates.service... Mar 17 19:06:08.255851 systemd[1]: Starting systemd-journal-catalog-update.service... Mar 17 19:06:08.258000 audit: BPF prog-id=39 op=LOAD Mar 17 19:06:08.260325 systemd[1]: Starting systemd-resolved.service... Mar 17 19:06:08.264923 kernel: audit: type=1334 audit(1742238368.258:209): prog-id=39 op=LOAD Mar 17 19:06:08.264000 audit: BPF prog-id=40 op=LOAD Mar 17 19:06:08.267128 systemd[1]: Starting systemd-timesyncd.service... Mar 17 19:06:08.270419 kernel: audit: type=1334 audit(1742238368.264:210): prog-id=40 op=LOAD Mar 17 19:06:08.273075 systemd[1]: Starting systemd-update-utmp.service... Mar 17 19:06:08.297000 audit[1370]: SYSTEM_BOOT pid=1370 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.310434 kernel: audit: type=1127 audit(1742238368.297:211): pid=1370 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.312794 systemd[1]: Finished systemd-update-utmp.service. Mar 17 19:06:08.314000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.328557 kernel: audit: type=1130 audit(1742238368.314:212): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.364751 systemd[1]: Finished clean-ca-certificates.service. Mar 17 19:06:08.365000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.367398 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 17 19:06:08.378404 kernel: audit: type=1130 audit(1742238368.365:213): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.478920 systemd[1]: Started systemd-timesyncd.service. Mar 17 19:06:08.481000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-timesyncd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.481743 systemd[1]: Reached target time-set.target. Mar 17 19:06:08.493788 kernel: audit: type=1130 audit(1742238368.481:214): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-timesyncd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.521049 systemd[1]: Finished systemd-journal-catalog-update.service. Mar 17 19:06:08.522000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.536504 kernel: audit: type=1130 audit(1742238368.522:215): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.610493 systemd-resolved[1363]: Positive Trust Anchors: Mar 17 19:06:08.610511 systemd-resolved[1363]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 19:06:08.610549 systemd-resolved[1363]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Mar 17 19:06:08.682175 systemd-resolved[1363]: Using system hostname 'ci-3510.3.7-a-a60daf7667'. Mar 17 19:06:08.683850 systemd[1]: Started systemd-resolved.service. Mar 17 19:06:08.685000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.686020 systemd[1]: Reached target network.target. Mar 17 19:06:08.701611 kernel: audit: type=1130 audit(1742238368.685:216): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 19:06:08.701673 systemd[1]: Reached target network-online.target. Mar 17 19:06:08.704029 augenrules[1380]: No rules Mar 17 19:06:08.701000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Mar 17 19:06:08.712937 systemd[1]: Reached target nss-lookup.target. Mar 17 19:06:08.701000 audit[1380]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fffc6f2e880 a2=420 a3=0 items=0 ppid=1359 pid=1380 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 19:06:08.701000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Mar 17 19:06:08.714599 kernel: audit: type=1305 audit(1742238368.701:217): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Mar 17 19:06:08.716626 systemd[1]: Finished audit-rules.service. Mar 17 19:06:08.747935 systemd-timesyncd[1364]: Contacted time server 193.1.12.167:123 (0.flatcar.pool.ntp.org). Mar 17 19:06:08.748018 systemd-timesyncd[1364]: Initial clock synchronization to Mon 2025-03-17 19:06:08.747365 UTC. Mar 17 19:06:13.962052 ldconfig[1246]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 17 19:06:13.972071 systemd[1]: Finished ldconfig.service. Mar 17 19:06:13.976004 systemd[1]: Starting systemd-update-done.service... Mar 17 19:06:13.984701 systemd[1]: Finished systemd-update-done.service. Mar 17 19:06:13.987316 systemd[1]: Reached target sysinit.target. Mar 17 19:06:13.989600 systemd[1]: Started motdgen.path. Mar 17 19:06:13.991468 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Mar 17 19:06:13.994501 systemd[1]: Started logrotate.timer. Mar 17 19:06:13.997593 systemd[1]: Started mdadm.timer. Mar 17 19:06:14.002538 systemd[1]: Started systemd-tmpfiles-clean.timer. Mar 17 19:06:14.004712 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 17 19:06:14.004757 systemd[1]: Reached target paths.target. Mar 17 19:06:14.007773 systemd[1]: Reached target timers.target. Mar 17 19:06:14.010761 systemd[1]: Listening on dbus.socket. Mar 17 19:06:14.013630 systemd[1]: Starting docker.socket... Mar 17 19:06:14.029803 systemd[1]: Listening on sshd.socket. Mar 17 19:06:14.031821 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 19:06:14.032314 systemd[1]: Listening on docker.socket. Mar 17 19:06:14.034137 systemd[1]: Reached target sockets.target. Mar 17 19:06:14.036004 systemd[1]: Reached target basic.target. Mar 17 19:06:14.037835 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Mar 17 19:06:14.037869 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Mar 17 19:06:14.038966 systemd[1]: Starting containerd.service... Mar 17 19:06:14.042153 systemd[1]: Starting dbus.service... Mar 17 19:06:14.045096 systemd[1]: Starting enable-oem-cloudinit.service... Mar 17 19:06:14.048355 systemd[1]: Starting extend-filesystems.service... Mar 17 19:06:14.050561 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Mar 17 19:06:14.051987 systemd[1]: Starting kubelet.service... Mar 17 19:06:14.057104 systemd[1]: Starting motdgen.service... Mar 17 19:06:14.060065 systemd[1]: Started nvidia.service. Mar 17 19:06:14.063043 systemd[1]: Starting prepare-helm.service... Mar 17 19:06:14.066132 systemd[1]: Starting ssh-key-proc-cmdline.service... Mar 17 19:06:14.069591 systemd[1]: Starting sshd-keygen.service... Mar 17 19:06:14.076471 systemd[1]: Starting systemd-logind.service... Mar 17 19:06:14.078495 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 19:06:14.078588 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 17 19:06:14.079119 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 17 19:06:14.080178 systemd[1]: Starting update-engine.service... Mar 17 19:06:14.085062 systemd[1]: Starting update-ssh-keys-after-ignition.service... Mar 17 19:06:14.094030 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 17 19:06:14.094286 systemd[1]: Finished ssh-key-proc-cmdline.service. Mar 17 19:06:14.129028 jq[1404]: true Mar 17 19:06:14.130414 jq[1390]: false Mar 17 19:06:14.131092 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 17 19:06:14.131346 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Mar 17 19:06:14.150348 jq[1413]: true Mar 17 19:06:14.167154 extend-filesystems[1391]: Found loop1 Mar 17 19:06:14.172521 extend-filesystems[1391]: Found sda Mar 17 19:06:14.176646 extend-filesystems[1391]: Found sda1 Mar 17 19:06:14.178282 extend-filesystems[1391]: Found sda2 Mar 17 19:06:14.178282 extend-filesystems[1391]: Found sda3 Mar 17 19:06:14.178282 extend-filesystems[1391]: Found usr Mar 17 19:06:14.178282 extend-filesystems[1391]: Found sda4 Mar 17 19:06:14.178282 extend-filesystems[1391]: Found sda6 Mar 17 19:06:14.178282 extend-filesystems[1391]: Found sda7 Mar 17 19:06:14.178282 extend-filesystems[1391]: Found sda9 Mar 17 19:06:14.178282 extend-filesystems[1391]: Checking size of /dev/sda9 Mar 17 19:06:14.189218 systemd[1]: motdgen.service: Deactivated successfully. Mar 17 19:06:14.189487 systemd[1]: Finished motdgen.service. Mar 17 19:06:14.232053 env[1416]: time="2025-03-17T19:06:14.231947884Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Mar 17 19:06:14.241742 systemd-logind[1401]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 17 19:06:14.247545 systemd-logind[1401]: New seat seat0. Mar 17 19:06:14.274062 env[1416]: time="2025-03-17T19:06:14.274016389Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 17 19:06:14.277489 env[1416]: time="2025-03-17T19:06:14.277460016Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 17 19:06:14.280410 env[1416]: time="2025-03-17T19:06:14.280345454Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.179-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 17 19:06:14.280410 env[1416]: time="2025-03-17T19:06:14.280395453Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 17 19:06:14.280719 env[1416]: time="2025-03-17T19:06:14.280685147Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 19:06:14.280778 env[1416]: time="2025-03-17T19:06:14.280719446Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 17 19:06:14.280778 env[1416]: time="2025-03-17T19:06:14.280739346Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Mar 17 19:06:14.280778 env[1416]: time="2025-03-17T19:06:14.280753345Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 17 19:06:14.280899 env[1416]: time="2025-03-17T19:06:14.280854943Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 17 19:06:14.281142 env[1416]: time="2025-03-17T19:06:14.281113738Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 17 19:06:14.281346 env[1416]: time="2025-03-17T19:06:14.281315834Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 19:06:14.281425 env[1416]: time="2025-03-17T19:06:14.281347033Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 17 19:06:14.281466 env[1416]: time="2025-03-17T19:06:14.281433931Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Mar 17 19:06:14.281466 env[1416]: time="2025-03-17T19:06:14.281450631Z" level=info msg="metadata content store policy set" policy=shared Mar 17 19:06:14.301254 tar[1408]: linux-amd64/helm Mar 17 19:06:14.341466 extend-filesystems[1391]: Old size kept for /dev/sda9 Mar 17 19:06:14.370022 extend-filesystems[1391]: Found sr0 Mar 17 19:06:14.344735 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 17 19:06:14.344922 systemd[1]: Finished extend-filesystems.service. Mar 17 19:06:14.378596 env[1416]: time="2025-03-17T19:06:14.378546265Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379610943Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379666142Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379718940Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379786339Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379804739Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379830338Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379847538Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379863537Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379880437Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379906336Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.379921836Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.380084533Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 17 19:06:14.380484 env[1416]: time="2025-03-17T19:06:14.380187031Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 17 19:06:14.380930 env[1416]: time="2025-03-17T19:06:14.380608722Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 17 19:06:14.380930 env[1416]: time="2025-03-17T19:06:14.380650221Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.380930 env[1416]: time="2025-03-17T19:06:14.380684720Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 17 19:06:14.380930 env[1416]: time="2025-03-17T19:06:14.380846116Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.380930 env[1416]: time="2025-03-17T19:06:14.380867716Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381109 env[1416]: time="2025-03-17T19:06:14.380885816Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381109 env[1416]: time="2025-03-17T19:06:14.380998113Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381109 env[1416]: time="2025-03-17T19:06:14.381018313Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381109 env[1416]: time="2025-03-17T19:06:14.381037312Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381109 env[1416]: time="2025-03-17T19:06:14.381053712Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381109 env[1416]: time="2025-03-17T19:06:14.381085511Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381109 env[1416]: time="2025-03-17T19:06:14.381105811Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 17 19:06:14.381355 env[1416]: time="2025-03-17T19:06:14.381280907Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381355 env[1416]: time="2025-03-17T19:06:14.381319606Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381355 env[1416]: time="2025-03-17T19:06:14.381339106Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.381482 env[1416]: time="2025-03-17T19:06:14.381356006Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 17 19:06:14.381482 env[1416]: time="2025-03-17T19:06:14.381393005Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Mar 17 19:06:14.381482 env[1416]: time="2025-03-17T19:06:14.381410704Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 17 19:06:14.381482 env[1416]: time="2025-03-17T19:06:14.381434204Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Mar 17 19:06:14.381621 env[1416]: time="2025-03-17T19:06:14.381489203Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 17 19:06:14.382406 env[1416]: time="2025-03-17T19:06:14.381787796Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 17 19:06:14.382406 env[1416]: time="2025-03-17T19:06:14.381870295Z" level=info msg="Connect containerd service" Mar 17 19:06:14.382406 env[1416]: time="2025-03-17T19:06:14.381911394Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.382780775Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.382909073Z" level=info msg="Start subscribing containerd event" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.382955972Z" level=info msg="Start recovering state" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.383020470Z" level=info msg="Start event monitor" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.383032970Z" level=info msg="Start snapshots syncer" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.383043170Z" level=info msg="Start cni network conf syncer for default" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.383052270Z" level=info msg="Start streaming server" Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.383787054Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.383840953Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 17 19:06:14.440450 env[1416]: time="2025-03-17T19:06:14.421926343Z" level=info msg="containerd successfully booted in 0.190933s" Mar 17 19:06:14.440794 bash[1436]: Updated "/home/core/.ssh/authorized_keys" Mar 17 19:06:14.396888 dbus-daemon[1389]: [system] SELinux support is enabled Mar 17 19:06:14.384022 systemd[1]: Started containerd.service. Mar 17 19:06:14.403006 dbus-daemon[1389]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 17 19:06:14.397044 systemd[1]: Started dbus.service. Mar 17 19:06:14.401665 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 17 19:06:14.401692 systemd[1]: Reached target system-config.target. Mar 17 19:06:14.404412 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 17 19:06:14.404435 systemd[1]: Reached target user-config.target. Mar 17 19:06:14.407346 systemd[1]: Started systemd-logind.service. Mar 17 19:06:14.415739 systemd[1]: Finished update-ssh-keys-after-ignition.service. Mar 17 19:06:14.502081 systemd[1]: nvidia.service: Deactivated successfully. Mar 17 19:06:15.098060 update_engine[1402]: I0317 19:06:15.097601 1402 main.cc:92] Flatcar Update Engine starting Mar 17 19:06:15.148008 systemd[1]: Started update-engine.service. Mar 17 19:06:15.150549 update_engine[1402]: I0317 19:06:15.149506 1402 update_check_scheduler.cc:74] Next update check in 10m41s Mar 17 19:06:15.153103 systemd[1]: Started locksmithd.service. Mar 17 19:06:15.221833 tar[1408]: linux-amd64/LICENSE Mar 17 19:06:15.221999 tar[1408]: linux-amd64/README.md Mar 17 19:06:15.231943 systemd[1]: Finished prepare-helm.service. Mar 17 19:06:15.590354 systemd[1]: Started kubelet.service. Mar 17 19:06:15.731989 sshd_keygen[1411]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 17 19:06:15.761088 systemd[1]: Finished sshd-keygen.service. Mar 17 19:06:15.765473 systemd[1]: Starting issuegen.service... Mar 17 19:06:15.770194 systemd[1]: Started waagent.service. Mar 17 19:06:15.777167 systemd[1]: issuegen.service: Deactivated successfully. Mar 17 19:06:15.777353 systemd[1]: Finished issuegen.service. Mar 17 19:06:15.781478 systemd[1]: Starting systemd-user-sessions.service... Mar 17 19:06:15.817701 systemd[1]: Finished systemd-user-sessions.service. Mar 17 19:06:15.822037 systemd[1]: Started getty@tty1.service. Mar 17 19:06:15.825876 systemd[1]: Started serial-getty@ttyS0.service. Mar 17 19:06:15.828834 systemd[1]: Reached target getty.target. Mar 17 19:06:15.830932 systemd[1]: Reached target multi-user.target. Mar 17 19:06:15.839493 systemd[1]: Starting systemd-update-utmp-runlevel.service... Mar 17 19:06:15.848604 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Mar 17 19:06:15.848779 systemd[1]: Finished systemd-update-utmp-runlevel.service. Mar 17 19:06:15.857010 systemd[1]: Startup finished in 808ms (firmware) + 22.487s (loader) + 924ms (kernel) + 14.509s (initrd) + 26.202s (userspace) = 1min 4.932s. Mar 17 19:06:16.187254 kubelet[1497]: E0317 19:06:16.187199 1497 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:06:16.188849 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:06:16.189017 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:06:16.382306 login[1520]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Mar 17 19:06:16.383889 login[1521]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Mar 17 19:06:16.447411 systemd[1]: Created slice user-500.slice. Mar 17 19:06:16.448901 systemd[1]: Starting user-runtime-dir@500.service... Mar 17 19:06:16.451404 systemd-logind[1401]: New session 1 of user core. Mar 17 19:06:16.457603 systemd-logind[1401]: New session 2 of user core. Mar 17 19:06:16.461905 systemd[1]: Finished user-runtime-dir@500.service. Mar 17 19:06:16.463650 systemd[1]: Starting user@500.service... Mar 17 19:06:16.476978 (systemd)[1524]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:06:16.585107 systemd[1524]: Queued start job for default target default.target. Mar 17 19:06:16.585740 systemd[1524]: Reached target paths.target. Mar 17 19:06:16.585768 systemd[1524]: Reached target sockets.target. Mar 17 19:06:16.585785 systemd[1524]: Reached target timers.target. Mar 17 19:06:16.585800 systemd[1524]: Reached target basic.target. Mar 17 19:06:16.585930 systemd[1]: Started user@500.service. Mar 17 19:06:16.587200 systemd[1]: Started session-1.scope. Mar 17 19:06:16.588009 systemd[1]: Started session-2.scope. Mar 17 19:06:16.589049 systemd[1524]: Reached target default.target. Mar 17 19:06:16.589256 systemd[1524]: Startup finished in 105ms. Mar 17 19:06:16.620003 locksmithd[1493]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 17 19:06:22.545001 waagent[1513]: 2025-03-17T19:06:22.544882Z INFO Daemon Daemon Azure Linux Agent Version:2.6.0.2 Mar 17 19:06:22.548680 waagent[1513]: 2025-03-17T19:06:22.548604Z INFO Daemon Daemon OS: flatcar 3510.3.7 Mar 17 19:06:22.550940 waagent[1513]: 2025-03-17T19:06:22.550877Z INFO Daemon Daemon Python: 3.9.16 Mar 17 19:06:22.553319 waagent[1513]: 2025-03-17T19:06:22.553247Z INFO Daemon Daemon Run daemon Mar 17 19:06:22.556049 waagent[1513]: 2025-03-17T19:06:22.555987Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='3510.3.7' Mar 17 19:06:22.568907 waagent[1513]: 2025-03-17T19:06:22.568788Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 1. Mar 17 19:06:22.576100 waagent[1513]: 2025-03-17T19:06:22.575991Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.577402Z INFO Daemon Daemon cloud-init is enabled: False Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.578171Z INFO Daemon Daemon Using waagent for provisioning Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.579611Z INFO Daemon Daemon Activate resource disk Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.580345Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.587963Z INFO Daemon Daemon Found device: None Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.588880Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.589726Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.591841Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.592927Z INFO Daemon Daemon Running default provisioning handler Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.602616Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 1. Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.605521Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.606465Z INFO Daemon Daemon cloud-init is enabled: False Mar 17 19:06:22.620688 waagent[1513]: 2025-03-17T19:06:22.607247Z INFO Daemon Daemon Copying ovf-env.xml Mar 17 19:06:22.629406 waagent[1513]: 2025-03-17T19:06:22.625986Z INFO Daemon Daemon Successfully mounted dvd Mar 17 19:06:22.715638 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Mar 17 19:06:22.734053 waagent[1513]: 2025-03-17T19:06:22.733897Z INFO Daemon Daemon Detect protocol endpoint Mar 17 19:06:22.735797 waagent[1513]: 2025-03-17T19:06:22.735723Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 17 19:06:22.736985 waagent[1513]: 2025-03-17T19:06:22.736816Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Mar 17 19:06:22.738517 waagent[1513]: 2025-03-17T19:06:22.738462Z INFO Daemon Daemon Test for route to 168.63.129.16 Mar 17 19:06:22.739571 waagent[1513]: 2025-03-17T19:06:22.739517Z INFO Daemon Daemon Route to 168.63.129.16 exists Mar 17 19:06:22.740687 waagent[1513]: 2025-03-17T19:06:22.740637Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Mar 17 19:06:22.832206 waagent[1513]: 2025-03-17T19:06:22.832063Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Mar 17 19:06:22.850157 waagent[1513]: 2025-03-17T19:06:22.833945Z INFO Daemon Daemon Wire protocol version:2012-11-30 Mar 17 19:06:22.850157 waagent[1513]: 2025-03-17T19:06:22.834978Z INFO Daemon Daemon Server preferred version:2015-04-05 Mar 17 19:06:23.101526 waagent[1513]: 2025-03-17T19:06:23.101297Z INFO Daemon Daemon Initializing goal state during protocol detection Mar 17 19:06:23.111241 waagent[1513]: 2025-03-17T19:06:23.111156Z INFO Daemon Daemon Forcing an update of the goal state.. Mar 17 19:06:23.116380 waagent[1513]: 2025-03-17T19:06:23.112507Z INFO Daemon Daemon Fetching goal state [incarnation 1] Mar 17 19:06:23.189648 waagent[1513]: 2025-03-17T19:06:23.189513Z INFO Daemon Daemon Found private key matching thumbprint 5F315EC6781F3F939053C3B61909605342E677AA Mar 17 19:06:23.194294 waagent[1513]: 2025-03-17T19:06:23.194207Z INFO Daemon Daemon Certificate with thumbprint 234E1AFB2B595C3A0DEA6FE8A9A8243568790FE9 has no matching private key. Mar 17 19:06:23.198947 waagent[1513]: 2025-03-17T19:06:23.198870Z INFO Daemon Daemon Fetch goal state completed Mar 17 19:06:23.240283 waagent[1513]: 2025-03-17T19:06:23.240189Z INFO Daemon Daemon Fetched new vmSettings [correlation ID: b90152e2-6b87-4c7f-9cbe-653f2ce6a0fc New eTag: 7455229735815960187] Mar 17 19:06:23.246217 waagent[1513]: 2025-03-17T19:06:23.246118Z INFO Daemon Daemon Status Blob type 'None' is not valid, assuming BlockBlob Mar 17 19:06:23.294722 waagent[1513]: 2025-03-17T19:06:23.294631Z INFO Daemon Daemon Starting provisioning Mar 17 19:06:23.297542 waagent[1513]: 2025-03-17T19:06:23.297449Z INFO Daemon Daemon Handle ovf-env.xml. Mar 17 19:06:23.299882 waagent[1513]: 2025-03-17T19:06:23.299811Z INFO Daemon Daemon Set hostname [ci-3510.3.7-a-a60daf7667] Mar 17 19:06:23.333713 waagent[1513]: 2025-03-17T19:06:23.333555Z INFO Daemon Daemon Publish hostname [ci-3510.3.7-a-a60daf7667] Mar 17 19:06:23.337439 waagent[1513]: 2025-03-17T19:06:23.337314Z INFO Daemon Daemon Examine /proc/net/route for primary interface Mar 17 19:06:23.340955 waagent[1513]: 2025-03-17T19:06:23.340882Z INFO Daemon Daemon Primary interface is [eth0] Mar 17 19:06:23.355456 systemd[1]: systemd-networkd-wait-online.service: Deactivated successfully. Mar 17 19:06:23.355722 systemd[1]: Stopped systemd-networkd-wait-online.service. Mar 17 19:06:23.355795 systemd[1]: Stopping systemd-networkd-wait-online.service... Mar 17 19:06:23.356118 systemd[1]: Stopping systemd-networkd.service... Mar 17 19:06:23.359423 systemd-networkd[1180]: eth0: DHCPv6 lease lost Mar 17 19:06:23.360945 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 17 19:06:23.361145 systemd[1]: Stopped systemd-networkd.service. Mar 17 19:06:23.363675 systemd[1]: Starting systemd-networkd.service... Mar 17 19:06:23.395613 systemd-networkd[1571]: enP55511s1: Link UP Mar 17 19:06:23.395623 systemd-networkd[1571]: enP55511s1: Gained carrier Mar 17 19:06:23.396988 systemd-networkd[1571]: eth0: Link UP Mar 17 19:06:23.396998 systemd-networkd[1571]: eth0: Gained carrier Mar 17 19:06:23.397451 systemd-networkd[1571]: lo: Link UP Mar 17 19:06:23.397459 systemd-networkd[1571]: lo: Gained carrier Mar 17 19:06:23.397779 systemd-networkd[1571]: eth0: Gained IPv6LL Mar 17 19:06:23.398060 systemd-networkd[1571]: Enumeration completed Mar 17 19:06:23.398182 systemd[1]: Started systemd-networkd.service. Mar 17 19:06:23.400648 systemd[1]: Starting systemd-networkd-wait-online.service... Mar 17 19:06:23.403580 systemd-networkd[1571]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 19:06:23.408075 waagent[1513]: 2025-03-17T19:06:23.405148Z INFO Daemon Daemon Create user account if not exists Mar 17 19:06:23.408568 waagent[1513]: 2025-03-17T19:06:23.408488Z INFO Daemon Daemon User core already exists, skip useradd Mar 17 19:06:23.412125 waagent[1513]: 2025-03-17T19:06:23.412045Z INFO Daemon Daemon Configure sudoer Mar 17 19:06:23.418210 waagent[1513]: 2025-03-17T19:06:23.414007Z INFO Daemon Daemon Configure sshd Mar 17 19:06:23.418210 waagent[1513]: 2025-03-17T19:06:23.415152Z INFO Daemon Daemon Deploy ssh public key. Mar 17 19:06:23.441462 systemd-networkd[1571]: eth0: DHCPv4 address 10.200.8.15/24, gateway 10.200.8.1 acquired from 168.63.129.16 Mar 17 19:06:23.444759 systemd[1]: Finished systemd-networkd-wait-online.service. Mar 17 19:06:23.508759 waagent[1513]: 2025-03-17T19:06:23.508626Z INFO Daemon Daemon Decode custom data Mar 17 19:06:23.511406 waagent[1513]: 2025-03-17T19:06:23.511315Z INFO Daemon Daemon Save custom data Mar 17 19:06:24.574223 waagent[1513]: 2025-03-17T19:06:24.574102Z INFO Daemon Daemon Provisioning complete Mar 17 19:06:24.588248 waagent[1513]: 2025-03-17T19:06:24.588174Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Mar 17 19:06:24.591854 waagent[1513]: 2025-03-17T19:06:24.591778Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Mar 17 19:06:24.597398 waagent[1513]: 2025-03-17T19:06:24.597315Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.6.0.2 is the most current agent Mar 17 19:06:24.867684 waagent[1580]: 2025-03-17T19:06:24.865780Z INFO ExtHandler ExtHandler Agent WALinuxAgent-2.6.0.2 is running as the goal state agent Mar 17 19:06:24.868052 waagent[1580]: 2025-03-17T19:06:24.867839Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 17 19:06:24.868123 waagent[1580]: 2025-03-17T19:06:24.868021Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 17 19:06:24.879287 waagent[1580]: 2025-03-17T19:06:24.879213Z INFO ExtHandler ExtHandler Forcing an update of the goal state.. Mar 17 19:06:24.879470 waagent[1580]: 2025-03-17T19:06:24.879414Z INFO ExtHandler ExtHandler Fetching goal state [incarnation 1] Mar 17 19:06:24.941826 waagent[1580]: 2025-03-17T19:06:24.941693Z INFO ExtHandler ExtHandler Found private key matching thumbprint 5F315EC6781F3F939053C3B61909605342E677AA Mar 17 19:06:24.942073 waagent[1580]: 2025-03-17T19:06:24.942007Z INFO ExtHandler ExtHandler Certificate with thumbprint 234E1AFB2B595C3A0DEA6FE8A9A8243568790FE9 has no matching private key. Mar 17 19:06:24.942309 waagent[1580]: 2025-03-17T19:06:24.942258Z INFO ExtHandler ExtHandler Fetch goal state completed Mar 17 19:06:24.957072 waagent[1580]: 2025-03-17T19:06:24.957001Z INFO ExtHandler ExtHandler Fetched new vmSettings [correlation ID: 1a00d225-377b-4544-b707-b251d9997fe0 New eTag: 7455229735815960187] Mar 17 19:06:24.957640 waagent[1580]: 2025-03-17T19:06:24.957580Z INFO ExtHandler ExtHandler Status Blob type 'None' is not valid, assuming BlockBlob Mar 17 19:06:25.014544 waagent[1580]: 2025-03-17T19:06:25.014350Z INFO ExtHandler ExtHandler Distro: flatcar-3510.3.7; OSUtil: CoreOSUtil; AgentService: waagent; Python: 3.9.16; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Mar 17 19:06:25.034644 waagent[1580]: 2025-03-17T19:06:25.034543Z INFO ExtHandler ExtHandler WALinuxAgent-2.6.0.2 running as process 1580 Mar 17 19:06:25.038117 waagent[1580]: 2025-03-17T19:06:25.038044Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '3510.3.7', '', 'Flatcar Container Linux by Kinvolk'] Mar 17 19:06:25.039394 waagent[1580]: 2025-03-17T19:06:25.039324Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Mar 17 19:06:25.165182 waagent[1580]: 2025-03-17T19:06:25.165114Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Mar 17 19:06:25.165653 waagent[1580]: 2025-03-17T19:06:25.165584Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Mar 17 19:06:25.173696 waagent[1580]: 2025-03-17T19:06:25.173635Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Mar 17 19:06:25.174200 waagent[1580]: 2025-03-17T19:06:25.174138Z ERROR ExtHandler ExtHandler Unable to setup the persistent firewall rules: [Errno 30] Read-only file system: '/lib/systemd/system/waagent-network-setup.service' Mar 17 19:06:25.175291 waagent[1580]: 2025-03-17T19:06:25.175223Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [False], cgroups enabled [False], python supported: [True] Mar 17 19:06:25.176585 waagent[1580]: 2025-03-17T19:06:25.176525Z INFO ExtHandler ExtHandler Starting env monitor service. Mar 17 19:06:25.176858 waagent[1580]: 2025-03-17T19:06:25.176801Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 17 19:06:25.177624 waagent[1580]: 2025-03-17T19:06:25.177567Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Mar 17 19:06:25.177751 waagent[1580]: 2025-03-17T19:06:25.177679Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 17 19:06:25.178662 waagent[1580]: 2025-03-17T19:06:25.178608Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Mar 17 19:06:25.178969 waagent[1580]: 2025-03-17T19:06:25.178910Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Mar 17 19:06:25.179329 waagent[1580]: 2025-03-17T19:06:25.179279Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 17 19:06:25.179479 waagent[1580]: 2025-03-17T19:06:25.179352Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Mar 17 19:06:25.179939 waagent[1580]: 2025-03-17T19:06:25.179885Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 17 19:06:25.180191 waagent[1580]: 2025-03-17T19:06:25.180130Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Mar 17 19:06:25.180191 waagent[1580]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Mar 17 19:06:25.180191 waagent[1580]: eth0 00000000 0108C80A 0003 0 0 1024 00000000 0 0 0 Mar 17 19:06:25.180191 waagent[1580]: eth0 0008C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Mar 17 19:06:25.180191 waagent[1580]: eth0 0108C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Mar 17 19:06:25.180191 waagent[1580]: eth0 10813FA8 0108C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 17 19:06:25.180191 waagent[1580]: eth0 FEA9FEA9 0108C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 17 19:06:25.184424 waagent[1580]: 2025-03-17T19:06:25.184055Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Mar 17 19:06:25.185603 waagent[1580]: 2025-03-17T19:06:25.185541Z INFO EnvHandler ExtHandler Configure routes Mar 17 19:06:25.185896 waagent[1580]: 2025-03-17T19:06:25.185846Z INFO EnvHandler ExtHandler Gateway:None Mar 17 19:06:25.186023 waagent[1580]: 2025-03-17T19:06:25.185971Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Mar 17 19:06:25.186341 waagent[1580]: 2025-03-17T19:06:25.186228Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Mar 17 19:06:25.186602 waagent[1580]: 2025-03-17T19:06:25.186555Z INFO EnvHandler ExtHandler Routes:None Mar 17 19:06:25.201082 waagent[1580]: 2025-03-17T19:06:25.201028Z INFO ExtHandler ExtHandler Checking for agent updates (family: Prod) Mar 17 19:06:25.201791 waagent[1580]: 2025-03-17T19:06:25.201751Z WARNING ExtHandler ExtHandler Fetch failed: [HttpError] HTTPS is unavailable and required Mar 17 19:06:25.202671 waagent[1580]: 2025-03-17T19:06:25.202626Z INFO ExtHandler ExtHandler [PERIODIC] Request failed using the direct channel. Error: 'NoneType' object has no attribute 'getheaders' Mar 17 19:06:25.219577 waagent[1580]: 2025-03-17T19:06:25.219497Z ERROR EnvHandler ExtHandler Failed to get the PID of the DHCP client: invalid literal for int() with base 10: 'MainPID=1571' Mar 17 19:06:25.238264 waagent[1580]: 2025-03-17T19:06:25.238193Z INFO ExtHandler ExtHandler Default channel changed to HostGA channel. Mar 17 19:06:25.319427 waagent[1580]: 2025-03-17T19:06:25.319253Z INFO MonitorHandler ExtHandler Network interfaces: Mar 17 19:06:25.319427 waagent[1580]: Executing ['ip', '-a', '-o', 'link']: Mar 17 19:06:25.319427 waagent[1580]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Mar 17 19:06:25.319427 waagent[1580]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 7c:1e:52:fc:a1:da brd ff:ff:ff:ff:ff:ff Mar 17 19:06:25.319427 waagent[1580]: 3: enP55511s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 7c:1e:52:fc:a1:da brd ff:ff:ff:ff:ff:ff\ altname enP55511p0s2 Mar 17 19:06:25.319427 waagent[1580]: Executing ['ip', '-4', '-a', '-o', 'address']: Mar 17 19:06:25.319427 waagent[1580]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Mar 17 19:06:25.319427 waagent[1580]: 2: eth0 inet 10.200.8.15/24 metric 1024 brd 10.200.8.255 scope global eth0\ valid_lft forever preferred_lft forever Mar 17 19:06:25.319427 waagent[1580]: Executing ['ip', '-6', '-a', '-o', 'address']: Mar 17 19:06:25.319427 waagent[1580]: 1: lo inet6 ::1/128 scope host \ valid_lft forever preferred_lft forever Mar 17 19:06:25.319427 waagent[1580]: 2: eth0 inet6 fe80::7e1e:52ff:fefc:a1da/64 scope link \ valid_lft forever preferred_lft forever Mar 17 19:06:25.362754 waagent[1580]: 2025-03-17T19:06:25.362664Z INFO ExtHandler ExtHandler Agent WALinuxAgent-2.6.0.2 discovered update WALinuxAgent-2.12.0.2 -- exiting Mar 17 19:06:25.600921 waagent[1513]: 2025-03-17T19:06:25.600718Z INFO Daemon Daemon Agent WALinuxAgent-2.6.0.2 launched with command '/usr/share/oem/python/bin/python -u /usr/share/oem/bin/waagent -run-exthandlers' is successfully running Mar 17 19:06:25.604690 waagent[1513]: 2025-03-17T19:06:25.604629Z INFO Daemon Daemon Determined Agent WALinuxAgent-2.12.0.2 to be the latest agent Mar 17 19:06:26.379142 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 17 19:06:26.379422 systemd[1]: Stopped kubelet.service. Mar 17 19:06:26.381301 systemd[1]: Starting kubelet.service... Mar 17 19:06:26.504739 systemd[1]: Started kubelet.service. Mar 17 19:06:26.570521 kubelet[1618]: E0317 19:06:26.570476 1618 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:06:26.574654 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:06:26.574819 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:06:26.763288 waagent[1611]: 2025-03-17T19:06:26.763115Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.12.0.2) Mar 17 19:06:26.763958 waagent[1611]: 2025-03-17T19:06:26.763891Z INFO ExtHandler ExtHandler OS: flatcar 3510.3.7 Mar 17 19:06:26.764109 waagent[1611]: 2025-03-17T19:06:26.764054Z INFO ExtHandler ExtHandler Python: 3.9.16 Mar 17 19:06:26.764259 waagent[1611]: 2025-03-17T19:06:26.764211Z INFO ExtHandler ExtHandler CPU Arch: x86_64 Mar 17 19:06:26.773955 waagent[1611]: 2025-03-17T19:06:26.773843Z INFO ExtHandler ExtHandler Distro: flatcar-3510.3.7; OSUtil: CoreOSUtil; AgentService: waagent; Python: 3.9.16; Arch: x86_64; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Mar 17 19:06:26.774351 waagent[1611]: 2025-03-17T19:06:26.774291Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 17 19:06:26.774537 waagent[1611]: 2025-03-17T19:06:26.774488Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 17 19:06:26.786102 waagent[1611]: 2025-03-17T19:06:26.786032Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Mar 17 19:06:26.794233 waagent[1611]: 2025-03-17T19:06:26.794169Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.164 Mar 17 19:06:26.795143 waagent[1611]: 2025-03-17T19:06:26.795081Z INFO ExtHandler Mar 17 19:06:26.795289 waagent[1611]: 2025-03-17T19:06:26.795238Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 85597f64-7a04-468a-bb78-1c7502cca761 eTag: 7455229735815960187 source: Fabric] Mar 17 19:06:26.796001 waagent[1611]: 2025-03-17T19:06:26.795944Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Mar 17 19:06:26.797070 waagent[1611]: 2025-03-17T19:06:26.797010Z INFO ExtHandler Mar 17 19:06:26.797209 waagent[1611]: 2025-03-17T19:06:26.797156Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Mar 17 19:06:26.804278 waagent[1611]: 2025-03-17T19:06:26.804224Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Mar 17 19:06:26.804740 waagent[1611]: 2025-03-17T19:06:26.804694Z WARNING ExtHandler ExtHandler Fetch failed: [HttpError] HTTPS is unavailable and required Mar 17 19:06:26.825096 waagent[1611]: 2025-03-17T19:06:26.825027Z INFO ExtHandler ExtHandler Default channel changed to HostGAPlugin channel. Mar 17 19:06:26.891937 waagent[1611]: 2025-03-17T19:06:26.891796Z INFO ExtHandler Downloaded certificate {'thumbprint': '5F315EC6781F3F939053C3B61909605342E677AA', 'hasPrivateKey': True} Mar 17 19:06:26.892923 waagent[1611]: 2025-03-17T19:06:26.892855Z INFO ExtHandler Downloaded certificate {'thumbprint': '234E1AFB2B595C3A0DEA6FE8A9A8243568790FE9', 'hasPrivateKey': False} Mar 17 19:06:26.893898 waagent[1611]: 2025-03-17T19:06:26.893837Z INFO ExtHandler Fetch goal state completed Mar 17 19:06:26.915110 waagent[1611]: 2025-03-17T19:06:26.915014Z INFO ExtHandler ExtHandler OpenSSL version: OpenSSL 3.0.15 3 Sep 2024 (Library: OpenSSL 3.0.15 3 Sep 2024) Mar 17 19:06:26.926337 waagent[1611]: 2025-03-17T19:06:26.926253Z INFO ExtHandler ExtHandler WALinuxAgent-2.12.0.2 running as process 1611 Mar 17 19:06:26.929288 waagent[1611]: 2025-03-17T19:06:26.929225Z INFO ExtHandler ExtHandler [CGI] Cgroup monitoring is not supported on ['flatcar', '3510.3.7', '', 'Flatcar Container Linux by Kinvolk'] Mar 17 19:06:26.930227 waagent[1611]: 2025-03-17T19:06:26.930168Z INFO ExtHandler ExtHandler [CGI] Agent will reset the quotas in case distro: ['flatcar', '3510.3.7', '', 'Flatcar Container Linux by Kinvolk'] went from supported to unsupported Mar 17 19:06:26.930528 waagent[1611]: 2025-03-17T19:06:26.930472Z INFO ExtHandler ExtHandler [CGI] Agent cgroups enabled: False Mar 17 19:06:26.932424 waagent[1611]: 2025-03-17T19:06:26.932351Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Mar 17 19:06:26.937135 waagent[1611]: 2025-03-17T19:06:26.937081Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Mar 17 19:06:26.937508 waagent[1611]: 2025-03-17T19:06:26.937453Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Mar 17 19:06:26.945487 waagent[1611]: 2025-03-17T19:06:26.945432Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Mar 17 19:06:26.945936 waagent[1611]: 2025-03-17T19:06:26.945879Z ERROR ExtHandler ExtHandler Unable to setup the persistent firewall rules: [Errno 30] Read-only file system: '/lib/systemd/system/waagent-network-setup.service' Mar 17 19:06:26.951745 waagent[1611]: 2025-03-17T19:06:26.951653Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Mar 17 19:06:26.952700 waagent[1611]: 2025-03-17T19:06:26.952634Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: 1. configuration enabled [True], 2. cgroups v1 enabled [False] OR cgroups v2 is in use and v2 resource limiting configuration enabled [False], 3. python supported: [True] Mar 17 19:06:26.954117 waagent[1611]: 2025-03-17T19:06:26.954056Z INFO ExtHandler ExtHandler Starting env monitor service. Mar 17 19:06:26.954992 waagent[1611]: 2025-03-17T19:06:26.954936Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Mar 17 19:06:26.955331 waagent[1611]: 2025-03-17T19:06:26.955279Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 17 19:06:26.955446 waagent[1611]: 2025-03-17T19:06:26.955366Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 17 19:06:26.955773 waagent[1611]: 2025-03-17T19:06:26.955717Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Mar 17 19:06:26.955963 waagent[1611]: 2025-03-17T19:06:26.955914Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 17 19:06:26.956068 waagent[1611]: 2025-03-17T19:06:26.956001Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 17 19:06:26.956722 waagent[1611]: 2025-03-17T19:06:26.956667Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Mar 17 19:06:26.957048 waagent[1611]: 2025-03-17T19:06:26.956995Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Mar 17 19:06:26.957541 waagent[1611]: 2025-03-17T19:06:26.957482Z INFO EnvHandler ExtHandler Configure routes Mar 17 19:06:26.957810 waagent[1611]: 2025-03-17T19:06:26.957754Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Mar 17 19:06:26.957810 waagent[1611]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Mar 17 19:06:26.957810 waagent[1611]: eth0 00000000 0108C80A 0003 0 0 1024 00000000 0 0 0 Mar 17 19:06:26.957810 waagent[1611]: eth0 0008C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Mar 17 19:06:26.957810 waagent[1611]: eth0 0108C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Mar 17 19:06:26.957810 waagent[1611]: eth0 10813FA8 0108C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 17 19:06:26.957810 waagent[1611]: eth0 FEA9FEA9 0108C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 17 19:06:26.958302 waagent[1611]: 2025-03-17T19:06:26.958245Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Mar 17 19:06:26.958756 waagent[1611]: 2025-03-17T19:06:26.958683Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Mar 17 19:06:26.958881 waagent[1611]: 2025-03-17T19:06:26.958829Z INFO EnvHandler ExtHandler Gateway:None Mar 17 19:06:26.959456 waagent[1611]: 2025-03-17T19:06:26.959404Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Mar 17 19:06:26.959841 waagent[1611]: 2025-03-17T19:06:26.959788Z INFO EnvHandler ExtHandler Routes:None Mar 17 19:06:26.975445 waagent[1611]: 2025-03-17T19:06:26.975368Z INFO ExtHandler ExtHandler Downloading agent manifest Mar 17 19:06:26.992262 waagent[1611]: 2025-03-17T19:06:26.992174Z INFO MonitorHandler ExtHandler Network interfaces: Mar 17 19:06:26.992262 waagent[1611]: Executing ['ip', '-a', '-o', 'link']: Mar 17 19:06:26.992262 waagent[1611]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Mar 17 19:06:26.992262 waagent[1611]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 7c:1e:52:fc:a1:da brd ff:ff:ff:ff:ff:ff Mar 17 19:06:26.992262 waagent[1611]: 3: enP55511s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 7c:1e:52:fc:a1:da brd ff:ff:ff:ff:ff:ff\ altname enP55511p0s2 Mar 17 19:06:26.992262 waagent[1611]: Executing ['ip', '-4', '-a', '-o', 'address']: Mar 17 19:06:26.992262 waagent[1611]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Mar 17 19:06:26.992262 waagent[1611]: 2: eth0 inet 10.200.8.15/24 metric 1024 brd 10.200.8.255 scope global eth0\ valid_lft forever preferred_lft forever Mar 17 19:06:26.992262 waagent[1611]: Executing ['ip', '-6', '-a', '-o', 'address']: Mar 17 19:06:26.992262 waagent[1611]: 1: lo inet6 ::1/128 scope host \ valid_lft forever preferred_lft forever Mar 17 19:06:26.992262 waagent[1611]: 2: eth0 inet6 fe80::7e1e:52ff:fefc:a1da/64 scope link \ valid_lft forever preferred_lft forever Mar 17 19:06:27.021187 waagent[1611]: 2025-03-17T19:06:27.021043Z INFO ExtHandler ExtHandler Mar 17 19:06:27.021350 waagent[1611]: 2025-03-17T19:06:27.021282Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: a6c18e02-bb90-479c-839f-5d07d2909587 correlation d1eb894e-54ce-442a-8a3f-0668f8f16d35 created: 2025-03-17T19:04:01.115570Z] Mar 17 19:06:27.022449 waagent[1611]: 2025-03-17T19:06:27.022371Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Mar 17 19:06:27.024338 waagent[1611]: 2025-03-17T19:06:27.024278Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 3 ms] Mar 17 19:06:27.044191 waagent[1611]: 2025-03-17T19:06:27.044124Z INFO ExtHandler ExtHandler Looking for existing remote access users. Mar 17 19:06:27.101021 waagent[1611]: 2025-03-17T19:06:27.100944Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.12.0.2 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 6BCA58B5-E1E7-446D-9ED5-4AF4C04EEEDD;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 1;UpdateMode: SelfUpdate;] Mar 17 19:06:27.203715 waagent[1611]: 2025-03-17T19:06:27.203591Z INFO EnvHandler ExtHandler Created firewall rules for the Azure Fabric: Mar 17 19:06:27.203715 waagent[1611]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 17 19:06:27.203715 waagent[1611]: pkts bytes target prot opt in out source destination Mar 17 19:06:27.203715 waagent[1611]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 17 19:06:27.203715 waagent[1611]: pkts bytes target prot opt in out source destination Mar 17 19:06:27.203715 waagent[1611]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 17 19:06:27.203715 waagent[1611]: pkts bytes target prot opt in out source destination Mar 17 19:06:27.203715 waagent[1611]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 17 19:06:27.203715 waagent[1611]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 17 19:06:27.203715 waagent[1611]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 17 19:06:27.210808 waagent[1611]: 2025-03-17T19:06:27.210702Z INFO EnvHandler ExtHandler Current Firewall rules: Mar 17 19:06:27.210808 waagent[1611]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 17 19:06:27.210808 waagent[1611]: pkts bytes target prot opt in out source destination Mar 17 19:06:27.210808 waagent[1611]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 17 19:06:27.210808 waagent[1611]: pkts bytes target prot opt in out source destination Mar 17 19:06:27.210808 waagent[1611]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 17 19:06:27.210808 waagent[1611]: pkts bytes target prot opt in out source destination Mar 17 19:06:27.210808 waagent[1611]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 17 19:06:27.210808 waagent[1611]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 17 19:06:27.210808 waagent[1611]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 17 19:06:27.211349 waagent[1611]: 2025-03-17T19:06:27.211296Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Mar 17 19:06:36.629069 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 17 19:06:36.629419 systemd[1]: Stopped kubelet.service. Mar 17 19:06:36.631430 systemd[1]: Starting kubelet.service... Mar 17 19:06:36.711951 systemd[1]: Started kubelet.service. Mar 17 19:06:37.323282 kubelet[1676]: E0317 19:06:37.323227 1676 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:06:37.324982 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:06:37.325140 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:06:47.379113 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 17 19:06:47.379488 systemd[1]: Stopped kubelet.service. Mar 17 19:06:47.381490 systemd[1]: Starting kubelet.service... Mar 17 19:06:47.461371 systemd[1]: Started kubelet.service. Mar 17 19:06:48.126091 kubelet[1686]: E0317 19:06:48.126036 1686 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:06:48.127817 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:06:48.127980 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:06:52.971442 kernel: hv_balloon: Max. dynamic memory size: 8192 MB Mar 17 19:06:58.129116 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 17 19:06:58.129475 systemd[1]: Stopped kubelet.service. Mar 17 19:06:58.131492 systemd[1]: Starting kubelet.service... Mar 17 19:06:58.212727 systemd[1]: Started kubelet.service. Mar 17 19:06:58.790894 kubelet[1695]: E0317 19:06:58.790841 1695 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:06:58.792461 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:06:58.792621 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:07:00.257534 update_engine[1402]: I0317 19:07:00.257444 1402 update_attempter.cc:509] Updating boot flags... Mar 17 19:07:08.879333 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Mar 17 19:07:08.879665 systemd[1]: Stopped kubelet.service. Mar 17 19:07:08.882544 systemd[1]: Starting kubelet.service... Mar 17 19:07:09.134494 systemd[1]: Started kubelet.service. Mar 17 19:07:09.508224 kubelet[1747]: E0317 19:07:09.508097 1747 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:07:09.509627 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:07:09.509786 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:07:09.757253 systemd[1]: Created slice system-sshd.slice. Mar 17 19:07:09.759408 systemd[1]: Started sshd@0-10.200.8.15:22-10.200.16.10:43928.service. Mar 17 19:07:10.553658 sshd[1753]: Accepted publickey for core from 10.200.16.10 port 43928 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:07:10.555317 sshd[1753]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:07:10.559687 systemd-logind[1401]: New session 3 of user core. Mar 17 19:07:10.560982 systemd[1]: Started session-3.scope. Mar 17 19:07:11.097266 systemd[1]: Started sshd@1-10.200.8.15:22-10.200.16.10:43932.service. Mar 17 19:07:11.720470 sshd[1758]: Accepted publickey for core from 10.200.16.10 port 43932 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:07:11.722142 sshd[1758]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:07:11.727198 systemd-logind[1401]: New session 4 of user core. Mar 17 19:07:11.727705 systemd[1]: Started session-4.scope. Mar 17 19:07:12.180150 sshd[1758]: pam_unix(sshd:session): session closed for user core Mar 17 19:07:12.183344 systemd[1]: sshd@1-10.200.8.15:22-10.200.16.10:43932.service: Deactivated successfully. Mar 17 19:07:12.184311 systemd[1]: session-4.scope: Deactivated successfully. Mar 17 19:07:12.184914 systemd-logind[1401]: Session 4 logged out. Waiting for processes to exit. Mar 17 19:07:12.185713 systemd-logind[1401]: Removed session 4. Mar 17 19:07:12.285660 systemd[1]: Started sshd@2-10.200.8.15:22-10.200.16.10:43936.service. Mar 17 19:07:12.909881 sshd[1764]: Accepted publickey for core from 10.200.16.10 port 43936 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:07:12.911580 sshd[1764]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:07:12.917346 systemd[1]: Started session-5.scope. Mar 17 19:07:12.918302 systemd-logind[1401]: New session 5 of user core. Mar 17 19:07:13.351796 sshd[1764]: pam_unix(sshd:session): session closed for user core Mar 17 19:07:13.355513 systemd[1]: sshd@2-10.200.8.15:22-10.200.16.10:43936.service: Deactivated successfully. Mar 17 19:07:13.356474 systemd[1]: session-5.scope: Deactivated successfully. Mar 17 19:07:13.357258 systemd-logind[1401]: Session 5 logged out. Waiting for processes to exit. Mar 17 19:07:13.358135 systemd-logind[1401]: Removed session 5. Mar 17 19:07:13.456753 systemd[1]: Started sshd@3-10.200.8.15:22-10.200.16.10:43950.service. Mar 17 19:07:14.081249 sshd[1770]: Accepted publickey for core from 10.200.16.10 port 43950 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:07:14.083570 sshd[1770]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:07:14.088259 systemd[1]: Started session-6.scope. Mar 17 19:07:14.088987 systemd-logind[1401]: New session 6 of user core. Mar 17 19:07:14.528117 sshd[1770]: pam_unix(sshd:session): session closed for user core Mar 17 19:07:14.531436 systemd[1]: sshd@3-10.200.8.15:22-10.200.16.10:43950.service: Deactivated successfully. Mar 17 19:07:14.532350 systemd[1]: session-6.scope: Deactivated successfully. Mar 17 19:07:14.532991 systemd-logind[1401]: Session 6 logged out. Waiting for processes to exit. Mar 17 19:07:14.533728 systemd-logind[1401]: Removed session 6. Mar 17 19:07:14.633246 systemd[1]: Started sshd@4-10.200.8.15:22-10.200.16.10:43956.service. Mar 17 19:07:15.258257 sshd[1776]: Accepted publickey for core from 10.200.16.10 port 43956 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:07:15.259912 sshd[1776]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:07:15.265587 systemd[1]: Started session-7.scope. Mar 17 19:07:15.266030 systemd-logind[1401]: New session 7 of user core. Mar 17 19:07:15.776591 sudo[1779]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 17 19:07:15.776889 sudo[1779]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Mar 17 19:07:15.799891 systemd[1]: Starting docker.service... Mar 17 19:07:15.837609 env[1789]: time="2025-03-17T19:07:15.837554732Z" level=info msg="Starting up" Mar 17 19:07:15.838753 env[1789]: time="2025-03-17T19:07:15.838728365Z" level=info msg="parsed scheme: \"unix\"" module=grpc Mar 17 19:07:15.838887 env[1789]: time="2025-03-17T19:07:15.838874069Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Mar 17 19:07:15.838950 env[1789]: time="2025-03-17T19:07:15.838938471Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Mar 17 19:07:15.838993 env[1789]: time="2025-03-17T19:07:15.838985373Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Mar 17 19:07:15.840586 env[1789]: time="2025-03-17T19:07:15.840552817Z" level=info msg="parsed scheme: \"unix\"" module=grpc Mar 17 19:07:15.840586 env[1789]: time="2025-03-17T19:07:15.840570118Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Mar 17 19:07:15.840586 env[1789]: time="2025-03-17T19:07:15.840586318Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Mar 17 19:07:15.840772 env[1789]: time="2025-03-17T19:07:15.840596819Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Mar 17 19:07:15.848265 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport46693542-merged.mount: Deactivated successfully. Mar 17 19:07:15.902055 env[1789]: time="2025-03-17T19:07:15.902016875Z" level=info msg="Loading containers: start." Mar 17 19:07:16.092422 kernel: Initializing XFRM netlink socket Mar 17 19:07:16.113149 env[1789]: time="2025-03-17T19:07:16.113098522Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Mar 17 19:07:16.212074 systemd-networkd[1571]: docker0: Link UP Mar 17 19:07:16.232542 env[1789]: time="2025-03-17T19:07:16.232502742Z" level=info msg="Loading containers: done." Mar 17 19:07:16.247877 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1041781508-merged.mount: Deactivated successfully. Mar 17 19:07:16.257859 env[1789]: time="2025-03-17T19:07:16.257803346Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 17 19:07:16.258055 env[1789]: time="2025-03-17T19:07:16.258036952Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Mar 17 19:07:16.258177 env[1789]: time="2025-03-17T19:07:16.258153556Z" level=info msg="Daemon has completed initialization" Mar 17 19:07:16.283845 systemd[1]: Started docker.service. Mar 17 19:07:16.288066 env[1789]: time="2025-03-17T19:07:16.288023886Z" level=info msg="API listen on /run/docker.sock" Mar 17 19:07:17.900967 env[1416]: time="2025-03-17T19:07:17.900913447Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\"" Mar 17 19:07:18.698949 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount600698793.mount: Deactivated successfully. Mar 17 19:07:19.629081 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Mar 17 19:07:19.629337 systemd[1]: Stopped kubelet.service. Mar 17 19:07:19.631245 systemd[1]: Starting kubelet.service... Mar 17 19:07:19.748206 systemd[1]: Started kubelet.service. Mar 17 19:07:19.812489 kubelet[1909]: E0317 19:07:19.812439 1909 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:07:19.814190 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:07:19.814348 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:07:20.375416 env[1416]: time="2025-03-17T19:07:20.375350360Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:20.380522 env[1416]: time="2025-03-17T19:07:20.380478188Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:20.383769 env[1416]: time="2025-03-17T19:07:20.383731469Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:20.386631 env[1416]: time="2025-03-17T19:07:20.386595540Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:22c19cc70fe5806d0a2cb28a6b6b33fd34e6f9e50616bdf6d53649bcfafbc277,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:20.387319 env[1416]: time="2025-03-17T19:07:20.387285157Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\" returns image reference \"sha256:f084bc047a8cf7c8484d47c51e70e646dde3977d916f282feb99207b7b9241af\"" Mar 17 19:07:20.389073 env[1416]: time="2025-03-17T19:07:20.389047301Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\"" Mar 17 19:07:21.316061 waagent[1611]: 2025-03-17T19:07:21.315969Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 2] Mar 17 19:07:21.327023 waagent[1611]: 2025-03-17T19:07:21.326957Z INFO ExtHandler Mar 17 19:07:21.327357 waagent[1611]: 2025-03-17T19:07:21.327305Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: cb36ce82-a216-4134-9321-134b71e0f700 eTag: 17046717506950291787 source: Fabric] Mar 17 19:07:21.328337 waagent[1611]: 2025-03-17T19:07:21.328281Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Mar 17 19:07:21.329807 waagent[1611]: 2025-03-17T19:07:21.329748Z INFO ExtHandler Mar 17 19:07:21.330042 waagent[1611]: 2025-03-17T19:07:21.329995Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 2] Mar 17 19:07:21.414608 waagent[1611]: 2025-03-17T19:07:21.414540Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Mar 17 19:07:21.516453 waagent[1611]: 2025-03-17T19:07:21.516290Z INFO ExtHandler Downloaded certificate {'thumbprint': '5F315EC6781F3F939053C3B61909605342E677AA', 'hasPrivateKey': True} Mar 17 19:07:21.517860 waagent[1611]: 2025-03-17T19:07:21.517792Z INFO ExtHandler Downloaded certificate {'thumbprint': '234E1AFB2B595C3A0DEA6FE8A9A8243568790FE9', 'hasPrivateKey': False} Mar 17 19:07:21.519196 waagent[1611]: 2025-03-17T19:07:21.519134Z INFO ExtHandler Fetch goal state completed Mar 17 19:07:21.520440 waagent[1611]: 2025-03-17T19:07:21.520357Z INFO ExtHandler ExtHandler VM enabled for RSM updates, switching to RSM update mode Mar 17 19:07:21.521985 waagent[1611]: 2025-03-17T19:07:21.521927Z INFO ExtHandler ExtHandler Mar 17 19:07:21.522241 waagent[1611]: 2025-03-17T19:07:21.522187Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_2 channel: WireServer source: Fabric activity: 9879b8a5-9f79-49e8-90a7-f75e1f0af8c7 correlation d1eb894e-54ce-442a-8a3f-0668f8f16d35 created: 2025-03-17T19:07:15.473679Z] Mar 17 19:07:21.523225 waagent[1611]: 2025-03-17T19:07:21.523166Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Mar 17 19:07:21.525637 waagent[1611]: 2025-03-17T19:07:21.525578Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_2 3 ms] Mar 17 19:07:22.100663 env[1416]: time="2025-03-17T19:07:22.100608201Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:22.108308 env[1416]: time="2025-03-17T19:07:22.108259482Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:22.112916 env[1416]: time="2025-03-17T19:07:22.112876390Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:22.118752 env[1416]: time="2025-03-17T19:07:22.118576225Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:6abe7a0accecf29db6ebab18a10f844678ffed693d79e2e51a18a6f2b4530cbb,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:22.119098 env[1416]: time="2025-03-17T19:07:22.119065636Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\" returns image reference \"sha256:652dcad615a9a0c252c253860d5b5b7bfebd3efe159dc033a8555bc15a6d1985\"" Mar 17 19:07:22.120289 env[1416]: time="2025-03-17T19:07:22.120235564Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\"" Mar 17 19:07:23.502752 env[1416]: time="2025-03-17T19:07:23.502693227Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:23.507632 env[1416]: time="2025-03-17T19:07:23.507593539Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:23.513906 env[1416]: time="2025-03-17T19:07:23.513872083Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:23.516774 env[1416]: time="2025-03-17T19:07:23.516741349Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:fb80249bcb77ee72b1c9fa5b70bc28a83ed107c9ca71957841ad91db379963bf,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:23.517394 env[1416]: time="2025-03-17T19:07:23.517345163Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\" returns image reference \"sha256:7f1f6a63d8aa14cf61d0045e912ad312b4ade24637cecccc933b163582eae68c\"" Mar 17 19:07:23.518041 env[1416]: time="2025-03-17T19:07:23.518014878Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\"" Mar 17 19:07:24.734458 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount316519467.mount: Deactivated successfully. Mar 17 19:07:25.366598 env[1416]: time="2025-03-17T19:07:25.366541222Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:25.372398 env[1416]: time="2025-03-17T19:07:25.372350548Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:25.376243 env[1416]: time="2025-03-17T19:07:25.376209332Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.31.7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:25.379709 env[1416]: time="2025-03-17T19:07:25.379676907Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:e5839270c96c3ad1bea1dce4935126d3281297527f3655408d2970aa4b5cf178,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:25.380043 env[1416]: time="2025-03-17T19:07:25.380011615Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\" returns image reference \"sha256:dcfc039c372ea285997a302d60e58a75b80905b4c4dba969993b9b22e8ac66d1\"" Mar 17 19:07:25.380734 env[1416]: time="2025-03-17T19:07:25.380707730Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Mar 17 19:07:25.905498 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3329449399.mount: Deactivated successfully. Mar 17 19:07:27.129856 env[1416]: time="2025-03-17T19:07:27.129799019Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.136114 env[1416]: time="2025-03-17T19:07:27.136065348Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.140997 env[1416]: time="2025-03-17T19:07:27.140955649Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.145630 env[1416]: time="2025-03-17T19:07:27.145592444Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.146357 env[1416]: time="2025-03-17T19:07:27.146323959Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Mar 17 19:07:27.146955 env[1416]: time="2025-03-17T19:07:27.146929172Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 17 19:07:27.677079 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2198197341.mount: Deactivated successfully. Mar 17 19:07:27.699784 env[1416]: time="2025-03-17T19:07:27.699729760Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.706465 env[1416]: time="2025-03-17T19:07:27.706407098Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.710233 env[1416]: time="2025-03-17T19:07:27.710190975Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.713869 env[1416]: time="2025-03-17T19:07:27.713827350Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:27.714256 env[1416]: time="2025-03-17T19:07:27.714225059Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Mar 17 19:07:27.714978 env[1416]: time="2025-03-17T19:07:27.714950374Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Mar 17 19:07:28.281889 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount330313398.mount: Deactivated successfully. Mar 17 19:07:29.879070 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Mar 17 19:07:29.879339 systemd[1]: Stopped kubelet.service. Mar 17 19:07:29.881165 systemd[1]: Starting kubelet.service... Mar 17 19:07:29.998802 systemd[1]: Started kubelet.service. Mar 17 19:07:30.059007 kubelet[1926]: E0317 19:07:30.058953 1926 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 19:07:30.060743 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 19:07:30.060908 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 19:07:30.898209 env[1416]: time="2025-03-17T19:07:30.898155250Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:30.904838 env[1416]: time="2025-03-17T19:07:30.904790176Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:30.907972 env[1416]: time="2025-03-17T19:07:30.907937236Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:30.911590 env[1416]: time="2025-03-17T19:07:30.911560505Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:30.912268 env[1416]: time="2025-03-17T19:07:30.912237218Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Mar 17 19:07:33.337535 systemd[1]: Stopped kubelet.service. Mar 17 19:07:33.340676 systemd[1]: Starting kubelet.service... Mar 17 19:07:33.382539 systemd[1]: Reloading. Mar 17 19:07:33.476215 /usr/lib/systemd/system-generators/torcx-generator[1974]: time="2025-03-17T19:07:33Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 19:07:33.476255 /usr/lib/systemd/system-generators/torcx-generator[1974]: time="2025-03-17T19:07:33Z" level=info msg="torcx already run" Mar 17 19:07:33.577729 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 19:07:33.577751 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 19:07:33.598455 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 19:07:33.767804 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 17 19:07:33.767930 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 17 19:07:33.768276 systemd[1]: Stopped kubelet.service. Mar 17 19:07:33.770781 systemd[1]: Starting kubelet.service... Mar 17 19:07:35.020175 systemd[1]: Started kubelet.service. Mar 17 19:07:35.062803 kubelet[2041]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 19:07:35.062803 kubelet[2041]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 19:07:35.062803 kubelet[2041]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 19:07:35.063335 kubelet[2041]: I0317 19:07:35.062876 2041 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 19:07:35.512601 kubelet[2041]: I0317 19:07:35.512559 2041 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Mar 17 19:07:35.512601 kubelet[2041]: I0317 19:07:35.512589 2041 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 19:07:35.513651 kubelet[2041]: I0317 19:07:35.513625 2041 server.go:929] "Client rotation is on, will bootstrap in background" Mar 17 19:07:35.846700 kubelet[2041]: E0317 19:07:35.846371 2041 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.200.8.15:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:35.850330 kubelet[2041]: I0317 19:07:35.850281 2041 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 19:07:35.858921 kubelet[2041]: E0317 19:07:35.858887 2041 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 17 19:07:35.859165 kubelet[2041]: I0317 19:07:35.859143 2041 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 17 19:07:35.863877 kubelet[2041]: I0317 19:07:35.863851 2041 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 17 19:07:35.865180 kubelet[2041]: I0317 19:07:35.865155 2041 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 17 19:07:35.865412 kubelet[2041]: I0317 19:07:35.865350 2041 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 19:07:35.865602 kubelet[2041]: I0317 19:07:35.865412 2041 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510.3.7-a-a60daf7667","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 17 19:07:35.865763 kubelet[2041]: I0317 19:07:35.865609 2041 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 19:07:35.865763 kubelet[2041]: I0317 19:07:35.865622 2041 container_manager_linux.go:300] "Creating device plugin manager" Mar 17 19:07:35.865763 kubelet[2041]: I0317 19:07:35.865744 2041 state_mem.go:36] "Initialized new in-memory state store" Mar 17 19:07:35.870987 kubelet[2041]: I0317 19:07:35.870959 2041 kubelet.go:408] "Attempting to sync node with API server" Mar 17 19:07:35.871103 kubelet[2041]: I0317 19:07:35.870994 2041 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 19:07:35.871103 kubelet[2041]: I0317 19:07:35.871036 2041 kubelet.go:314] "Adding apiserver pod source" Mar 17 19:07:35.871103 kubelet[2041]: I0317 19:07:35.871050 2041 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 19:07:35.876134 kubelet[2041]: W0317 19:07:35.876076 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.8.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.7-a-a60daf7667&limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:35.876295 kubelet[2041]: E0317 19:07:35.876273 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.8.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.7-a-a60daf7667&limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:35.886105 kubelet[2041]: W0317 19:07:35.885804 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.8.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:35.886105 kubelet[2041]: E0317 19:07:35.885893 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.8.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:35.890666 kubelet[2041]: I0317 19:07:35.886180 2041 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Mar 17 19:07:35.890666 kubelet[2041]: I0317 19:07:35.890427 2041 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 19:07:35.895661 kubelet[2041]: W0317 19:07:35.895629 2041 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 17 19:07:35.896492 kubelet[2041]: I0317 19:07:35.896469 2041 server.go:1269] "Started kubelet" Mar 17 19:07:35.896934 kubelet[2041]: I0317 19:07:35.896876 2041 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 19:07:35.904219 kubelet[2041]: I0317 19:07:35.903981 2041 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 19:07:35.904433 kubelet[2041]: I0317 19:07:35.904412 2041 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 19:07:35.906077 kubelet[2041]: E0317 19:07:35.904619 2041 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.8.15:6443/api/v1/namespaces/default/events\": dial tcp 10.200.8.15:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-3510.3.7-a-a60daf7667.182daca515bb3e34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-3510.3.7-a-a60daf7667,UID:ci-3510.3.7-a-a60daf7667,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-3510.3.7-a-a60daf7667,},FirstTimestamp:2025-03-17 19:07:35.896423988 +0000 UTC m=+0.870614170,LastTimestamp:2025-03-17 19:07:35.896423988 +0000 UTC m=+0.870614170,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-3510.3.7-a-a60daf7667,}" Mar 17 19:07:35.910399 kubelet[2041]: I0317 19:07:35.907144 2041 server.go:460] "Adding debug handlers to kubelet server" Mar 17 19:07:35.911238 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Mar 17 19:07:35.911348 kubelet[2041]: I0317 19:07:35.911173 2041 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 19:07:35.913656 kubelet[2041]: E0317 19:07:35.913636 2041 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 17 19:07:35.914054 kubelet[2041]: I0317 19:07:35.914034 2041 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 17 19:07:35.917493 kubelet[2041]: I0317 19:07:35.917467 2041 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 17 19:07:35.917742 kubelet[2041]: E0317 19:07:35.917722 2041 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-3510.3.7-a-a60daf7667\" not found" Mar 17 19:07:35.919024 kubelet[2041]: I0317 19:07:35.918999 2041 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 17 19:07:35.919112 kubelet[2041]: I0317 19:07:35.919069 2041 reconciler.go:26] "Reconciler: start to sync state" Mar 17 19:07:35.919195 kubelet[2041]: E0317 19:07:35.919163 2041 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.8.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.7-a-a60daf7667?timeout=10s\": dial tcp 10.200.8.15:6443: connect: connection refused" interval="200ms" Mar 17 19:07:35.919456 kubelet[2041]: I0317 19:07:35.919433 2041 factory.go:221] Registration of the systemd container factory successfully Mar 17 19:07:35.919537 kubelet[2041]: I0317 19:07:35.919521 2041 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 17 19:07:35.921505 kubelet[2041]: I0317 19:07:35.921481 2041 factory.go:221] Registration of the containerd container factory successfully Mar 17 19:07:35.936104 kubelet[2041]: W0317 19:07:35.936049 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.8.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:35.936279 kubelet[2041]: E0317 19:07:35.936117 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.8.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:35.963995 kubelet[2041]: I0317 19:07:35.963956 2041 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 17 19:07:35.963995 kubelet[2041]: I0317 19:07:35.963981 2041 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 17 19:07:35.964244 kubelet[2041]: I0317 19:07:35.964007 2041 state_mem.go:36] "Initialized new in-memory state store" Mar 17 19:07:35.968426 kubelet[2041]: I0317 19:07:35.968398 2041 policy_none.go:49] "None policy: Start" Mar 17 19:07:35.969251 kubelet[2041]: I0317 19:07:35.969218 2041 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 19:07:35.969251 kubelet[2041]: I0317 19:07:35.969251 2041 state_mem.go:35] "Initializing new in-memory state store" Mar 17 19:07:35.977479 systemd[1]: Created slice kubepods.slice. Mar 17 19:07:35.981956 systemd[1]: Created slice kubepods-burstable.slice. Mar 17 19:07:35.985235 systemd[1]: Created slice kubepods-besteffort.slice. Mar 17 19:07:35.991079 kubelet[2041]: I0317 19:07:35.991060 2041 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 19:07:35.991323 kubelet[2041]: I0317 19:07:35.991314 2041 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 17 19:07:35.991443 kubelet[2041]: I0317 19:07:35.991413 2041 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 19:07:35.993501 kubelet[2041]: I0317 19:07:35.993486 2041 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 19:07:35.994880 kubelet[2041]: E0317 19:07:35.994530 2041 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-3510.3.7-a-a60daf7667\" not found" Mar 17 19:07:36.019171 kubelet[2041]: I0317 19:07:36.019111 2041 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 19:07:36.021532 kubelet[2041]: I0317 19:07:36.021505 2041 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 19:07:36.021740 kubelet[2041]: I0317 19:07:36.021723 2041 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 19:07:36.021828 kubelet[2041]: I0317 19:07:36.021772 2041 kubelet.go:2321] "Starting kubelet main sync loop" Mar 17 19:07:36.021887 kubelet[2041]: E0317 19:07:36.021832 2041 kubelet.go:2345] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Mar 17 19:07:36.022921 kubelet[2041]: W0317 19:07:36.022897 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.8.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:36.023085 kubelet[2041]: E0317 19:07:36.023067 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.8.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:36.093718 kubelet[2041]: I0317 19:07:36.093670 2041 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.094136 kubelet[2041]: E0317 19:07:36.094061 2041 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.8.15:6443/api/v1/nodes\": dial tcp 10.200.8.15:6443: connect: connection refused" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.120145 kubelet[2041]: E0317 19:07:36.119991 2041 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.8.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.7-a-a60daf7667?timeout=10s\": dial tcp 10.200.8.15:6443: connect: connection refused" interval="400ms" Mar 17 19:07:36.131097 systemd[1]: Created slice kubepods-burstable-pode92fe45c7cfc2f23b1bb39feb06b5aa6.slice. Mar 17 19:07:36.147823 systemd[1]: Created slice kubepods-burstable-pod16c7ada29ea6e9ec2982b3924b087dc1.slice. Mar 17 19:07:36.151520 systemd[1]: Created slice kubepods-burstable-podef206eb641edff80b4e41fe9729d5889.slice. Mar 17 19:07:36.221747 kubelet[2041]: I0317 19:07:36.221691 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ef206eb641edff80b4e41fe9729d5889-k8s-certs\") pod \"kube-apiserver-ci-3510.3.7-a-a60daf7667\" (UID: \"ef206eb641edff80b4e41fe9729d5889\") " pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.221747 kubelet[2041]: I0317 19:07:36.221741 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ef206eb641edff80b4e41fe9729d5889-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.7-a-a60daf7667\" (UID: \"ef206eb641edff80b4e41fe9729d5889\") " pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.222006 kubelet[2041]: I0317 19:07:36.221768 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-ca-certs\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.222006 kubelet[2041]: I0317 19:07:36.221792 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.222006 kubelet[2041]: I0317 19:07:36.221816 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.222006 kubelet[2041]: I0317 19:07:36.221838 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ef206eb641edff80b4e41fe9729d5889-ca-certs\") pod \"kube-apiserver-ci-3510.3.7-a-a60daf7667\" (UID: \"ef206eb641edff80b4e41fe9729d5889\") " pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.222006 kubelet[2041]: I0317 19:07:36.221860 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.222223 kubelet[2041]: I0317 19:07:36.221917 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.222223 kubelet[2041]: I0317 19:07:36.221945 2041 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/16c7ada29ea6e9ec2982b3924b087dc1-kubeconfig\") pod \"kube-scheduler-ci-3510.3.7-a-a60daf7667\" (UID: \"16c7ada29ea6e9ec2982b3924b087dc1\") " pod="kube-system/kube-scheduler-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.296693 kubelet[2041]: I0317 19:07:36.296654 2041 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.297097 kubelet[2041]: E0317 19:07:36.297061 2041 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.8.15:6443/api/v1/nodes\": dial tcp 10.200.8.15:6443: connect: connection refused" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.446660 env[1416]: time="2025-03-17T19:07:36.446602909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.7-a-a60daf7667,Uid:e92fe45c7cfc2f23b1bb39feb06b5aa6,Namespace:kube-system,Attempt:0,}" Mar 17 19:07:36.451307 env[1416]: time="2025-03-17T19:07:36.451085282Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.7-a-a60daf7667,Uid:16c7ada29ea6e9ec2982b3924b087dc1,Namespace:kube-system,Attempt:0,}" Mar 17 19:07:36.454476 env[1416]: time="2025-03-17T19:07:36.454435236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.7-a-a60daf7667,Uid:ef206eb641edff80b4e41fe9729d5889,Namespace:kube-system,Attempt:0,}" Mar 17 19:07:36.520483 kubelet[2041]: E0317 19:07:36.520436 2041 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.8.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.7-a-a60daf7667?timeout=10s\": dial tcp 10.200.8.15:6443: connect: connection refused" interval="800ms" Mar 17 19:07:36.699409 kubelet[2041]: I0317 19:07:36.699008 2041 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.699682 kubelet[2041]: E0317 19:07:36.699649 2041 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.8.15:6443/api/v1/nodes\": dial tcp 10.200.8.15:6443: connect: connection refused" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:36.906516 kubelet[2041]: W0317 19:07:36.906389 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.8.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:36.906516 kubelet[2041]: E0317 19:07:36.906481 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.8.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:37.039661 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1770312548.mount: Deactivated successfully. Mar 17 19:07:37.057824 kubelet[2041]: W0317 19:07:37.057700 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.8.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:37.057824 kubelet[2041]: E0317 19:07:37.057760 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.8.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:37.075580 env[1416]: time="2025-03-17T19:07:37.075523840Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.079697 env[1416]: time="2025-03-17T19:07:37.079647006Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.088718 env[1416]: time="2025-03-17T19:07:37.088674749Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.092395 env[1416]: time="2025-03-17T19:07:37.092345108Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.096079 env[1416]: time="2025-03-17T19:07:37.096038867Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.099894 env[1416]: time="2025-03-17T19:07:37.099855527Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.103828 env[1416]: time="2025-03-17T19:07:37.103783690Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.107308 env[1416]: time="2025-03-17T19:07:37.107269745Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.114746 env[1416]: time="2025-03-17T19:07:37.114703063Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.118584 env[1416]: time="2025-03-17T19:07:37.118547825Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.131005 env[1416]: time="2025-03-17T19:07:37.130951822Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.142103 env[1416]: time="2025-03-17T19:07:37.142051599Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:37.194535 env[1416]: time="2025-03-17T19:07:37.194465333Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:07:37.194758 env[1416]: time="2025-03-17T19:07:37.194503033Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:07:37.194758 env[1416]: time="2025-03-17T19:07:37.194517933Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:07:37.194758 env[1416]: time="2025-03-17T19:07:37.194691236Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/f3153e890fe6eb69654480b794da2dce37729b55f74df13f92f72651c044329b pid=2080 runtime=io.containerd.runc.v2 Mar 17 19:07:37.220997 env[1416]: time="2025-03-17T19:07:37.220926954Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:07:37.221233 env[1416]: time="2025-03-17T19:07:37.221202158Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:07:37.221356 env[1416]: time="2025-03-17T19:07:37.221333260Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:07:37.221663 env[1416]: time="2025-03-17T19:07:37.221630765Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/852b902aa7d354728b4a7511c81ee56b766aa43ecb8081a259448841fcd1613c pid=2105 runtime=io.containerd.runc.v2 Mar 17 19:07:37.222855 systemd[1]: Started cri-containerd-f3153e890fe6eb69654480b794da2dce37729b55f74df13f92f72651c044329b.scope. Mar 17 19:07:37.252881 env[1416]: time="2025-03-17T19:07:37.247806781Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:07:37.252881 env[1416]: time="2025-03-17T19:07:37.247848482Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:07:37.252881 env[1416]: time="2025-03-17T19:07:37.247863782Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:07:37.252881 env[1416]: time="2025-03-17T19:07:37.248006185Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ab9a77f3117785cafd4ea3cf61fa2a52e09d0c86b68228ac7326f85fd20a73cf pid=2125 runtime=io.containerd.runc.v2 Mar 17 19:07:37.259329 systemd[1]: Started cri-containerd-852b902aa7d354728b4a7511c81ee56b766aa43ecb8081a259448841fcd1613c.scope. Mar 17 19:07:37.281083 systemd[1]: Started cri-containerd-ab9a77f3117785cafd4ea3cf61fa2a52e09d0c86b68228ac7326f85fd20a73cf.scope. Mar 17 19:07:37.322405 kubelet[2041]: E0317 19:07:37.321182 2041 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.8.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.7-a-a60daf7667?timeout=10s\": dial tcp 10.200.8.15:6443: connect: connection refused" interval="1.6s" Mar 17 19:07:37.340542 env[1416]: time="2025-03-17T19:07:37.340491456Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.7-a-a60daf7667,Uid:ef206eb641edff80b4e41fe9729d5889,Namespace:kube-system,Attempt:0,} returns sandbox id \"f3153e890fe6eb69654480b794da2dce37729b55f74df13f92f72651c044329b\"" Mar 17 19:07:37.355503 env[1416]: time="2025-03-17T19:07:37.355454894Z" level=info msg="CreateContainer within sandbox \"f3153e890fe6eb69654480b794da2dce37729b55f74df13f92f72651c044329b\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 17 19:07:37.364555 env[1416]: time="2025-03-17T19:07:37.364291735Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.7-a-a60daf7667,Uid:16c7ada29ea6e9ec2982b3924b087dc1,Namespace:kube-system,Attempt:0,} returns sandbox id \"852b902aa7d354728b4a7511c81ee56b766aa43ecb8081a259448841fcd1613c\"" Mar 17 19:07:37.372260 env[1416]: time="2025-03-17T19:07:37.371951157Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.7-a-a60daf7667,Uid:e92fe45c7cfc2f23b1bb39feb06b5aa6,Namespace:kube-system,Attempt:0,} returns sandbox id \"ab9a77f3117785cafd4ea3cf61fa2a52e09d0c86b68228ac7326f85fd20a73cf\"" Mar 17 19:07:37.372457 env[1416]: time="2025-03-17T19:07:37.372398264Z" level=info msg="CreateContainer within sandbox \"852b902aa7d354728b4a7511c81ee56b766aa43ecb8081a259448841fcd1613c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 17 19:07:37.374815 env[1416]: time="2025-03-17T19:07:37.374783102Z" level=info msg="CreateContainer within sandbox \"ab9a77f3117785cafd4ea3cf61fa2a52e09d0c86b68228ac7326f85fd20a73cf\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 17 19:07:37.407101 env[1416]: time="2025-03-17T19:07:37.407043415Z" level=info msg="CreateContainer within sandbox \"f3153e890fe6eb69654480b794da2dce37729b55f74df13f92f72651c044329b\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9388f30ae0a277b3a9df44c8345bc725cc045de7f9609d5cb65c0480b8d99967\"" Mar 17 19:07:37.407908 env[1416]: time="2025-03-17T19:07:37.407846628Z" level=info msg="StartContainer for \"9388f30ae0a277b3a9df44c8345bc725cc045de7f9609d5cb65c0480b8d99967\"" Mar 17 19:07:37.428156 systemd[1]: Started cri-containerd-9388f30ae0a277b3a9df44c8345bc725cc045de7f9609d5cb65c0480b8d99967.scope. Mar 17 19:07:37.439528 env[1416]: time="2025-03-17T19:07:37.439479031Z" level=info msg="CreateContainer within sandbox \"852b902aa7d354728b4a7511c81ee56b766aa43ecb8081a259448841fcd1613c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"c9c3be76e13e342d402b8d9d6701d95009d4347345d5c7c745f1c4cd546e3904\"" Mar 17 19:07:37.443371 env[1416]: time="2025-03-17T19:07:37.441548964Z" level=info msg="StartContainer for \"c9c3be76e13e342d402b8d9d6701d95009d4347345d5c7c745f1c4cd546e3904\"" Mar 17 19:07:37.445662 kubelet[2041]: W0317 19:07:37.445500 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.8.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.7-a-a60daf7667&limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:37.445662 kubelet[2041]: E0317 19:07:37.445603 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.8.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.7-a-a60daf7667&limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:37.457402 env[1416]: time="2025-03-17T19:07:37.453762858Z" level=info msg="CreateContainer within sandbox \"ab9a77f3117785cafd4ea3cf61fa2a52e09d0c86b68228ac7326f85fd20a73cf\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b95102badbabd8f886206ac3d6ce655b2d2807a80166b0312591d92caae9847a\"" Mar 17 19:07:37.457402 env[1416]: time="2025-03-17T19:07:37.454291567Z" level=info msg="StartContainer for \"b95102badbabd8f886206ac3d6ce655b2d2807a80166b0312591d92caae9847a\"" Mar 17 19:07:37.486349 systemd[1]: Started cri-containerd-c9c3be76e13e342d402b8d9d6701d95009d4347345d5c7c745f1c4cd546e3904.scope. Mar 17 19:07:37.497427 systemd[1]: Started cri-containerd-b95102badbabd8f886206ac3d6ce655b2d2807a80166b0312591d92caae9847a.scope. Mar 17 19:07:37.501927 kubelet[2041]: I0317 19:07:37.501896 2041 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:37.502256 kubelet[2041]: E0317 19:07:37.502219 2041 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.8.15:6443/api/v1/nodes\": dial tcp 10.200.8.15:6443: connect: connection refused" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:37.527222 env[1416]: time="2025-03-17T19:07:37.527162226Z" level=info msg="StartContainer for \"9388f30ae0a277b3a9df44c8345bc725cc045de7f9609d5cb65c0480b8d99967\" returns successfully" Mar 17 19:07:37.570636 env[1416]: time="2025-03-17T19:07:37.570575717Z" level=info msg="StartContainer for \"c9c3be76e13e342d402b8d9d6701d95009d4347345d5c7c745f1c4cd546e3904\" returns successfully" Mar 17 19:07:37.575825 kubelet[2041]: W0317 19:07:37.575674 2041 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.8.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.8.15:6443: connect: connection refused Mar 17 19:07:37.575825 kubelet[2041]: E0317 19:07:37.575761 2041 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.8.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.8.15:6443: connect: connection refused" logger="UnhandledError" Mar 17 19:07:37.630205 env[1416]: time="2025-03-17T19:07:37.630145865Z" level=info msg="StartContainer for \"b95102badbabd8f886206ac3d6ce655b2d2807a80166b0312591d92caae9847a\" returns successfully" Mar 17 19:07:39.104076 kubelet[2041]: I0317 19:07:39.104038 2041 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:39.731041 kubelet[2041]: E0317 19:07:39.730991 2041 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-3510.3.7-a-a60daf7667\" not found" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:39.877997 kubelet[2041]: I0317 19:07:39.877952 2041 apiserver.go:52] "Watching apiserver" Mar 17 19:07:39.895194 kubelet[2041]: I0317 19:07:39.895156 2041 kubelet_node_status.go:75] "Successfully registered node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:39.919897 kubelet[2041]: I0317 19:07:39.919856 2041 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 17 19:07:40.488095 kubelet[2041]: E0317 19:07:40.488056 2041 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-3510.3.7-a-a60daf7667\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:42.638526 systemd[1]: Reloading. Mar 17 19:07:42.726230 /usr/lib/systemd/system-generators/torcx-generator[2330]: time="2025-03-17T19:07:42Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 19:07:42.729748 /usr/lib/systemd/system-generators/torcx-generator[2330]: time="2025-03-17T19:07:42Z" level=info msg="torcx already run" Mar 17 19:07:42.825034 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 19:07:42.825057 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 19:07:42.841986 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 19:07:42.956264 systemd[1]: Stopping kubelet.service... Mar 17 19:07:42.978887 systemd[1]: kubelet.service: Deactivated successfully. Mar 17 19:07:42.979103 systemd[1]: Stopped kubelet.service. Mar 17 19:07:42.981220 systemd[1]: Starting kubelet.service... Mar 17 19:07:43.336583 systemd[1]: Started kubelet.service. Mar 17 19:07:43.400706 kubelet[2396]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 19:07:43.400706 kubelet[2396]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 19:07:43.400706 kubelet[2396]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 19:07:43.401198 kubelet[2396]: I0317 19:07:43.400790 2396 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 19:07:43.412401 kubelet[2396]: I0317 19:07:43.412348 2396 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Mar 17 19:07:43.412401 kubelet[2396]: I0317 19:07:43.412394 2396 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 19:07:43.413363 kubelet[2396]: I0317 19:07:43.413328 2396 server.go:929] "Client rotation is on, will bootstrap in background" Mar 17 19:07:43.418818 kubelet[2396]: I0317 19:07:43.418784 2396 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 17 19:07:43.421907 kubelet[2396]: I0317 19:07:43.421876 2396 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 19:07:43.429448 kubelet[2396]: E0317 19:07:43.429398 2396 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Mar 17 19:07:43.429686 kubelet[2396]: I0317 19:07:43.429669 2396 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Mar 17 19:07:43.435294 kubelet[2396]: I0317 19:07:43.435264 2396 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 17 19:07:43.435614 kubelet[2396]: I0317 19:07:43.435456 2396 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 17 19:07:43.435614 kubelet[2396]: I0317 19:07:43.435586 2396 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 19:07:43.435886 kubelet[2396]: I0317 19:07:43.435628 2396 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510.3.7-a-a60daf7667","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 17 19:07:43.436046 kubelet[2396]: I0317 19:07:43.435897 2396 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 19:07:43.436046 kubelet[2396]: I0317 19:07:43.435913 2396 container_manager_linux.go:300] "Creating device plugin manager" Mar 17 19:07:43.436046 kubelet[2396]: I0317 19:07:43.435959 2396 state_mem.go:36] "Initialized new in-memory state store" Mar 17 19:07:43.436187 kubelet[2396]: I0317 19:07:43.436098 2396 kubelet.go:408] "Attempting to sync node with API server" Mar 17 19:07:43.436187 kubelet[2396]: I0317 19:07:43.436113 2396 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 19:07:43.436448 kubelet[2396]: I0317 19:07:43.436432 2396 kubelet.go:314] "Adding apiserver pod source" Mar 17 19:07:43.436561 kubelet[2396]: I0317 19:07:43.436551 2396 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 19:07:43.439370 kubelet[2396]: I0317 19:07:43.439343 2396 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Mar 17 19:07:43.440027 kubelet[2396]: I0317 19:07:43.440003 2396 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 19:07:43.440733 kubelet[2396]: I0317 19:07:43.440708 2396 server.go:1269] "Started kubelet" Mar 17 19:07:43.449661 kubelet[2396]: I0317 19:07:43.449635 2396 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 19:07:43.456961 kubelet[2396]: I0317 19:07:43.456916 2396 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 19:07:43.458097 kubelet[2396]: I0317 19:07:43.458065 2396 server.go:460] "Adding debug handlers to kubelet server" Mar 17 19:07:43.468354 kubelet[2396]: E0317 19:07:43.467452 2396 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-3510.3.7-a-a60daf7667\" not found" Mar 17 19:07:43.472483 kubelet[2396]: I0317 19:07:43.467490 2396 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 19:07:43.472762 kubelet[2396]: I0317 19:07:43.472742 2396 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 19:07:43.480191 kubelet[2396]: I0317 19:07:43.480162 2396 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 17 19:07:43.480622 kubelet[2396]: I0317 19:07:43.480591 2396 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 17 19:07:43.481424 kubelet[2396]: I0317 19:07:43.481389 2396 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 19:07:43.482946 kubelet[2396]: I0317 19:07:43.482782 2396 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 19:07:43.482946 kubelet[2396]: I0317 19:07:43.482819 2396 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 19:07:43.482946 kubelet[2396]: I0317 19:07:43.482840 2396 kubelet.go:2321] "Starting kubelet main sync loop" Mar 17 19:07:43.482946 kubelet[2396]: E0317 19:07:43.482884 2396 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 19:07:43.491194 kubelet[2396]: I0317 19:07:43.491172 2396 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 17 19:07:43.492118 kubelet[2396]: I0317 19:07:43.492096 2396 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 17 19:07:43.492478 kubelet[2396]: I0317 19:07:43.492460 2396 reconciler.go:26] "Reconciler: start to sync state" Mar 17 19:07:43.498526 kubelet[2396]: E0317 19:07:43.498497 2396 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 17 19:07:43.503612 kubelet[2396]: I0317 19:07:43.503588 2396 factory.go:221] Registration of the containerd container factory successfully Mar 17 19:07:43.503780 kubelet[2396]: I0317 19:07:43.503768 2396 factory.go:221] Registration of the systemd container factory successfully Mar 17 19:07:43.545013 kubelet[2396]: I0317 19:07:43.544983 2396 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 17 19:07:43.545013 kubelet[2396]: I0317 19:07:43.544999 2396 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 17 19:07:43.545013 kubelet[2396]: I0317 19:07:43.545021 2396 state_mem.go:36] "Initialized new in-memory state store" Mar 17 19:07:43.545279 kubelet[2396]: I0317 19:07:43.545232 2396 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 17 19:07:43.545279 kubelet[2396]: I0317 19:07:43.545248 2396 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 17 19:07:43.545279 kubelet[2396]: I0317 19:07:43.545271 2396 policy_none.go:49] "None policy: Start" Mar 17 19:07:43.545935 kubelet[2396]: I0317 19:07:43.545915 2396 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 19:07:43.545935 kubelet[2396]: I0317 19:07:43.545938 2396 state_mem.go:35] "Initializing new in-memory state store" Mar 17 19:07:43.546121 kubelet[2396]: I0317 19:07:43.546104 2396 state_mem.go:75] "Updated machine memory state" Mar 17 19:07:43.553033 kubelet[2396]: I0317 19:07:43.552999 2396 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 19:07:43.553211 kubelet[2396]: I0317 19:07:43.553196 2396 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 17 19:07:43.553266 kubelet[2396]: I0317 19:07:43.553209 2396 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 19:07:43.553918 kubelet[2396]: I0317 19:07:43.553897 2396 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 19:07:43.591877 kubelet[2396]: W0317 19:07:43.591769 2396 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 17 19:07:43.595721 kubelet[2396]: W0317 19:07:43.595686 2396 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 17 19:07:43.595911 kubelet[2396]: W0317 19:07:43.595891 2396 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 17 19:07:43.657079 kubelet[2396]: I0317 19:07:43.657046 2396 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.666881 kubelet[2396]: I0317 19:07:43.666849 2396 kubelet_node_status.go:111] "Node was previously registered" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.667063 kubelet[2396]: I0317 19:07:43.666938 2396 kubelet_node_status.go:75] "Successfully registered node" node="ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.692971 kubelet[2396]: I0317 19:07:43.692925 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693240 kubelet[2396]: I0317 19:07:43.693217 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693385 kubelet[2396]: I0317 19:07:43.693354 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/16c7ada29ea6e9ec2982b3924b087dc1-kubeconfig\") pod \"kube-scheduler-ci-3510.3.7-a-a60daf7667\" (UID: \"16c7ada29ea6e9ec2982b3924b087dc1\") " pod="kube-system/kube-scheduler-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693499 kubelet[2396]: I0317 19:07:43.693418 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ef206eb641edff80b4e41fe9729d5889-ca-certs\") pod \"kube-apiserver-ci-3510.3.7-a-a60daf7667\" (UID: \"ef206eb641edff80b4e41fe9729d5889\") " pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693499 kubelet[2396]: I0317 19:07:43.693442 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-ca-certs\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693499 kubelet[2396]: I0317 19:07:43.693466 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693499 kubelet[2396]: I0317 19:07:43.693487 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ef206eb641edff80b4e41fe9729d5889-k8s-certs\") pod \"kube-apiserver-ci-3510.3.7-a-a60daf7667\" (UID: \"ef206eb641edff80b4e41fe9729d5889\") " pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693669 kubelet[2396]: I0317 19:07:43.693513 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ef206eb641edff80b4e41fe9729d5889-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.7-a-a60daf7667\" (UID: \"ef206eb641edff80b4e41fe9729d5889\") " pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:43.693669 kubelet[2396]: I0317 19:07:43.693539 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e92fe45c7cfc2f23b1bb39feb06b5aa6-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.7-a-a60daf7667\" (UID: \"e92fe45c7cfc2f23b1bb39feb06b5aa6\") " pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:44.120458 sudo[2428]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 17 19:07:44.120762 sudo[2428]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Mar 17 19:07:44.447934 kubelet[2396]: I0317 19:07:44.447603 2396 apiserver.go:52] "Watching apiserver" Mar 17 19:07:44.492835 kubelet[2396]: I0317 19:07:44.492797 2396 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 17 19:07:44.557064 kubelet[2396]: W0317 19:07:44.557018 2396 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 17 19:07:44.557359 kubelet[2396]: E0317 19:07:44.557341 2396 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-3510.3.7-a-a60daf7667\" already exists" pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:44.562947 kubelet[2396]: W0317 19:07:44.562920 2396 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Mar 17 19:07:44.563213 kubelet[2396]: E0317 19:07:44.563195 2396 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-3510.3.7-a-a60daf7667\" already exists" pod="kube-system/kube-scheduler-ci-3510.3.7-a-a60daf7667" Mar 17 19:07:44.582038 kubelet[2396]: I0317 19:07:44.581951 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-3510.3.7-a-a60daf7667" podStartSLOduration=1.581914306 podStartE2EDuration="1.581914306s" podCreationTimestamp="2025-03-17 19:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 19:07:44.568829031 +0000 UTC m=+1.223945828" watchObservedRunningTime="2025-03-17 19:07:44.581914306 +0000 UTC m=+1.237031203" Mar 17 19:07:44.602034 kubelet[2396]: I0317 19:07:44.601966 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-3510.3.7-a-a60daf7667" podStartSLOduration=1.601940075 podStartE2EDuration="1.601940075s" podCreationTimestamp="2025-03-17 19:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 19:07:44.594679177 +0000 UTC m=+1.249795974" watchObservedRunningTime="2025-03-17 19:07:44.601940075 +0000 UTC m=+1.257056872" Mar 17 19:07:44.602425 kubelet[2396]: I0317 19:07:44.602391 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-3510.3.7-a-a60daf7667" podStartSLOduration=1.60236058 podStartE2EDuration="1.60236058s" podCreationTimestamp="2025-03-17 19:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 19:07:44.583143122 +0000 UTC m=+1.238259919" watchObservedRunningTime="2025-03-17 19:07:44.60236058 +0000 UTC m=+1.257477377" Mar 17 19:07:44.679540 sudo[2428]: pam_unix(sudo:session): session closed for user root Mar 17 19:07:46.217343 sudo[1779]: pam_unix(sudo:session): session closed for user root Mar 17 19:07:46.318565 sshd[1776]: pam_unix(sshd:session): session closed for user core Mar 17 19:07:46.322216 systemd[1]: sshd@4-10.200.8.15:22-10.200.16.10:43956.service: Deactivated successfully. Mar 17 19:07:46.323077 systemd[1]: session-7.scope: Deactivated successfully. Mar 17 19:07:46.323256 systemd[1]: session-7.scope: Consumed 3.859s CPU time. Mar 17 19:07:46.323809 systemd-logind[1401]: Session 7 logged out. Waiting for processes to exit. Mar 17 19:07:46.324748 systemd-logind[1401]: Removed session 7. Mar 17 19:07:47.909778 kubelet[2396]: I0317 19:07:47.909732 2396 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 17 19:07:47.910490 env[1416]: time="2025-03-17T19:07:47.910451681Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 17 19:07:47.910856 kubelet[2396]: I0317 19:07:47.910669 2396 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 17 19:07:48.699808 systemd[1]: Created slice kubepods-besteffort-podef97debc_db00_4269_956b_9db15d4eb20a.slice. Mar 17 19:07:48.715489 systemd[1]: Created slice kubepods-burstable-pod626fe57b_3983_4ba1_9e43_d893ba16fe35.slice. Mar 17 19:07:48.729562 kubelet[2396]: I0317 19:07:48.729526 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ef97debc-db00-4269-956b-9db15d4eb20a-xtables-lock\") pod \"kube-proxy-kn8bq\" (UID: \"ef97debc-db00-4269-956b-9db15d4eb20a\") " pod="kube-system/kube-proxy-kn8bq" Mar 17 19:07:48.729817 kubelet[2396]: I0317 19:07:48.729794 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g8dv\" (UniqueName: \"kubernetes.io/projected/ef97debc-db00-4269-956b-9db15d4eb20a-kube-api-access-9g8dv\") pod \"kube-proxy-kn8bq\" (UID: \"ef97debc-db00-4269-956b-9db15d4eb20a\") " pod="kube-system/kube-proxy-kn8bq" Mar 17 19:07:48.729921 kubelet[2396]: I0317 19:07:48.729907 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-bpf-maps\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730011 kubelet[2396]: I0317 19:07:48.729998 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-cgroup\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730109 kubelet[2396]: I0317 19:07:48.730088 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-config-path\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730189 kubelet[2396]: I0317 19:07:48.730178 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/ef97debc-db00-4269-956b-9db15d4eb20a-kube-proxy\") pod \"kube-proxy-kn8bq\" (UID: \"ef97debc-db00-4269-956b-9db15d4eb20a\") " pod="kube-system/kube-proxy-kn8bq" Mar 17 19:07:48.730268 kubelet[2396]: I0317 19:07:48.730257 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ef97debc-db00-4269-956b-9db15d4eb20a-lib-modules\") pod \"kube-proxy-kn8bq\" (UID: \"ef97debc-db00-4269-956b-9db15d4eb20a\") " pod="kube-system/kube-proxy-kn8bq" Mar 17 19:07:48.730351 kubelet[2396]: I0317 19:07:48.730336 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-kernel\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730485 kubelet[2396]: I0317 19:07:48.730468 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8rzv\" (UniqueName: \"kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-kube-api-access-c8rzv\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730588 kubelet[2396]: I0317 19:07:48.730574 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-run\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730674 kubelet[2396]: I0317 19:07:48.730659 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-etc-cni-netd\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730753 kubelet[2396]: I0317 19:07:48.730741 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-lib-modules\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730839 kubelet[2396]: I0317 19:07:48.730825 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-hubble-tls\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.730934 kubelet[2396]: I0317 19:07:48.730922 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-hostproc\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.731027 kubelet[2396]: I0317 19:07:48.731013 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/626fe57b-3983-4ba1-9e43-d893ba16fe35-clustermesh-secrets\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.731107 kubelet[2396]: I0317 19:07:48.731094 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-net\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.731191 kubelet[2396]: I0317 19:07:48.731178 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cni-path\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.731272 kubelet[2396]: I0317 19:07:48.731259 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-xtables-lock\") pod \"cilium-5q82h\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " pod="kube-system/cilium-5q82h" Mar 17 19:07:48.833002 kubelet[2396]: I0317 19:07:48.832954 2396 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 17 19:07:49.011200 env[1416]: time="2025-03-17T19:07:49.010120134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kn8bq,Uid:ef97debc-db00-4269-956b-9db15d4eb20a,Namespace:kube-system,Attempt:0,}" Mar 17 19:07:49.019743 env[1416]: time="2025-03-17T19:07:49.019678548Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-5q82h,Uid:626fe57b-3983-4ba1-9e43-d893ba16fe35,Namespace:kube-system,Attempt:0,}" Mar 17 19:07:49.054771 systemd[1]: Created slice kubepods-besteffort-pod22948d77_ed73_4aeb_91e0_8b4f22a58625.slice. Mar 17 19:07:49.071680 env[1416]: time="2025-03-17T19:07:49.071596568Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:07:49.071860 env[1416]: time="2025-03-17T19:07:49.071693669Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:07:49.071860 env[1416]: time="2025-03-17T19:07:49.071722969Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:07:49.071979 env[1416]: time="2025-03-17T19:07:49.071924072Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/075d7c468075542d64a0006bf137b0702bba30ea545ddd1c9cb82ed54429b528 pid=2474 runtime=io.containerd.runc.v2 Mar 17 19:07:49.115013 systemd[1]: Started cri-containerd-075d7c468075542d64a0006bf137b0702bba30ea545ddd1c9cb82ed54429b528.scope. Mar 17 19:07:49.127744 env[1416]: time="2025-03-17T19:07:49.127666337Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:07:49.127969 env[1416]: time="2025-03-17T19:07:49.127940840Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:07:49.128087 env[1416]: time="2025-03-17T19:07:49.128057142Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:07:49.128360 env[1416]: time="2025-03-17T19:07:49.128318545Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4 pid=2505 runtime=io.containerd.runc.v2 Mar 17 19:07:49.134062 kubelet[2396]: I0317 19:07:49.133992 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/22948d77-ed73-4aeb-91e0-8b4f22a58625-cilium-config-path\") pod \"cilium-operator-5d85765b45-lz4dx\" (UID: \"22948d77-ed73-4aeb-91e0-8b4f22a58625\") " pod="kube-system/cilium-operator-5d85765b45-lz4dx" Mar 17 19:07:49.134477 kubelet[2396]: I0317 19:07:49.134079 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v8sw\" (UniqueName: \"kubernetes.io/projected/22948d77-ed73-4aeb-91e0-8b4f22a58625-kube-api-access-7v8sw\") pod \"cilium-operator-5d85765b45-lz4dx\" (UID: \"22948d77-ed73-4aeb-91e0-8b4f22a58625\") " pod="kube-system/cilium-operator-5d85765b45-lz4dx" Mar 17 19:07:49.149105 systemd[1]: Started cri-containerd-d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4.scope. Mar 17 19:07:49.192412 env[1416]: time="2025-03-17T19:07:49.192338909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kn8bq,Uid:ef97debc-db00-4269-956b-9db15d4eb20a,Namespace:kube-system,Attempt:0,} returns sandbox id \"075d7c468075542d64a0006bf137b0702bba30ea545ddd1c9cb82ed54429b528\"" Mar 17 19:07:49.199353 env[1416]: time="2025-03-17T19:07:49.199311792Z" level=info msg="CreateContainer within sandbox \"075d7c468075542d64a0006bf137b0702bba30ea545ddd1c9cb82ed54429b528\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 17 19:07:49.213844 env[1416]: time="2025-03-17T19:07:49.213313459Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-5q82h,Uid:626fe57b-3983-4ba1-9e43-d893ba16fe35,Namespace:kube-system,Attempt:0,} returns sandbox id \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\"" Mar 17 19:07:49.215119 env[1416]: time="2025-03-17T19:07:49.215075880Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 17 19:07:49.269988 env[1416]: time="2025-03-17T19:07:49.269270927Z" level=info msg="CreateContainer within sandbox \"075d7c468075542d64a0006bf137b0702bba30ea545ddd1c9cb82ed54429b528\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"01e41cfb15159bc23dd312891be9a65751bb01675815f8fbc5eedcfddacb5b82\"" Mar 17 19:07:49.272156 env[1416]: time="2025-03-17T19:07:49.270932047Z" level=info msg="StartContainer for \"01e41cfb15159bc23dd312891be9a65751bb01675815f8fbc5eedcfddacb5b82\"" Mar 17 19:07:49.289537 systemd[1]: Started cri-containerd-01e41cfb15159bc23dd312891be9a65751bb01675815f8fbc5eedcfddacb5b82.scope. Mar 17 19:07:49.328289 env[1416]: time="2025-03-17T19:07:49.328238431Z" level=info msg="StartContainer for \"01e41cfb15159bc23dd312891be9a65751bb01675815f8fbc5eedcfddacb5b82\" returns successfully" Mar 17 19:07:49.361136 env[1416]: time="2025-03-17T19:07:49.361085123Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-lz4dx,Uid:22948d77-ed73-4aeb-91e0-8b4f22a58625,Namespace:kube-system,Attempt:0,}" Mar 17 19:07:49.405425 env[1416]: time="2025-03-17T19:07:49.405325551Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:07:49.405708 env[1416]: time="2025-03-17T19:07:49.405666955Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:07:49.405848 env[1416]: time="2025-03-17T19:07:49.405821657Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:07:49.406242 env[1416]: time="2025-03-17T19:07:49.406198262Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b pid=2597 runtime=io.containerd.runc.v2 Mar 17 19:07:49.423450 systemd[1]: Started cri-containerd-5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b.scope. Mar 17 19:07:49.479180 env[1416]: time="2025-03-17T19:07:49.479130732Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-lz4dx,Uid:22948d77-ed73-4aeb-91e0-8b4f22a58625,Namespace:kube-system,Attempt:0,} returns sandbox id \"5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b\"" Mar 17 19:07:53.030973 kubelet[2396]: I0317 19:07:53.030815 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-kn8bq" podStartSLOduration=5.030791518 podStartE2EDuration="5.030791518s" podCreationTimestamp="2025-03-17 19:07:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 19:07:49.562571828 +0000 UTC m=+6.217688625" watchObservedRunningTime="2025-03-17 19:07:53.030791518 +0000 UTC m=+9.685908415" Mar 17 19:07:55.274517 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3583050120.mount: Deactivated successfully. Mar 17 19:07:57.982752 env[1416]: time="2025-03-17T19:07:57.982697287Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:57.987337 env[1416]: time="2025-03-17T19:07:57.987280533Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:57.992318 env[1416]: time="2025-03-17T19:07:57.992277283Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:07:57.993553 env[1416]: time="2025-03-17T19:07:57.993506895Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Mar 17 19:07:57.996819 env[1416]: time="2025-03-17T19:07:57.996783528Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 17 19:07:57.998032 env[1416]: time="2025-03-17T19:07:57.997974040Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 17 19:07:58.023714 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount948011103.mount: Deactivated successfully. Mar 17 19:07:58.031194 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1723494266.mount: Deactivated successfully. Mar 17 19:07:58.042604 env[1416]: time="2025-03-17T19:07:58.042546977Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\"" Mar 17 19:07:58.044103 env[1416]: time="2025-03-17T19:07:58.043119383Z" level=info msg="StartContainer for \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\"" Mar 17 19:07:58.063125 systemd[1]: Started cri-containerd-0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6.scope. Mar 17 19:07:58.095877 env[1416]: time="2025-03-17T19:07:58.095828200Z" level=info msg="StartContainer for \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\" returns successfully" Mar 17 19:07:58.126863 systemd[1]: cri-containerd-0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6.scope: Deactivated successfully. Mar 17 19:07:59.022208 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6-rootfs.mount: Deactivated successfully. Mar 17 19:08:01.867052 env[1416]: time="2025-03-17T19:08:01.866987069Z" level=info msg="shim disconnected" id=0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6 Mar 17 19:08:01.867525 env[1416]: time="2025-03-17T19:08:01.867422873Z" level=warning msg="cleaning up after shim disconnected" id=0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6 namespace=k8s.io Mar 17 19:08:01.867525 env[1416]: time="2025-03-17T19:08:01.867465474Z" level=info msg="cleaning up dead shim" Mar 17 19:08:01.875929 env[1416]: time="2025-03-17T19:08:01.875892751Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:08:01Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2810 runtime=io.containerd.runc.v2\n" Mar 17 19:08:02.582838 env[1416]: time="2025-03-17T19:08:02.582779262Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 17 19:08:02.619450 env[1416]: time="2025-03-17T19:08:02.619408993Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\"" Mar 17 19:08:02.620054 env[1416]: time="2025-03-17T19:08:02.620023698Z" level=info msg="StartContainer for \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\"" Mar 17 19:08:02.643171 systemd[1]: Started cri-containerd-99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67.scope. Mar 17 19:08:02.679718 env[1416]: time="2025-03-17T19:08:02.679664037Z" level=info msg="StartContainer for \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\" returns successfully" Mar 17 19:08:02.689775 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 17 19:08:02.690302 systemd[1]: Stopped systemd-sysctl.service. Mar 17 19:08:02.690678 systemd[1]: Stopping systemd-sysctl.service... Mar 17 19:08:02.693583 systemd[1]: Starting systemd-sysctl.service... Mar 17 19:08:02.705752 systemd[1]: Finished systemd-sysctl.service. Mar 17 19:08:02.707770 systemd[1]: cri-containerd-99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67.scope: Deactivated successfully. Mar 17 19:08:02.733600 env[1416]: time="2025-03-17T19:08:02.733543524Z" level=info msg="shim disconnected" id=99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67 Mar 17 19:08:02.733600 env[1416]: time="2025-03-17T19:08:02.733598825Z" level=warning msg="cleaning up after shim disconnected" id=99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67 namespace=k8s.io Mar 17 19:08:02.733884 env[1416]: time="2025-03-17T19:08:02.733610425Z" level=info msg="cleaning up dead shim" Mar 17 19:08:02.740939 env[1416]: time="2025-03-17T19:08:02.740901091Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:08:02Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2872 runtime=io.containerd.runc.v2\n" Mar 17 19:08:03.603750 env[1416]: time="2025-03-17T19:08:03.603705580Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 17 19:08:03.609995 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67-rootfs.mount: Deactivated successfully. Mar 17 19:08:03.656582 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1349504998.mount: Deactivated successfully. Mar 17 19:08:03.676298 env[1416]: time="2025-03-17T19:08:03.676246423Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\"" Mar 17 19:08:03.678597 env[1416]: time="2025-03-17T19:08:03.678559943Z" level=info msg="StartContainer for \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\"" Mar 17 19:08:03.712747 systemd[1]: Started cri-containerd-c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1.scope. Mar 17 19:08:03.764506 systemd[1]: cri-containerd-c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1.scope: Deactivated successfully. Mar 17 19:08:03.766136 env[1416]: time="2025-03-17T19:08:03.766091118Z" level=info msg="StartContainer for \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\" returns successfully" Mar 17 19:08:04.136602 env[1416]: time="2025-03-17T19:08:04.136548077Z" level=info msg="shim disconnected" id=c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1 Mar 17 19:08:04.136909 env[1416]: time="2025-03-17T19:08:04.136888280Z" level=warning msg="cleaning up after shim disconnected" id=c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1 namespace=k8s.io Mar 17 19:08:04.137005 env[1416]: time="2025-03-17T19:08:04.136991181Z" level=info msg="cleaning up dead shim" Mar 17 19:08:04.161333 env[1416]: time="2025-03-17T19:08:04.161282392Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:08:04Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2930 runtime=io.containerd.runc.v2\n" Mar 17 19:08:04.303834 env[1416]: time="2025-03-17T19:08:04.303773730Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:08:04.308837 env[1416]: time="2025-03-17T19:08:04.308795573Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:08:04.312142 env[1416]: time="2025-03-17T19:08:04.312104002Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 19:08:04.312711 env[1416]: time="2025-03-17T19:08:04.312679607Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Mar 17 19:08:04.316141 env[1416]: time="2025-03-17T19:08:04.316102437Z" level=info msg="CreateContainer within sandbox \"5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 17 19:08:04.344857 env[1416]: time="2025-03-17T19:08:04.344806786Z" level=info msg="CreateContainer within sandbox \"5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\"" Mar 17 19:08:04.346793 env[1416]: time="2025-03-17T19:08:04.345597993Z" level=info msg="StartContainer for \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\"" Mar 17 19:08:04.362685 systemd[1]: Started cri-containerd-afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f.scope. Mar 17 19:08:04.395659 env[1416]: time="2025-03-17T19:08:04.394896421Z" level=info msg="StartContainer for \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\" returns successfully" Mar 17 19:08:04.592745 env[1416]: time="2025-03-17T19:08:04.592699139Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 17 19:08:04.614806 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1-rootfs.mount: Deactivated successfully. Mar 17 19:08:04.621009 kubelet[2396]: I0317 19:08:04.620948 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-lz4dx" podStartSLOduration=0.787532716 podStartE2EDuration="15.620917585s" podCreationTimestamp="2025-03-17 19:07:49 +0000 UTC" firstStartedPulling="2025-03-17 19:07:49.480363247 +0000 UTC m=+6.135480044" lastFinishedPulling="2025-03-17 19:08:04.313748116 +0000 UTC m=+20.968864913" observedRunningTime="2025-03-17 19:08:04.620510581 +0000 UTC m=+21.275627478" watchObservedRunningTime="2025-03-17 19:08:04.620917585 +0000 UTC m=+21.276034382" Mar 17 19:08:04.658617 env[1416]: time="2025-03-17T19:08:04.658408210Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\"" Mar 17 19:08:04.661066 env[1416]: time="2025-03-17T19:08:04.660111525Z" level=info msg="StartContainer for \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\"" Mar 17 19:08:04.694124 systemd[1]: Started cri-containerd-e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e.scope. Mar 17 19:08:04.797539 systemd[1]: cri-containerd-e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e.scope: Deactivated successfully. Mar 17 19:08:04.801711 env[1416]: time="2025-03-17T19:08:04.801658555Z" level=info msg="StartContainer for \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\" returns successfully" Mar 17 19:08:04.916295 env[1416]: time="2025-03-17T19:08:04.916160649Z" level=info msg="shim disconnected" id=e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e Mar 17 19:08:04.916295 env[1416]: time="2025-03-17T19:08:04.916218550Z" level=warning msg="cleaning up after shim disconnected" id=e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e namespace=k8s.io Mar 17 19:08:04.916295 env[1416]: time="2025-03-17T19:08:04.916229250Z" level=info msg="cleaning up dead shim" Mar 17 19:08:04.934325 env[1416]: time="2025-03-17T19:08:04.934266007Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:08:04Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3026 runtime=io.containerd.runc.v2\n" Mar 17 19:08:05.596852 env[1416]: time="2025-03-17T19:08:05.596803863Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 17 19:08:05.613652 systemd[1]: run-containerd-runc-k8s.io-e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e-runc.k8NosA.mount: Deactivated successfully. Mar 17 19:08:05.613782 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e-rootfs.mount: Deactivated successfully. Mar 17 19:08:05.628823 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4082850699.mount: Deactivated successfully. Mar 17 19:08:05.641193 env[1416]: time="2025-03-17T19:08:05.641146841Z" level=info msg="CreateContainer within sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\"" Mar 17 19:08:05.642000 env[1416]: time="2025-03-17T19:08:05.641969748Z" level=info msg="StartContainer for \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\"" Mar 17 19:08:05.668326 systemd[1]: Started cri-containerd-fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059.scope. Mar 17 19:08:05.711419 env[1416]: time="2025-03-17T19:08:05.709671624Z" level=info msg="StartContainer for \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\" returns successfully" Mar 17 19:08:05.822839 kubelet[2396]: I0317 19:08:05.822658 2396 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Mar 17 19:08:05.868141 systemd[1]: Created slice kubepods-burstable-pod384b4865_e704_48cc_9b39_0a1e31dff77f.slice. Mar 17 19:08:05.878649 systemd[1]: Created slice kubepods-burstable-pod793047ab_6f25_4c89_84cc_3c3f86f40add.slice. Mar 17 19:08:05.954763 kubelet[2396]: I0317 19:08:05.954547 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4gqr\" (UniqueName: \"kubernetes.io/projected/384b4865-e704-48cc-9b39-0a1e31dff77f-kube-api-access-h4gqr\") pod \"coredns-6f6b679f8f-9cdz5\" (UID: \"384b4865-e704-48cc-9b39-0a1e31dff77f\") " pod="kube-system/coredns-6f6b679f8f-9cdz5" Mar 17 19:08:05.954763 kubelet[2396]: I0317 19:08:05.954627 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/793047ab-6f25-4c89-84cc-3c3f86f40add-config-volume\") pod \"coredns-6f6b679f8f-9t7qj\" (UID: \"793047ab-6f25-4c89-84cc-3c3f86f40add\") " pod="kube-system/coredns-6f6b679f8f-9t7qj" Mar 17 19:08:05.954763 kubelet[2396]: I0317 19:08:05.954699 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpdnw\" (UniqueName: \"kubernetes.io/projected/793047ab-6f25-4c89-84cc-3c3f86f40add-kube-api-access-xpdnw\") pod \"coredns-6f6b679f8f-9t7qj\" (UID: \"793047ab-6f25-4c89-84cc-3c3f86f40add\") " pod="kube-system/coredns-6f6b679f8f-9t7qj" Mar 17 19:08:05.954763 kubelet[2396]: I0317 19:08:05.954741 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/384b4865-e704-48cc-9b39-0a1e31dff77f-config-volume\") pod \"coredns-6f6b679f8f-9cdz5\" (UID: \"384b4865-e704-48cc-9b39-0a1e31dff77f\") " pod="kube-system/coredns-6f6b679f8f-9cdz5" Mar 17 19:08:06.173791 env[1416]: time="2025-03-17T19:08:06.173736951Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-9cdz5,Uid:384b4865-e704-48cc-9b39-0a1e31dff77f,Namespace:kube-system,Attempt:0,}" Mar 17 19:08:06.183940 env[1416]: time="2025-03-17T19:08:06.183902036Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-9t7qj,Uid:793047ab-6f25-4c89-84cc-3c3f86f40add,Namespace:kube-system,Attempt:0,}" Mar 17 19:08:06.622802 kubelet[2396]: I0317 19:08:06.622628 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-5q82h" podStartSLOduration=9.842114667 podStartE2EDuration="18.622602403s" podCreationTimestamp="2025-03-17 19:07:48 +0000 UTC" firstStartedPulling="2025-03-17 19:07:49.214592675 +0000 UTC m=+5.869709472" lastFinishedPulling="2025-03-17 19:07:57.995080411 +0000 UTC m=+14.650197208" observedRunningTime="2025-03-17 19:08:06.620532185 +0000 UTC m=+23.275649082" watchObservedRunningTime="2025-03-17 19:08:06.622602403 +0000 UTC m=+23.277719200" Mar 17 19:08:08.145333 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Mar 17 19:08:08.145481 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Mar 17 19:08:08.140455 systemd-networkd[1571]: cilium_host: Link UP Mar 17 19:08:08.140626 systemd-networkd[1571]: cilium_net: Link UP Mar 17 19:08:08.141851 systemd-networkd[1571]: cilium_net: Gained carrier Mar 17 19:08:08.146739 systemd-networkd[1571]: cilium_host: Gained carrier Mar 17 19:08:08.303670 systemd-networkd[1571]: cilium_vxlan: Link UP Mar 17 19:08:08.303680 systemd-networkd[1571]: cilium_vxlan: Gained carrier Mar 17 19:08:08.598411 kernel: NET: Registered PF_ALG protocol family Mar 17 19:08:08.698574 systemd-networkd[1571]: cilium_host: Gained IPv6LL Mar 17 19:08:08.954561 systemd-networkd[1571]: cilium_net: Gained IPv6LL Mar 17 19:08:09.316980 systemd-networkd[1571]: lxc_health: Link UP Mar 17 19:08:09.334256 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Mar 17 19:08:09.333662 systemd-networkd[1571]: lxc_health: Gained carrier Mar 17 19:08:09.722588 systemd-networkd[1571]: cilium_vxlan: Gained IPv6LL Mar 17 19:08:09.744543 kernel: eth0: renamed from tmpc7876 Mar 17 19:08:09.749961 systemd-networkd[1571]: lxc6db295f518e6: Link UP Mar 17 19:08:09.755599 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc6db295f518e6: link becomes ready Mar 17 19:08:09.756021 systemd-networkd[1571]: lxc6db295f518e6: Gained carrier Mar 17 19:08:09.758629 systemd-networkd[1571]: lxc0e3caf2b9976: Link UP Mar 17 19:08:09.767483 kernel: eth0: renamed from tmpdad57 Mar 17 19:08:09.778965 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc0e3caf2b9976: link becomes ready Mar 17 19:08:09.779233 systemd-networkd[1571]: lxc0e3caf2b9976: Gained carrier Mar 17 19:08:10.939552 systemd-networkd[1571]: lxc6db295f518e6: Gained IPv6LL Mar 17 19:08:11.322526 systemd-networkd[1571]: lxc_health: Gained IPv6LL Mar 17 19:08:11.706521 systemd-networkd[1571]: lxc0e3caf2b9976: Gained IPv6LL Mar 17 19:08:13.449405 env[1416]: time="2025-03-17T19:08:13.448715104Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:08:13.449405 env[1416]: time="2025-03-17T19:08:13.448805104Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:08:13.449405 env[1416]: time="2025-03-17T19:08:13.448834804Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:08:13.449405 env[1416]: time="2025-03-17T19:08:13.448987006Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/dad57f0795e31c34d351ad2ebd306aa97504d7414b53642700b5b82dc18791fa pid=3583 runtime=io.containerd.runc.v2 Mar 17 19:08:13.477947 systemd[1]: Started cri-containerd-dad57f0795e31c34d351ad2ebd306aa97504d7414b53642700b5b82dc18791fa.scope. Mar 17 19:08:13.502641 env[1416]: time="2025-03-17T19:08:13.502569100Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:08:13.504089 env[1416]: time="2025-03-17T19:08:13.503979910Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:08:13.504329 env[1416]: time="2025-03-17T19:08:13.504298713Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:08:13.505439 env[1416]: time="2025-03-17T19:08:13.504857017Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c78768dc3137843e29cceeefb23a6420c45b3b3dd724266ca81d5b779db81f84 pid=3610 runtime=io.containerd.runc.v2 Mar 17 19:08:13.538346 systemd[1]: Started cri-containerd-c78768dc3137843e29cceeefb23a6420c45b3b3dd724266ca81d5b779db81f84.scope. Mar 17 19:08:13.610461 env[1416]: time="2025-03-17T19:08:13.610401594Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-9t7qj,Uid:793047ab-6f25-4c89-84cc-3c3f86f40add,Namespace:kube-system,Attempt:0,} returns sandbox id \"dad57f0795e31c34d351ad2ebd306aa97504d7414b53642700b5b82dc18791fa\"" Mar 17 19:08:13.614396 env[1416]: time="2025-03-17T19:08:13.614344023Z" level=info msg="CreateContainer within sandbox \"dad57f0795e31c34d351ad2ebd306aa97504d7414b53642700b5b82dc18791fa\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 17 19:08:13.655862 env[1416]: time="2025-03-17T19:08:13.655816328Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-9cdz5,Uid:384b4865-e704-48cc-9b39-0a1e31dff77f,Namespace:kube-system,Attempt:0,} returns sandbox id \"c78768dc3137843e29cceeefb23a6420c45b3b3dd724266ca81d5b779db81f84\"" Mar 17 19:08:13.664454 env[1416]: time="2025-03-17T19:08:13.662095375Z" level=info msg="CreateContainer within sandbox \"c78768dc3137843e29cceeefb23a6420c45b3b3dd724266ca81d5b779db81f84\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 17 19:08:13.692535 env[1416]: time="2025-03-17T19:08:13.692333097Z" level=info msg="CreateContainer within sandbox \"dad57f0795e31c34d351ad2ebd306aa97504d7414b53642700b5b82dc18791fa\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"758decc6da089aadebd8ab007644c9f4d1cd3695db543e828e434b049fa9b744\"" Mar 17 19:08:13.695390 env[1416]: time="2025-03-17T19:08:13.695336719Z" level=info msg="StartContainer for \"758decc6da089aadebd8ab007644c9f4d1cd3695db543e828e434b049fa9b744\"" Mar 17 19:08:13.713763 env[1416]: time="2025-03-17T19:08:13.713005250Z" level=info msg="CreateContainer within sandbox \"c78768dc3137843e29cceeefb23a6420c45b3b3dd724266ca81d5b779db81f84\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"69fdb757a094f68b63848fccc952ef7f8a5139b5bff178c18c81f0e26e7821e0\"" Mar 17 19:08:13.715644 env[1416]: time="2025-03-17T19:08:13.715604569Z" level=info msg="StartContainer for \"69fdb757a094f68b63848fccc952ef7f8a5139b5bff178c18c81f0e26e7821e0\"" Mar 17 19:08:13.723055 systemd[1]: Started cri-containerd-758decc6da089aadebd8ab007644c9f4d1cd3695db543e828e434b049fa9b744.scope. Mar 17 19:08:13.749052 systemd[1]: Started cri-containerd-69fdb757a094f68b63848fccc952ef7f8a5139b5bff178c18c81f0e26e7821e0.scope. Mar 17 19:08:13.776592 env[1416]: time="2025-03-17T19:08:13.776536717Z" level=info msg="StartContainer for \"758decc6da089aadebd8ab007644c9f4d1cd3695db543e828e434b049fa9b744\" returns successfully" Mar 17 19:08:13.809886 env[1416]: time="2025-03-17T19:08:13.809813662Z" level=info msg="StartContainer for \"69fdb757a094f68b63848fccc952ef7f8a5139b5bff178c18c81f0e26e7821e0\" returns successfully" Mar 17 19:08:14.458681 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3497423529.mount: Deactivated successfully. Mar 17 19:08:14.656232 kubelet[2396]: I0317 19:08:14.656144 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-9t7qj" podStartSLOduration=25.656117112 podStartE2EDuration="25.656117112s" podCreationTimestamp="2025-03-17 19:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 19:08:14.637099574 +0000 UTC m=+31.292216471" watchObservedRunningTime="2025-03-17 19:08:14.656117112 +0000 UTC m=+31.311234009" Mar 17 19:08:14.681970 kubelet[2396]: I0317 19:08:14.681903 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-9cdz5" podStartSLOduration=25.681880899 podStartE2EDuration="25.681880899s" podCreationTimestamp="2025-03-17 19:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 19:08:14.659031533 +0000 UTC m=+31.314148430" watchObservedRunningTime="2025-03-17 19:08:14.681880899 +0000 UTC m=+31.336997796" Mar 17 19:10:09.554611 systemd[1]: Started sshd@5-10.200.8.15:22-10.200.16.10:58066.service. Mar 17 19:10:10.178268 sshd[3760]: Accepted publickey for core from 10.200.16.10 port 58066 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:10.179991 sshd[3760]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:10.185063 systemd[1]: Started session-8.scope. Mar 17 19:10:10.185582 systemd-logind[1401]: New session 8 of user core. Mar 17 19:10:10.707098 sshd[3760]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:10.710723 systemd[1]: sshd@5-10.200.8.15:22-10.200.16.10:58066.service: Deactivated successfully. Mar 17 19:10:10.711877 systemd[1]: session-8.scope: Deactivated successfully. Mar 17 19:10:10.712806 systemd-logind[1401]: Session 8 logged out. Waiting for processes to exit. Mar 17 19:10:10.713715 systemd-logind[1401]: Removed session 8. Mar 17 19:10:15.815911 systemd[1]: Started sshd@6-10.200.8.15:22-10.200.16.10:58074.service. Mar 17 19:10:16.440430 sshd[3772]: Accepted publickey for core from 10.200.16.10 port 58074 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:16.442154 sshd[3772]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:16.447390 systemd[1]: Started session-9.scope. Mar 17 19:10:16.447996 systemd-logind[1401]: New session 9 of user core. Mar 17 19:10:16.939942 sshd[3772]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:16.942722 systemd[1]: sshd@6-10.200.8.15:22-10.200.16.10:58074.service: Deactivated successfully. Mar 17 19:10:16.943655 systemd[1]: session-9.scope: Deactivated successfully. Mar 17 19:10:16.944407 systemd-logind[1401]: Session 9 logged out. Waiting for processes to exit. Mar 17 19:10:16.945195 systemd-logind[1401]: Removed session 9. Mar 17 19:10:22.046800 systemd[1]: Started sshd@7-10.200.8.15:22-10.200.16.10:45036.service. Mar 17 19:10:22.670992 sshd[3788]: Accepted publickey for core from 10.200.16.10 port 45036 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:22.672397 sshd[3788]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:22.677702 systemd[1]: Started session-10.scope. Mar 17 19:10:22.678168 systemd-logind[1401]: New session 10 of user core. Mar 17 19:10:23.184348 sshd[3788]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:23.187663 systemd[1]: sshd@7-10.200.8.15:22-10.200.16.10:45036.service: Deactivated successfully. Mar 17 19:10:23.188710 systemd[1]: session-10.scope: Deactivated successfully. Mar 17 19:10:23.189437 systemd-logind[1401]: Session 10 logged out. Waiting for processes to exit. Mar 17 19:10:23.190229 systemd-logind[1401]: Removed session 10. Mar 17 19:10:28.371573 systemd[1]: Started sshd@8-10.200.8.15:22-10.200.16.10:45042.service. Mar 17 19:10:29.047960 sshd[3800]: Accepted publickey for core from 10.200.16.10 port 45042 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:29.049414 sshd[3800]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:29.054479 systemd-logind[1401]: New session 11 of user core. Mar 17 19:10:29.055153 systemd[1]: Started session-11.scope. Mar 17 19:10:29.579639 sshd[3800]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:29.582884 systemd[1]: sshd@8-10.200.8.15:22-10.200.16.10:45042.service: Deactivated successfully. Mar 17 19:10:29.584048 systemd[1]: session-11.scope: Deactivated successfully. Mar 17 19:10:29.584886 systemd-logind[1401]: Session 11 logged out. Waiting for processes to exit. Mar 17 19:10:29.585840 systemd-logind[1401]: Removed session 11. Mar 17 19:10:34.702876 systemd[1]: Started sshd@9-10.200.8.15:22-10.200.16.10:52420.service. Mar 17 19:10:35.337041 sshd[3812]: Accepted publickey for core from 10.200.16.10 port 52420 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:35.338740 sshd[3812]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:35.343628 systemd[1]: Started session-12.scope. Mar 17 19:10:35.344398 systemd-logind[1401]: New session 12 of user core. Mar 17 19:10:35.849183 sshd[3812]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:35.851938 systemd[1]: sshd@9-10.200.8.15:22-10.200.16.10:52420.service: Deactivated successfully. Mar 17 19:10:35.853078 systemd-logind[1401]: Session 12 logged out. Waiting for processes to exit. Mar 17 19:10:35.853165 systemd[1]: session-12.scope: Deactivated successfully. Mar 17 19:10:35.854415 systemd-logind[1401]: Removed session 12. Mar 17 19:10:35.973404 systemd[1]: Started sshd@10-10.200.8.15:22-10.200.16.10:52424.service. Mar 17 19:10:36.595732 sshd[3824]: Accepted publickey for core from 10.200.16.10 port 52424 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:36.597153 sshd[3824]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:36.602044 systemd[1]: Started session-13.scope. Mar 17 19:10:36.602668 systemd-logind[1401]: New session 13 of user core. Mar 17 19:10:37.130068 sshd[3824]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:37.132755 systemd[1]: sshd@10-10.200.8.15:22-10.200.16.10:52424.service: Deactivated successfully. Mar 17 19:10:37.133602 systemd[1]: session-13.scope: Deactivated successfully. Mar 17 19:10:37.134326 systemd-logind[1401]: Session 13 logged out. Waiting for processes to exit. Mar 17 19:10:37.135278 systemd-logind[1401]: Removed session 13. Mar 17 19:10:37.240523 systemd[1]: Started sshd@11-10.200.8.15:22-10.200.16.10:52426.service. Mar 17 19:10:37.872440 sshd[3834]: Accepted publickey for core from 10.200.16.10 port 52426 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:37.873991 sshd[3834]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:37.879344 systemd-logind[1401]: New session 14 of user core. Mar 17 19:10:37.879915 systemd[1]: Started session-14.scope. Mar 17 19:10:38.377063 sshd[3834]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:38.380398 systemd[1]: sshd@11-10.200.8.15:22-10.200.16.10:52426.service: Deactivated successfully. Mar 17 19:10:38.381530 systemd[1]: session-14.scope: Deactivated successfully. Mar 17 19:10:38.382230 systemd-logind[1401]: Session 14 logged out. Waiting for processes to exit. Mar 17 19:10:38.383132 systemd-logind[1401]: Removed session 14. Mar 17 19:10:43.482716 systemd[1]: Started sshd@12-10.200.8.15:22-10.200.16.10:47100.service. Mar 17 19:10:44.114024 sshd[3846]: Accepted publickey for core from 10.200.16.10 port 47100 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:44.115896 sshd[3846]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:44.121085 systemd[1]: Started session-15.scope. Mar 17 19:10:44.121586 systemd-logind[1401]: New session 15 of user core. Mar 17 19:10:44.612092 sshd[3846]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:44.615366 systemd[1]: sshd@12-10.200.8.15:22-10.200.16.10:47100.service: Deactivated successfully. Mar 17 19:10:44.616271 systemd[1]: session-15.scope: Deactivated successfully. Mar 17 19:10:44.617061 systemd-logind[1401]: Session 15 logged out. Waiting for processes to exit. Mar 17 19:10:44.617900 systemd-logind[1401]: Removed session 15. Mar 17 19:10:44.717974 systemd[1]: Started sshd@13-10.200.8.15:22-10.200.16.10:47104.service. Mar 17 19:10:45.342259 sshd[3860]: Accepted publickey for core from 10.200.16.10 port 47104 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:45.343952 sshd[3860]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:45.349135 systemd[1]: Started session-16.scope. Mar 17 19:10:45.349599 systemd-logind[1401]: New session 16 of user core. Mar 17 19:10:45.917031 sshd[3860]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:45.920086 systemd[1]: sshd@13-10.200.8.15:22-10.200.16.10:47104.service: Deactivated successfully. Mar 17 19:10:45.921286 systemd[1]: session-16.scope: Deactivated successfully. Mar 17 19:10:45.922202 systemd-logind[1401]: Session 16 logged out. Waiting for processes to exit. Mar 17 19:10:45.923218 systemd-logind[1401]: Removed session 16. Mar 17 19:10:46.025529 systemd[1]: Started sshd@14-10.200.8.15:22-10.200.16.10:47112.service. Mar 17 19:10:46.655562 sshd[3869]: Accepted publickey for core from 10.200.16.10 port 47112 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:46.657213 sshd[3869]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:46.662906 systemd-logind[1401]: New session 17 of user core. Mar 17 19:10:46.663546 systemd[1]: Started session-17.scope. Mar 17 19:10:48.685854 sshd[3869]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:48.688967 systemd[1]: sshd@14-10.200.8.15:22-10.200.16.10:47112.service: Deactivated successfully. Mar 17 19:10:48.689857 systemd[1]: session-17.scope: Deactivated successfully. Mar 17 19:10:48.690634 systemd-logind[1401]: Session 17 logged out. Waiting for processes to exit. Mar 17 19:10:48.691447 systemd-logind[1401]: Removed session 17. Mar 17 19:10:48.794811 systemd[1]: Started sshd@15-10.200.8.15:22-10.200.16.10:39434.service. Mar 17 19:10:49.419220 sshd[3886]: Accepted publickey for core from 10.200.16.10 port 39434 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:49.420734 sshd[3886]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:49.425724 systemd-logind[1401]: New session 18 of user core. Mar 17 19:10:49.426220 systemd[1]: Started session-18.scope. Mar 17 19:10:50.032052 sshd[3886]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:50.035502 systemd[1]: sshd@15-10.200.8.15:22-10.200.16.10:39434.service: Deactivated successfully. Mar 17 19:10:50.036641 systemd[1]: session-18.scope: Deactivated successfully. Mar 17 19:10:50.037540 systemd-logind[1401]: Session 18 logged out. Waiting for processes to exit. Mar 17 19:10:50.038656 systemd-logind[1401]: Removed session 18. Mar 17 19:10:50.137297 systemd[1]: Started sshd@16-10.200.8.15:22-10.200.16.10:39446.service. Mar 17 19:10:50.761931 sshd[3898]: Accepted publickey for core from 10.200.16.10 port 39446 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:50.763627 sshd[3898]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:50.768658 systemd[1]: Started session-19.scope. Mar 17 19:10:50.769443 systemd-logind[1401]: New session 19 of user core. Mar 17 19:10:51.270932 sshd[3898]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:51.274800 systemd[1]: sshd@16-10.200.8.15:22-10.200.16.10:39446.service: Deactivated successfully. Mar 17 19:10:51.275733 systemd[1]: session-19.scope: Deactivated successfully. Mar 17 19:10:51.276435 systemd-logind[1401]: Session 19 logged out. Waiting for processes to exit. Mar 17 19:10:51.277655 systemd-logind[1401]: Removed session 19. Mar 17 19:10:56.377134 systemd[1]: Started sshd@17-10.200.8.15:22-10.200.16.10:39456.service. Mar 17 19:10:57.002433 sshd[3913]: Accepted publickey for core from 10.200.16.10 port 39456 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:10:57.004184 sshd[3913]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:10:57.009555 systemd-logind[1401]: New session 20 of user core. Mar 17 19:10:57.010069 systemd[1]: Started session-20.scope. Mar 17 19:10:57.502059 sshd[3913]: pam_unix(sshd:session): session closed for user core Mar 17 19:10:57.504900 systemd[1]: sshd@17-10.200.8.15:22-10.200.16.10:39456.service: Deactivated successfully. Mar 17 19:10:57.505861 systemd[1]: session-20.scope: Deactivated successfully. Mar 17 19:10:57.506543 systemd-logind[1401]: Session 20 logged out. Waiting for processes to exit. Mar 17 19:10:57.507325 systemd-logind[1401]: Removed session 20. Mar 17 19:11:02.608790 systemd[1]: Started sshd@18-10.200.8.15:22-10.200.16.10:49360.service. Mar 17 19:11:03.235735 sshd[3925]: Accepted publickey for core from 10.200.16.10 port 49360 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:11:03.237092 sshd[3925]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:11:03.241975 systemd[1]: Started session-21.scope. Mar 17 19:11:03.242346 systemd-logind[1401]: New session 21 of user core. Mar 17 19:11:03.734889 sshd[3925]: pam_unix(sshd:session): session closed for user core Mar 17 19:11:03.738019 systemd[1]: sshd@18-10.200.8.15:22-10.200.16.10:49360.service: Deactivated successfully. Mar 17 19:11:03.738974 systemd[1]: session-21.scope: Deactivated successfully. Mar 17 19:11:03.739649 systemd-logind[1401]: Session 21 logged out. Waiting for processes to exit. Mar 17 19:11:03.740468 systemd-logind[1401]: Removed session 21. Mar 17 19:11:08.841863 systemd[1]: Started sshd@19-10.200.8.15:22-10.200.16.10:48260.service. Mar 17 19:11:09.468039 sshd[3939]: Accepted publickey for core from 10.200.16.10 port 48260 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:11:09.469896 sshd[3939]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:11:09.477299 systemd[1]: Started session-22.scope. Mar 17 19:11:09.479844 systemd-logind[1401]: New session 22 of user core. Mar 17 19:11:09.965187 sshd[3939]: pam_unix(sshd:session): session closed for user core Mar 17 19:11:09.968546 systemd[1]: sshd@19-10.200.8.15:22-10.200.16.10:48260.service: Deactivated successfully. Mar 17 19:11:09.969665 systemd[1]: session-22.scope: Deactivated successfully. Mar 17 19:11:09.970482 systemd-logind[1401]: Session 22 logged out. Waiting for processes to exit. Mar 17 19:11:09.971291 systemd-logind[1401]: Removed session 22. Mar 17 19:11:10.072151 systemd[1]: Started sshd@20-10.200.8.15:22-10.200.16.10:48270.service. Mar 17 19:11:10.699032 sshd[3954]: Accepted publickey for core from 10.200.16.10 port 48270 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:11:10.700650 sshd[3954]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:11:10.706100 systemd[1]: Started session-23.scope. Mar 17 19:11:10.706752 systemd-logind[1401]: New session 23 of user core. Mar 17 19:11:12.663045 systemd[1]: run-containerd-runc-k8s.io-fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059-runc.qAodvC.mount: Deactivated successfully. Mar 17 19:11:12.682885 env[1416]: time="2025-03-17T19:11:12.682033985Z" level=info msg="StopContainer for \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\" with timeout 30 (s)" Mar 17 19:11:12.683529 env[1416]: time="2025-03-17T19:11:12.683493402Z" level=info msg="Stop container \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\" with signal terminated" Mar 17 19:11:12.691412 env[1416]: time="2025-03-17T19:11:12.691331194Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 19:11:12.698495 systemd[1]: cri-containerd-afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f.scope: Deactivated successfully. Mar 17 19:11:12.704778 env[1416]: time="2025-03-17T19:11:12.704737351Z" level=info msg="StopContainer for \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\" with timeout 2 (s)" Mar 17 19:11:12.705199 env[1416]: time="2025-03-17T19:11:12.705162456Z" level=info msg="Stop container \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\" with signal terminated" Mar 17 19:11:12.718271 systemd-networkd[1571]: lxc_health: Link DOWN Mar 17 19:11:12.718281 systemd-networkd[1571]: lxc_health: Lost carrier Mar 17 19:11:12.726565 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f-rootfs.mount: Deactivated successfully. Mar 17 19:11:12.740719 systemd[1]: cri-containerd-fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059.scope: Deactivated successfully. Mar 17 19:11:12.740959 systemd[1]: cri-containerd-fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059.scope: Consumed 7.153s CPU time. Mar 17 19:11:12.762035 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059-rootfs.mount: Deactivated successfully. Mar 17 19:11:12.779001 env[1416]: time="2025-03-17T19:11:12.778946822Z" level=info msg="shim disconnected" id=fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059 Mar 17 19:11:12.779001 env[1416]: time="2025-03-17T19:11:12.779002323Z" level=warning msg="cleaning up after shim disconnected" id=fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059 namespace=k8s.io Mar 17 19:11:12.779358 env[1416]: time="2025-03-17T19:11:12.779014923Z" level=info msg="cleaning up dead shim" Mar 17 19:11:12.779708 env[1416]: time="2025-03-17T19:11:12.779664331Z" level=info msg="shim disconnected" id=afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f Mar 17 19:11:12.779708 env[1416]: time="2025-03-17T19:11:12.779708631Z" level=warning msg="cleaning up after shim disconnected" id=afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f namespace=k8s.io Mar 17 19:11:12.779929 env[1416]: time="2025-03-17T19:11:12.779720731Z" level=info msg="cleaning up dead shim" Mar 17 19:11:12.789808 env[1416]: time="2025-03-17T19:11:12.789766949Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4021 runtime=io.containerd.runc.v2\n" Mar 17 19:11:12.790925 env[1416]: time="2025-03-17T19:11:12.790885562Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4020 runtime=io.containerd.runc.v2\n" Mar 17 19:11:12.794882 env[1416]: time="2025-03-17T19:11:12.794851909Z" level=info msg="StopContainer for \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\" returns successfully" Mar 17 19:11:12.795566 env[1416]: time="2025-03-17T19:11:12.795539117Z" level=info msg="StopPodSandbox for \"5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b\"" Mar 17 19:11:12.795663 env[1416]: time="2025-03-17T19:11:12.795609118Z" level=info msg="Container to stop \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 19:11:12.798207 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b-shm.mount: Deactivated successfully. Mar 17 19:11:12.809359 env[1416]: time="2025-03-17T19:11:12.808573670Z" level=info msg="StopContainer for \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\" returns successfully" Mar 17 19:11:12.809359 env[1416]: time="2025-03-17T19:11:12.809187577Z" level=info msg="StopPodSandbox for \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\"" Mar 17 19:11:12.809359 env[1416]: time="2025-03-17T19:11:12.809254878Z" level=info msg="Container to stop \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 19:11:12.809359 env[1416]: time="2025-03-17T19:11:12.809274578Z" level=info msg="Container to stop \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 19:11:12.809359 env[1416]: time="2025-03-17T19:11:12.809290078Z" level=info msg="Container to stop \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 19:11:12.809359 env[1416]: time="2025-03-17T19:11:12.809306879Z" level=info msg="Container to stop \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 19:11:12.809359 env[1416]: time="2025-03-17T19:11:12.809321079Z" level=info msg="Container to stop \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 19:11:12.808824 systemd[1]: cri-containerd-5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b.scope: Deactivated successfully. Mar 17 19:11:12.823075 systemd[1]: cri-containerd-d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4.scope: Deactivated successfully. Mar 17 19:11:12.840628 env[1416]: time="2025-03-17T19:11:12.840520645Z" level=info msg="shim disconnected" id=5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b Mar 17 19:11:12.840628 env[1416]: time="2025-03-17T19:11:12.840573646Z" level=warning msg="cleaning up after shim disconnected" id=5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b namespace=k8s.io Mar 17 19:11:12.840628 env[1416]: time="2025-03-17T19:11:12.840585446Z" level=info msg="cleaning up dead shim" Mar 17 19:11:12.853256 env[1416]: time="2025-03-17T19:11:12.853214694Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4082 runtime=io.containerd.runc.v2\n" Mar 17 19:11:12.853629 env[1416]: time="2025-03-17T19:11:12.853595898Z" level=info msg="TearDown network for sandbox \"5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b\" successfully" Mar 17 19:11:12.853719 env[1416]: time="2025-03-17T19:11:12.853624499Z" level=info msg="StopPodSandbox for \"5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b\" returns successfully" Mar 17 19:11:12.855585 env[1416]: time="2025-03-17T19:11:12.855353219Z" level=info msg="shim disconnected" id=d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4 Mar 17 19:11:12.855585 env[1416]: time="2025-03-17T19:11:12.855420920Z" level=warning msg="cleaning up after shim disconnected" id=d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4 namespace=k8s.io Mar 17 19:11:12.855585 env[1416]: time="2025-03-17T19:11:12.855434620Z" level=info msg="cleaning up dead shim" Mar 17 19:11:12.868268 env[1416]: time="2025-03-17T19:11:12.868224670Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4097 runtime=io.containerd.runc.v2\n" Mar 17 19:11:12.868720 env[1416]: time="2025-03-17T19:11:12.868684975Z" level=info msg="TearDown network for sandbox \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" successfully" Mar 17 19:11:12.868720 env[1416]: time="2025-03-17T19:11:12.868717476Z" level=info msg="StopPodSandbox for \"d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4\" returns successfully" Mar 17 19:11:12.986970 kubelet[2396]: I0317 19:11:12.986837 2396 scope.go:117] "RemoveContainer" containerID="afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f" Mar 17 19:11:12.991432 env[1416]: time="2025-03-17T19:11:12.991336115Z" level=info msg="RemoveContainer for \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\"" Mar 17 19:11:13.003012 env[1416]: time="2025-03-17T19:11:13.002965751Z" level=info msg="RemoveContainer for \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\" returns successfully" Mar 17 19:11:13.003320 kubelet[2396]: I0317 19:11:13.003294 2396 scope.go:117] "RemoveContainer" containerID="afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f" Mar 17 19:11:13.003675 env[1416]: time="2025-03-17T19:11:13.003599959Z" level=error msg="ContainerStatus for \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\": not found" Mar 17 19:11:13.003809 kubelet[2396]: E0317 19:11:13.003781 2396 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\": not found" containerID="afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f" Mar 17 19:11:13.003931 kubelet[2396]: I0317 19:11:13.003820 2396 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f"} err="failed to get container status \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\": rpc error: code = NotFound desc = an error occurred when try to find container \"afe3abe2f6131b4dbfd2fd9bbab5bf4485bbb4d6b3ce63c3900ec24419e1e47f\": not found" Mar 17 19:11:13.004013 kubelet[2396]: I0317 19:11:13.003936 2396 scope.go:117] "RemoveContainer" containerID="fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059" Mar 17 19:11:13.004924 env[1416]: time="2025-03-17T19:11:13.004894074Z" level=info msg="RemoveContainer for \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\"" Mar 17 19:11:13.011386 env[1416]: time="2025-03-17T19:11:13.011351949Z" level=info msg="RemoveContainer for \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\" returns successfully" Mar 17 19:11:13.011556 kubelet[2396]: I0317 19:11:13.011539 2396 scope.go:117] "RemoveContainer" containerID="e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e" Mar 17 19:11:13.012652 env[1416]: time="2025-03-17T19:11:13.012624164Z" level=info msg="RemoveContainer for \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\"" Mar 17 19:11:13.019901 env[1416]: time="2025-03-17T19:11:13.019864048Z" level=info msg="RemoveContainer for \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\" returns successfully" Mar 17 19:11:13.020136 kubelet[2396]: I0317 19:11:13.020096 2396 scope.go:117] "RemoveContainer" containerID="c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1" Mar 17 19:11:13.021132 env[1416]: time="2025-03-17T19:11:13.021107163Z" level=info msg="RemoveContainer for \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\"" Mar 17 19:11:13.027501 env[1416]: time="2025-03-17T19:11:13.027467737Z" level=info msg="RemoveContainer for \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\" returns successfully" Mar 17 19:11:13.027664 kubelet[2396]: I0317 19:11:13.027645 2396 scope.go:117] "RemoveContainer" containerID="99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67" Mar 17 19:11:13.028674 env[1416]: time="2025-03-17T19:11:13.028647851Z" level=info msg="RemoveContainer for \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\"" Mar 17 19:11:13.034887 kubelet[2396]: I0317 19:11:13.034858 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8rzv\" (UniqueName: \"kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-kube-api-access-c8rzv\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.034995 kubelet[2396]: I0317 19:11:13.034902 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-hubble-tls\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.034995 kubelet[2396]: I0317 19:11:13.034925 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-run\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.034995 kubelet[2396]: I0317 19:11:13.034944 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-hostproc\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.034995 kubelet[2396]: I0317 19:11:13.034968 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/626fe57b-3983-4ba1-9e43-d893ba16fe35-clustermesh-secrets\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.034995 kubelet[2396]: I0317 19:11:13.034991 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v8sw\" (UniqueName: \"kubernetes.io/projected/22948d77-ed73-4aeb-91e0-8b4f22a58625-kube-api-access-7v8sw\") pod \"22948d77-ed73-4aeb-91e0-8b4f22a58625\" (UID: \"22948d77-ed73-4aeb-91e0-8b4f22a58625\") " Mar 17 19:11:13.035218 kubelet[2396]: I0317 19:11:13.035019 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-etc-cni-netd\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035218 kubelet[2396]: I0317 19:11:13.035041 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-kernel\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035218 kubelet[2396]: I0317 19:11:13.035061 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-net\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035218 kubelet[2396]: I0317 19:11:13.035086 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/22948d77-ed73-4aeb-91e0-8b4f22a58625-cilium-config-path\") pod \"22948d77-ed73-4aeb-91e0-8b4f22a58625\" (UID: \"22948d77-ed73-4aeb-91e0-8b4f22a58625\") " Mar 17 19:11:13.035218 kubelet[2396]: I0317 19:11:13.035107 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-cgroup\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035218 kubelet[2396]: I0317 19:11:13.035126 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-lib-modules\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035510 kubelet[2396]: I0317 19:11:13.035151 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cni-path\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035510 kubelet[2396]: I0317 19:11:13.035172 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-xtables-lock\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035510 kubelet[2396]: I0317 19:11:13.035191 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-bpf-maps\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035510 kubelet[2396]: I0317 19:11:13.035214 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-config-path\") pod \"626fe57b-3983-4ba1-9e43-d893ba16fe35\" (UID: \"626fe57b-3983-4ba1-9e43-d893ba16fe35\") " Mar 17 19:11:13.035758 kubelet[2396]: I0317 19:11:13.035734 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.037882 kubelet[2396]: I0317 19:11:13.037848 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 19:11:13.037984 kubelet[2396]: I0317 19:11:13.037916 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.038948 kubelet[2396]: I0317 19:11:13.038923 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.039103 kubelet[2396]: I0317 19:11:13.039084 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-hostproc" (OuterVolumeSpecName: "hostproc") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.043035 env[1416]: time="2025-03-17T19:11:13.042993918Z" level=info msg="RemoveContainer for \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\" returns successfully" Mar 17 19:11:13.043195 kubelet[2396]: I0317 19:11:13.043168 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.043275 kubelet[2396]: I0317 19:11:13.043213 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.043275 kubelet[2396]: I0317 19:11:13.043235 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cni-path" (OuterVolumeSpecName: "cni-path") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.043275 kubelet[2396]: I0317 19:11:13.043255 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.043425 kubelet[2396]: I0317 19:11:13.043275 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.043425 kubelet[2396]: I0317 19:11:13.043407 2396 scope.go:117] "RemoveContainer" containerID="0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6" Mar 17 19:11:13.045420 kubelet[2396]: I0317 19:11:13.045367 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:13.045527 kubelet[2396]: I0317 19:11:13.045501 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-kube-api-access-c8rzv" (OuterVolumeSpecName: "kube-api-access-c8rzv") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "kube-api-access-c8rzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 19:11:13.045606 kubelet[2396]: I0317 19:11:13.045588 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626fe57b-3983-4ba1-9e43-d893ba16fe35-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 19:11:13.045678 kubelet[2396]: I0317 19:11:13.045660 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "626fe57b-3983-4ba1-9e43-d893ba16fe35" (UID: "626fe57b-3983-4ba1-9e43-d893ba16fe35"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 19:11:13.046533 kubelet[2396]: I0317 19:11:13.046509 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22948d77-ed73-4aeb-91e0-8b4f22a58625-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "22948d77-ed73-4aeb-91e0-8b4f22a58625" (UID: "22948d77-ed73-4aeb-91e0-8b4f22a58625"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 19:11:13.046617 env[1416]: time="2025-03-17T19:11:13.046592760Z" level=info msg="RemoveContainer for \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\"" Mar 17 19:11:13.048901 kubelet[2396]: I0317 19:11:13.048868 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22948d77-ed73-4aeb-91e0-8b4f22a58625-kube-api-access-7v8sw" (OuterVolumeSpecName: "kube-api-access-7v8sw") pod "22948d77-ed73-4aeb-91e0-8b4f22a58625" (UID: "22948d77-ed73-4aeb-91e0-8b4f22a58625"). InnerVolumeSpecName "kube-api-access-7v8sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 19:11:13.052651 env[1416]: time="2025-03-17T19:11:13.052618930Z" level=info msg="RemoveContainer for \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\" returns successfully" Mar 17 19:11:13.052864 kubelet[2396]: I0317 19:11:13.052794 2396 scope.go:117] "RemoveContainer" containerID="fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059" Mar 17 19:11:13.053262 env[1416]: time="2025-03-17T19:11:13.053173837Z" level=error msg="ContainerStatus for \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\": not found" Mar 17 19:11:13.053527 kubelet[2396]: E0317 19:11:13.053500 2396 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\": not found" containerID="fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059" Mar 17 19:11:13.053631 kubelet[2396]: I0317 19:11:13.053533 2396 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059"} err="failed to get container status \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\": rpc error: code = NotFound desc = an error occurred when try to find container \"fb201325c879b2f1686e9933318d0c06382b3b7e5f8623ac03ce15940d6f5059\": not found" Mar 17 19:11:13.053631 kubelet[2396]: I0317 19:11:13.053562 2396 scope.go:117] "RemoveContainer" containerID="e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e" Mar 17 19:11:13.053882 env[1416]: time="2025-03-17T19:11:13.053831944Z" level=error msg="ContainerStatus for \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\": not found" Mar 17 19:11:13.054053 kubelet[2396]: E0317 19:11:13.054026 2396 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\": not found" containerID="e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e" Mar 17 19:11:13.054127 kubelet[2396]: I0317 19:11:13.054051 2396 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e"} err="failed to get container status \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\": rpc error: code = NotFound desc = an error occurred when try to find container \"e61d5da82d65a3500cd377260065360a97e76609910a7c48e6141cb9a4db542e\": not found" Mar 17 19:11:13.054127 kubelet[2396]: I0317 19:11:13.054073 2396 scope.go:117] "RemoveContainer" containerID="c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1" Mar 17 19:11:13.054317 env[1416]: time="2025-03-17T19:11:13.054251549Z" level=error msg="ContainerStatus for \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\": not found" Mar 17 19:11:13.054503 kubelet[2396]: E0317 19:11:13.054481 2396 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\": not found" containerID="c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1" Mar 17 19:11:13.054589 kubelet[2396]: I0317 19:11:13.054508 2396 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1"} err="failed to get container status \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\": rpc error: code = NotFound desc = an error occurred when try to find container \"c99025021ea3b707fb799ba97b884836c553a1ac4528aa85aa2da521cde9fbe1\": not found" Mar 17 19:11:13.054589 kubelet[2396]: I0317 19:11:13.054527 2396 scope.go:117] "RemoveContainer" containerID="99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67" Mar 17 19:11:13.054849 env[1416]: time="2025-03-17T19:11:13.054802356Z" level=error msg="ContainerStatus for \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\": not found" Mar 17 19:11:13.055013 kubelet[2396]: E0317 19:11:13.054984 2396 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\": not found" containerID="99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67" Mar 17 19:11:13.055086 kubelet[2396]: I0317 19:11:13.055007 2396 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67"} err="failed to get container status \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\": rpc error: code = NotFound desc = an error occurred when try to find container \"99f3d8c43b36779b62300c7e070e7046addda2c985b6a8f1e273aec80ee49a67\": not found" Mar 17 19:11:13.055086 kubelet[2396]: I0317 19:11:13.055026 2396 scope.go:117] "RemoveContainer" containerID="0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6" Mar 17 19:11:13.055252 env[1416]: time="2025-03-17T19:11:13.055199160Z" level=error msg="ContainerStatus for \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\": not found" Mar 17 19:11:13.055350 kubelet[2396]: E0317 19:11:13.055329 2396 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\": not found" containerID="0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6" Mar 17 19:11:13.055432 kubelet[2396]: I0317 19:11:13.055358 2396 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6"} err="failed to get container status \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\": rpc error: code = NotFound desc = an error occurred when try to find container \"0b80913b7680b09e64aabfe7eb5b8e9a9be0c9669edd3855420bd7356ab9ccc6\": not found" Mar 17 19:11:13.135915 kubelet[2396]: I0317 19:11:13.135850 2396 reconciler_common.go:288] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-bpf-maps\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.135915 kubelet[2396]: I0317 19:11:13.135901 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-config-path\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.135915 kubelet[2396]: I0317 19:11:13.135921 2396 reconciler_common.go:288] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-xtables-lock\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.135940 2396 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-c8rzv\" (UniqueName: \"kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-kube-api-access-c8rzv\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.135957 2396 reconciler_common.go:288] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/626fe57b-3983-4ba1-9e43-d893ba16fe35-hubble-tls\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.135972 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-run\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.135983 2396 reconciler_common.go:288] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-hostproc\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.135994 2396 reconciler_common.go:288] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/626fe57b-3983-4ba1-9e43-d893ba16fe35-clustermesh-secrets\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.136007 2396 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-7v8sw\" (UniqueName: \"kubernetes.io/projected/22948d77-ed73-4aeb-91e0-8b4f22a58625-kube-api-access-7v8sw\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.136020 2396 reconciler_common.go:288] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-etc-cni-netd\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136256 kubelet[2396]: I0317 19:11:13.136034 2396 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-net\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136584 kubelet[2396]: I0317 19:11:13.136047 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/22948d77-ed73-4aeb-91e0-8b4f22a58625-cilium-config-path\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136584 kubelet[2396]: I0317 19:11:13.136060 2396 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-host-proc-sys-kernel\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136584 kubelet[2396]: I0317 19:11:13.136079 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cilium-cgroup\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136584 kubelet[2396]: I0317 19:11:13.136092 2396 reconciler_common.go:288] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-lib-modules\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.136584 kubelet[2396]: I0317 19:11:13.136110 2396 reconciler_common.go:288] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/626fe57b-3983-4ba1-9e43-d893ba16fe35-cni-path\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:13.294030 systemd[1]: Removed slice kubepods-besteffort-pod22948d77_ed73_4aeb_91e0_8b4f22a58625.slice. Mar 17 19:11:13.302549 systemd[1]: Removed slice kubepods-burstable-pod626fe57b_3983_4ba1_9e43_d893ba16fe35.slice. Mar 17 19:11:13.302663 systemd[1]: kubepods-burstable-pod626fe57b_3983_4ba1_9e43_d893ba16fe35.slice: Consumed 7.277s CPU time. Mar 17 19:11:13.486858 kubelet[2396]: I0317 19:11:13.486807 2396 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22948d77-ed73-4aeb-91e0-8b4f22a58625" path="/var/lib/kubelet/pods/22948d77-ed73-4aeb-91e0-8b4f22a58625/volumes" Mar 17 19:11:13.487409 kubelet[2396]: I0317 19:11:13.487355 2396 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="626fe57b-3983-4ba1-9e43-d893ba16fe35" path="/var/lib/kubelet/pods/626fe57b-3983-4ba1-9e43-d893ba16fe35/volumes" Mar 17 19:11:13.597318 kubelet[2396]: E0317 19:11:13.597208 2396 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 17 19:11:13.659337 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5dd056bb6f9c716414877471ad6d9dda1e10aaa408d7ee01066f0d7d21b5850b-rootfs.mount: Deactivated successfully. Mar 17 19:11:13.659499 systemd[1]: var-lib-kubelet-pods-22948d77\x2ded73\x2d4aeb\x2d91e0\x2d8b4f22a58625-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d7v8sw.mount: Deactivated successfully. Mar 17 19:11:13.659610 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4-rootfs.mount: Deactivated successfully. Mar 17 19:11:13.659704 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d28d3d231622255ee000838c3f6158da5a149dd1ad5138a696fe7461e8a851a4-shm.mount: Deactivated successfully. Mar 17 19:11:13.659799 systemd[1]: var-lib-kubelet-pods-626fe57b\x2d3983\x2d4ba1\x2d9e43\x2dd893ba16fe35-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dc8rzv.mount: Deactivated successfully. Mar 17 19:11:13.659897 systemd[1]: var-lib-kubelet-pods-626fe57b\x2d3983\x2d4ba1\x2d9e43\x2dd893ba16fe35-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 17 19:11:13.659989 systemd[1]: var-lib-kubelet-pods-626fe57b\x2d3983\x2d4ba1\x2d9e43\x2dd893ba16fe35-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 17 19:11:14.718309 sshd[3954]: pam_unix(sshd:session): session closed for user core Mar 17 19:11:14.722113 systemd[1]: sshd@20-10.200.8.15:22-10.200.16.10:48270.service: Deactivated successfully. Mar 17 19:11:14.723172 systemd[1]: session-23.scope: Deactivated successfully. Mar 17 19:11:14.723415 systemd[1]: session-23.scope: Consumed 1.008s CPU time. Mar 17 19:11:14.724298 systemd-logind[1401]: Session 23 logged out. Waiting for processes to exit. Mar 17 19:11:14.725362 systemd-logind[1401]: Removed session 23. Mar 17 19:11:14.823973 systemd[1]: Started sshd@21-10.200.8.15:22-10.200.16.10:48280.service. Mar 17 19:11:15.449754 sshd[4119]: Accepted publickey for core from 10.200.16.10 port 48280 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:11:15.451494 sshd[4119]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:11:15.457426 systemd[1]: Started session-24.scope. Mar 17 19:11:15.458228 systemd-logind[1401]: New session 24 of user core. Mar 17 19:11:16.376902 kubelet[2396]: E0317 19:11:16.376860 2396 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="626fe57b-3983-4ba1-9e43-d893ba16fe35" containerName="apply-sysctl-overwrites" Mar 17 19:11:16.376902 kubelet[2396]: E0317 19:11:16.376897 2396 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="626fe57b-3983-4ba1-9e43-d893ba16fe35" containerName="mount-bpf-fs" Mar 17 19:11:16.376902 kubelet[2396]: E0317 19:11:16.376907 2396 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="22948d77-ed73-4aeb-91e0-8b4f22a58625" containerName="cilium-operator" Mar 17 19:11:16.376902 kubelet[2396]: E0317 19:11:16.376915 2396 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="626fe57b-3983-4ba1-9e43-d893ba16fe35" containerName="clean-cilium-state" Mar 17 19:11:16.377525 kubelet[2396]: E0317 19:11:16.376925 2396 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="626fe57b-3983-4ba1-9e43-d893ba16fe35" containerName="cilium-agent" Mar 17 19:11:16.377525 kubelet[2396]: E0317 19:11:16.376933 2396 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="626fe57b-3983-4ba1-9e43-d893ba16fe35" containerName="mount-cgroup" Mar 17 19:11:16.377525 kubelet[2396]: I0317 19:11:16.376993 2396 memory_manager.go:354] "RemoveStaleState removing state" podUID="626fe57b-3983-4ba1-9e43-d893ba16fe35" containerName="cilium-agent" Mar 17 19:11:16.377525 kubelet[2396]: I0317 19:11:16.377004 2396 memory_manager.go:354] "RemoveStaleState removing state" podUID="22948d77-ed73-4aeb-91e0-8b4f22a58625" containerName="cilium-operator" Mar 17 19:11:16.383925 systemd[1]: Created slice kubepods-burstable-podc0ec4ce1_3e3f_435d_9b27_ae37fac7cd98.slice. Mar 17 19:11:16.471255 sshd[4119]: pam_unix(sshd:session): session closed for user core Mar 17 19:11:16.474846 systemd[1]: sshd@21-10.200.8.15:22-10.200.16.10:48280.service: Deactivated successfully. Mar 17 19:11:16.476350 systemd[1]: session-24.scope: Deactivated successfully. Mar 17 19:11:16.476586 systemd-logind[1401]: Session 24 logged out. Waiting for processes to exit. Mar 17 19:11:16.477920 systemd-logind[1401]: Removed session 24. Mar 17 19:11:16.555628 kubelet[2396]: I0317 19:11:16.555562 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-run\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.555628 kubelet[2396]: I0317 19:11:16.555622 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-lib-modules\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.555929 kubelet[2396]: I0317 19:11:16.555654 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hubble-tls\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.555929 kubelet[2396]: I0317 19:11:16.555681 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-etc-cni-netd\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.555929 kubelet[2396]: I0317 19:11:16.555705 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-kernel\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.555929 kubelet[2396]: I0317 19:11:16.555730 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-ipsec-secrets\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.555929 kubelet[2396]: I0317 19:11:16.555755 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-cgroup\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.555929 kubelet[2396]: I0317 19:11:16.555777 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-config-path\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.556238 kubelet[2396]: I0317 19:11:16.555802 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkf6l\" (UniqueName: \"kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-kube-api-access-kkf6l\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.556238 kubelet[2396]: I0317 19:11:16.555828 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-xtables-lock\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.556238 kubelet[2396]: I0317 19:11:16.555855 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-net\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.556238 kubelet[2396]: I0317 19:11:16.555879 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-bpf-maps\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.556238 kubelet[2396]: I0317 19:11:16.555905 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hostproc\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.556238 kubelet[2396]: I0317 19:11:16.555928 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cni-path\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.556470 kubelet[2396]: I0317 19:11:16.555953 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-clustermesh-secrets\") pod \"cilium-n8h48\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " pod="kube-system/cilium-n8h48" Mar 17 19:11:16.577119 systemd[1]: Started sshd@22-10.200.8.15:22-10.200.16.10:48292.service. Mar 17 19:11:16.990433 env[1416]: time="2025-03-17T19:11:16.988970983Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-n8h48,Uid:c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98,Namespace:kube-system,Attempt:0,}" Mar 17 19:11:17.022366 env[1416]: time="2025-03-17T19:11:17.022283763Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:11:17.022366 env[1416]: time="2025-03-17T19:11:17.022321163Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:11:17.022366 env[1416]: time="2025-03-17T19:11:17.022334963Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:11:17.022742 env[1416]: time="2025-03-17T19:11:17.022695967Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02 pid=4141 runtime=io.containerd.runc.v2 Mar 17 19:11:17.038432 systemd[1]: Started cri-containerd-9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02.scope. Mar 17 19:11:17.065938 env[1416]: time="2025-03-17T19:11:17.065890158Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-n8h48,Uid:c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98,Namespace:kube-system,Attempt:0,} returns sandbox id \"9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02\"" Mar 17 19:11:17.069079 env[1416]: time="2025-03-17T19:11:17.069040894Z" level=info msg="CreateContainer within sandbox \"9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 17 19:11:17.095218 env[1416]: time="2025-03-17T19:11:17.095161791Z" level=info msg="CreateContainer within sandbox \"9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\"" Mar 17 19:11:17.095815 env[1416]: time="2025-03-17T19:11:17.095784498Z" level=info msg="StartContainer for \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\"" Mar 17 19:11:17.115666 systemd[1]: Started cri-containerd-b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a.scope. Mar 17 19:11:17.129003 systemd[1]: cri-containerd-b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a.scope: Deactivated successfully. Mar 17 19:11:17.186355 env[1416]: time="2025-03-17T19:11:17.186289627Z" level=info msg="shim disconnected" id=b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a Mar 17 19:11:17.186355 env[1416]: time="2025-03-17T19:11:17.186353927Z" level=warning msg="cleaning up after shim disconnected" id=b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a namespace=k8s.io Mar 17 19:11:17.186355 env[1416]: time="2025-03-17T19:11:17.186366828Z" level=info msg="cleaning up dead shim" Mar 17 19:11:17.193947 env[1416]: time="2025-03-17T19:11:17.193895013Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:17Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4202 runtime=io.containerd.runc.v2\ntime=\"2025-03-17T19:11:17Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" Mar 17 19:11:17.194272 env[1416]: time="2025-03-17T19:11:17.194167716Z" level=error msg="copy shim log" error="read /proc/self/fd/30: file already closed" Mar 17 19:11:17.194619 env[1416]: time="2025-03-17T19:11:17.194571021Z" level=error msg="Failed to pipe stdout of container \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\"" error="reading from a closed fifo" Mar 17 19:11:17.194791 env[1416]: time="2025-03-17T19:11:17.194751023Z" level=error msg="Failed to pipe stderr of container \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\"" error="reading from a closed fifo" Mar 17 19:11:17.198601 env[1416]: time="2025-03-17T19:11:17.198545766Z" level=error msg="StartContainer for \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" Mar 17 19:11:17.198859 kubelet[2396]: E0317 19:11:17.198813 2396 log.go:32] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a" Mar 17 19:11:17.199021 kubelet[2396]: E0317 19:11:17.198998 2396 kuberuntime_manager.go:1272] "Unhandled Error" err=< Mar 17 19:11:17.199021 kubelet[2396]: init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; Mar 17 19:11:17.199021 kubelet[2396]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; Mar 17 19:11:17.199021 kubelet[2396]: rm /hostbin/cilium-mount Mar 17 19:11:17.199195 kubelet[2396]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kkf6l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:&AppArmorProfile{Type:Unconfined,LocalhostProfile:nil,},},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cilium-n8h48_kube-system(c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown Mar 17 19:11:17.199195 kubelet[2396]: > logger="UnhandledError" Mar 17 19:11:17.200986 kubelet[2396]: E0317 19:11:17.200953 2396 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-n8h48" podUID="c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" Mar 17 19:11:17.202315 sshd[4129]: Accepted publickey for core from 10.200.16.10 port 48292 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:11:17.204048 sshd[4129]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:11:17.208902 systemd[1]: Started session-25.scope. Mar 17 19:11:17.209348 systemd-logind[1401]: New session 25 of user core. Mar 17 19:11:17.726435 sshd[4129]: pam_unix(sshd:session): session closed for user core Mar 17 19:11:17.729968 systemd[1]: sshd@22-10.200.8.15:22-10.200.16.10:48292.service: Deactivated successfully. Mar 17 19:11:17.730964 systemd[1]: session-25.scope: Deactivated successfully. Mar 17 19:11:17.732088 systemd-logind[1401]: Session 25 logged out. Waiting for processes to exit. Mar 17 19:11:17.733338 systemd-logind[1401]: Removed session 25. Mar 17 19:11:17.831632 systemd[1]: Started sshd@23-10.200.8.15:22-10.200.16.10:48294.service. Mar 17 19:11:17.970913 kubelet[2396]: I0317 19:11:17.970854 2396 setters.go:600] "Node became not ready" node="ci-3510.3.7-a-a60daf7667" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-03-17T19:11:17Z","lastTransitionTime":"2025-03-17T19:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Mar 17 19:11:18.007320 env[1416]: time="2025-03-17T19:11:18.007174156Z" level=info msg="StopPodSandbox for \"9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02\"" Mar 17 19:11:18.007320 env[1416]: time="2025-03-17T19:11:18.007242157Z" level=info msg="Container to stop \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 19:11:18.010535 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02-shm.mount: Deactivated successfully. Mar 17 19:11:18.018820 systemd[1]: cri-containerd-9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02.scope: Deactivated successfully. Mar 17 19:11:18.050858 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02-rootfs.mount: Deactivated successfully. Mar 17 19:11:18.065835 env[1416]: time="2025-03-17T19:11:18.065780718Z" level=info msg="shim disconnected" id=9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02 Mar 17 19:11:18.066498 env[1416]: time="2025-03-17T19:11:18.066467325Z" level=warning msg="cleaning up after shim disconnected" id=9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02 namespace=k8s.io Mar 17 19:11:18.066642 env[1416]: time="2025-03-17T19:11:18.066626427Z" level=info msg="cleaning up dead shim" Mar 17 19:11:18.074829 env[1416]: time="2025-03-17T19:11:18.074788719Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:18Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4245 runtime=io.containerd.runc.v2\n" Mar 17 19:11:18.075150 env[1416]: time="2025-03-17T19:11:18.075117723Z" level=info msg="TearDown network for sandbox \"9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02\" successfully" Mar 17 19:11:18.075230 env[1416]: time="2025-03-17T19:11:18.075149224Z" level=info msg="StopPodSandbox for \"9f31403bac3c8de5f33b605f56deae140c381dfe46cd42a372c3c3980e622e02\" returns successfully" Mar 17 19:11:18.268271 kubelet[2396]: I0317 19:11:18.268139 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hubble-tls\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268271 kubelet[2396]: I0317 19:11:18.268192 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkf6l\" (UniqueName: \"kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-kube-api-access-kkf6l\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268271 kubelet[2396]: I0317 19:11:18.268219 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-lib-modules\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268271 kubelet[2396]: I0317 19:11:18.268245 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-config-path\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268271 kubelet[2396]: I0317 19:11:18.268266 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-net\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268291 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-ipsec-secrets\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268308 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-run\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268325 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-xtables-lock\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268345 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-etc-cni-netd\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268361 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-bpf-maps\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268394 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-clustermesh-secrets\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268413 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-cgroup\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268433 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hostproc\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268454 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-kernel\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268474 2396 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cni-path\") pod \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\" (UID: \"c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98\") " Mar 17 19:11:18.268649 kubelet[2396]: I0317 19:11:18.268540 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cni-path" (OuterVolumeSpecName: "cni-path") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.269269 kubelet[2396]: I0317 19:11:18.269239 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.271677 kubelet[2396]: I0317 19:11:18.271644 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.271786 kubelet[2396]: I0317 19:11:18.271692 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.274603 systemd[1]: var-lib-kubelet-pods-c0ec4ce1\x2d3e3f\x2d435d\x2d9b27\x2dae37fac7cd98-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkkf6l.mount: Deactivated successfully. Mar 17 19:11:18.276498 kubelet[2396]: I0317 19:11:18.275033 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.277268 kubelet[2396]: I0317 19:11:18.277237 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.277363 kubelet[2396]: I0317 19:11:18.277281 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hostproc" (OuterVolumeSpecName: "hostproc") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.277363 kubelet[2396]: I0317 19:11:18.277306 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.278730 kubelet[2396]: I0317 19:11:18.278697 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 19:11:18.279029 kubelet[2396]: I0317 19:11:18.279008 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-kube-api-access-kkf6l" (OuterVolumeSpecName: "kube-api-access-kkf6l") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "kube-api-access-kkf6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 19:11:18.279220 kubelet[2396]: I0317 19:11:18.279198 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.279336 kubelet[2396]: I0317 19:11:18.279321 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 19:11:18.282135 systemd[1]: var-lib-kubelet-pods-c0ec4ce1\x2d3e3f\x2d435d\x2d9b27\x2dae37fac7cd98-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Mar 17 19:11:18.283915 kubelet[2396]: I0317 19:11:18.283887 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 19:11:18.285405 kubelet[2396]: I0317 19:11:18.285346 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 19:11:18.286107 kubelet[2396]: I0317 19:11:18.286074 2396 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" (UID: "c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 19:11:18.368656 kubelet[2396]: I0317 19:11:18.368610 2396 reconciler_common.go:288] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hubble-tls\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368656 kubelet[2396]: I0317 19:11:18.368647 2396 reconciler_common.go:288] "Volume detached for volume \"kube-api-access-kkf6l\" (UniqueName: \"kubernetes.io/projected/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-kube-api-access-kkf6l\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368656 kubelet[2396]: I0317 19:11:18.368662 2396 reconciler_common.go:288] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-lib-modules\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368673 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-config-path\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368687 2396 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-net\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368700 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-run\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368712 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-ipsec-secrets\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368721 2396 reconciler_common.go:288] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-xtables-lock\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368731 2396 reconciler_common.go:288] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-etc-cni-netd\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368741 2396 reconciler_common.go:288] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-bpf-maps\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368751 2396 reconciler_common.go:288] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-clustermesh-secrets\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368762 2396 reconciler_common.go:288] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cilium-cgroup\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368772 2396 reconciler_common.go:288] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-hostproc\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368782 2396 reconciler_common.go:288] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-cni-path\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.368954 kubelet[2396]: I0317 19:11:18.368793 2396 reconciler_common.go:288] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98-host-proc-sys-kernel\") on node \"ci-3510.3.7-a-a60daf7667\" DevicePath \"\"" Mar 17 19:11:18.458002 sshd[4224]: Accepted publickey for core from 10.200.16.10 port 48294 ssh2: RSA SHA256:4lz3AVC5ESuijCWyVe68PdTUtpte7kghS3r+sVRnQ+c Mar 17 19:11:18.459581 sshd[4224]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 19:11:18.464520 systemd[1]: Started session-26.scope. Mar 17 19:11:18.465274 systemd-logind[1401]: New session 26 of user core. Mar 17 19:11:18.599473 kubelet[2396]: E0317 19:11:18.599297 2396 kubelet.go:2901] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 17 19:11:18.674540 systemd[1]: var-lib-kubelet-pods-c0ec4ce1\x2d3e3f\x2d435d\x2d9b27\x2dae37fac7cd98-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 17 19:11:18.674713 systemd[1]: var-lib-kubelet-pods-c0ec4ce1\x2d3e3f\x2d435d\x2d9b27\x2dae37fac7cd98-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 17 19:11:19.010761 kubelet[2396]: I0317 19:11:19.010729 2396 scope.go:117] "RemoveContainer" containerID="b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a" Mar 17 19:11:19.015329 env[1416]: time="2025-03-17T19:11:19.015202640Z" level=info msg="RemoveContainer for \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\"" Mar 17 19:11:19.017711 systemd[1]: Removed slice kubepods-burstable-podc0ec4ce1_3e3f_435d_9b27_ae37fac7cd98.slice. Mar 17 19:11:19.027316 env[1416]: time="2025-03-17T19:11:19.027231275Z" level=info msg="RemoveContainer for \"b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a\" returns successfully" Mar 17 19:11:19.058580 kubelet[2396]: E0317 19:11:19.058543 2396 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" containerName="mount-cgroup" Mar 17 19:11:19.058792 kubelet[2396]: I0317 19:11:19.058599 2396 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" containerName="mount-cgroup" Mar 17 19:11:19.064908 systemd[1]: Created slice kubepods-burstable-pod82b2df0a_1f7d_4c9a_b123_172657502d38.slice. Mar 17 19:11:19.171966 kubelet[2396]: I0317 19:11:19.171918 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-etc-cni-netd\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.171966 kubelet[2396]: I0317 19:11:19.171960 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/82b2df0a-1f7d-4c9a-b123-172657502d38-clustermesh-secrets\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.171983 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-host-proc-sys-net\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172011 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-host-proc-sys-kernel\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172029 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-cilium-cgroup\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172051 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-lib-modules\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172075 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gfkn\" (UniqueName: \"kubernetes.io/projected/82b2df0a-1f7d-4c9a-b123-172657502d38-kube-api-access-7gfkn\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172093 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-xtables-lock\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172114 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/82b2df0a-1f7d-4c9a-b123-172657502d38-cilium-ipsec-secrets\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172133 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-cni-path\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172153 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/82b2df0a-1f7d-4c9a-b123-172657502d38-hubble-tls\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172173 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-cilium-run\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172196 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-bpf-maps\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172224 kubelet[2396]: I0317 19:11:19.172216 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/82b2df0a-1f7d-4c9a-b123-172657502d38-hostproc\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.172637 kubelet[2396]: I0317 19:11:19.172236 2396 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/82b2df0a-1f7d-4c9a-b123-172657502d38-cilium-config-path\") pod \"cilium-s46g7\" (UID: \"82b2df0a-1f7d-4c9a-b123-172657502d38\") " pod="kube-system/cilium-s46g7" Mar 17 19:11:19.369436 env[1416]: time="2025-03-17T19:11:19.368460805Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-s46g7,Uid:82b2df0a-1f7d-4c9a-b123-172657502d38,Namespace:kube-system,Attempt:0,}" Mar 17 19:11:19.403425 env[1416]: time="2025-03-17T19:11:19.403332796Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 19:11:19.403640 env[1416]: time="2025-03-17T19:11:19.403404497Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 19:11:19.403640 env[1416]: time="2025-03-17T19:11:19.403418897Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 19:11:19.403880 env[1416]: time="2025-03-17T19:11:19.403838002Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776 pid=4279 runtime=io.containerd.runc.v2 Mar 17 19:11:19.416343 systemd[1]: Started cri-containerd-cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776.scope. Mar 17 19:11:19.439582 env[1416]: time="2025-03-17T19:11:19.439526203Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-s46g7,Uid:82b2df0a-1f7d-4c9a-b123-172657502d38,Namespace:kube-system,Attempt:0,} returns sandbox id \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\"" Mar 17 19:11:19.442944 env[1416]: time="2025-03-17T19:11:19.442904141Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 17 19:11:19.477744 env[1416]: time="2025-03-17T19:11:19.477684931Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d\"" Mar 17 19:11:19.480087 env[1416]: time="2025-03-17T19:11:19.478539841Z" level=info msg="StartContainer for \"796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d\"" Mar 17 19:11:19.486936 kubelet[2396]: I0317 19:11:19.486897 2396 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98" path="/var/lib/kubelet/pods/c0ec4ce1-3e3f-435d-9b27-ae37fac7cd98/volumes" Mar 17 19:11:19.498458 systemd[1]: Started cri-containerd-796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d.scope. Mar 17 19:11:19.529624 env[1416]: time="2025-03-17T19:11:19.529572613Z" level=info msg="StartContainer for \"796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d\" returns successfully" Mar 17 19:11:19.535203 systemd[1]: cri-containerd-796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d.scope: Deactivated successfully. Mar 17 19:11:19.580700 env[1416]: time="2025-03-17T19:11:19.580643587Z" level=info msg="shim disconnected" id=796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d Mar 17 19:11:19.580700 env[1416]: time="2025-03-17T19:11:19.580699387Z" level=warning msg="cleaning up after shim disconnected" id=796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d namespace=k8s.io Mar 17 19:11:19.581026 env[1416]: time="2025-03-17T19:11:19.580710687Z" level=info msg="cleaning up dead shim" Mar 17 19:11:19.588539 env[1416]: time="2025-03-17T19:11:19.588495775Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:19Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4366 runtime=io.containerd.runc.v2\n" Mar 17 19:11:20.018271 env[1416]: time="2025-03-17T19:11:20.017355987Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 17 19:11:20.048684 env[1416]: time="2025-03-17T19:11:20.048641036Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501\"" Mar 17 19:11:20.049217 env[1416]: time="2025-03-17T19:11:20.049143342Z" level=info msg="StartContainer for \"8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501\"" Mar 17 19:11:20.076482 systemd[1]: Started cri-containerd-8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501.scope. Mar 17 19:11:20.120514 env[1416]: time="2025-03-17T19:11:20.120456838Z" level=info msg="StartContainer for \"8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501\" returns successfully" Mar 17 19:11:20.127155 systemd[1]: cri-containerd-8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501.scope: Deactivated successfully. Mar 17 19:11:20.162355 env[1416]: time="2025-03-17T19:11:20.162301004Z" level=info msg="shim disconnected" id=8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501 Mar 17 19:11:20.162355 env[1416]: time="2025-03-17T19:11:20.162354305Z" level=warning msg="cleaning up after shim disconnected" id=8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501 namespace=k8s.io Mar 17 19:11:20.162646 env[1416]: time="2025-03-17T19:11:20.162369005Z" level=info msg="cleaning up dead shim" Mar 17 19:11:20.171000 env[1416]: time="2025-03-17T19:11:20.170955701Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:20Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4428 runtime=io.containerd.runc.v2\n" Mar 17 19:11:20.290981 kubelet[2396]: W0317 19:11:20.290817 2396 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0ec4ce1_3e3f_435d_9b27_ae37fac7cd98.slice/cri-containerd-b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a.scope WatchSource:0}: container "b17307b120521ed94499ce249ccc9447493e354af215e820328a872363ac740a" in namespace "k8s.io": not found Mar 17 19:11:20.674714 systemd[1]: run-containerd-runc-k8s.io-8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501-runc.W1ERsv.mount: Deactivated successfully. Mar 17 19:11:20.674856 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501-rootfs.mount: Deactivated successfully. Mar 17 19:11:21.021639 env[1416]: time="2025-03-17T19:11:21.021514188Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 17 19:11:21.056542 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount517923448.mount: Deactivated successfully. Mar 17 19:11:21.068249 env[1416]: time="2025-03-17T19:11:21.068198706Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1\"" Mar 17 19:11:21.068853 env[1416]: time="2025-03-17T19:11:21.068809913Z" level=info msg="StartContainer for \"fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1\"" Mar 17 19:11:21.091110 systemd[1]: Started cri-containerd-fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1.scope. Mar 17 19:11:21.127991 env[1416]: time="2025-03-17T19:11:21.127947569Z" level=info msg="StartContainer for \"fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1\" returns successfully" Mar 17 19:11:21.128274 systemd[1]: cri-containerd-fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1.scope: Deactivated successfully. Mar 17 19:11:21.167370 env[1416]: time="2025-03-17T19:11:21.167322305Z" level=info msg="shim disconnected" id=fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1 Mar 17 19:11:21.167661 env[1416]: time="2025-03-17T19:11:21.167631909Z" level=warning msg="cleaning up after shim disconnected" id=fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1 namespace=k8s.io Mar 17 19:11:21.167661 env[1416]: time="2025-03-17T19:11:21.167652909Z" level=info msg="cleaning up dead shim" Mar 17 19:11:21.176150 env[1416]: time="2025-03-17T19:11:21.176109003Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:21Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4490 runtime=io.containerd.runc.v2\n" Mar 17 19:11:21.674815 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1-rootfs.mount: Deactivated successfully. Mar 17 19:11:22.027182 env[1416]: time="2025-03-17T19:11:22.025204516Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 17 19:11:22.050321 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3167196915.mount: Deactivated successfully. Mar 17 19:11:22.059833 env[1416]: time="2025-03-17T19:11:22.059779997Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a\"" Mar 17 19:11:22.060556 env[1416]: time="2025-03-17T19:11:22.060502405Z" level=info msg="StartContainer for \"c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a\"" Mar 17 19:11:22.085707 systemd[1]: Started cri-containerd-c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a.scope. Mar 17 19:11:22.110562 systemd[1]: cri-containerd-c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a.scope: Deactivated successfully. Mar 17 19:11:22.117825 env[1416]: time="2025-03-17T19:11:22.117773336Z" level=info msg="StartContainer for \"c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a\" returns successfully" Mar 17 19:11:22.151971 env[1416]: time="2025-03-17T19:11:22.151909913Z" level=info msg="shim disconnected" id=c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a Mar 17 19:11:22.151971 env[1416]: time="2025-03-17T19:11:22.151964013Z" level=warning msg="cleaning up after shim disconnected" id=c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a namespace=k8s.io Mar 17 19:11:22.151971 env[1416]: time="2025-03-17T19:11:22.151977213Z" level=info msg="cleaning up dead shim" Mar 17 19:11:22.159946 env[1416]: time="2025-03-17T19:11:22.159899401Z" level=warning msg="cleanup warnings time=\"2025-03-17T19:11:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4549 runtime=io.containerd.runc.v2\n" Mar 17 19:11:22.674877 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a-rootfs.mount: Deactivated successfully. Mar 17 19:11:23.030774 env[1416]: time="2025-03-17T19:11:23.030650796Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 17 19:11:23.067329 env[1416]: time="2025-03-17T19:11:23.067273997Z" level=info msg="CreateContainer within sandbox \"cc838ad9afdd9215505b03b22512dc26f88ad4d602ce339afde55a8d465cb776\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685\"" Mar 17 19:11:23.068154 env[1416]: time="2025-03-17T19:11:23.068112606Z" level=info msg="StartContainer for \"67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685\"" Mar 17 19:11:23.097409 systemd[1]: Started cri-containerd-67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685.scope. Mar 17 19:11:23.136574 env[1416]: time="2025-03-17T19:11:23.136522956Z" level=info msg="StartContainer for \"67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685\" returns successfully" Mar 17 19:11:23.404682 kubelet[2396]: W0317 19:11:23.404628 2396 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82b2df0a_1f7d_4c9a_b123_172657502d38.slice/cri-containerd-796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d.scope WatchSource:0}: task 796e698ece8b44549eaec2a5032f712c0a570d0f9673b4b0cdcec7af1dc2505d not found: not found Mar 17 19:11:23.613413 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Mar 17 19:11:23.674733 systemd[1]: run-containerd-runc-k8s.io-67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685-runc.5L3rZx.mount: Deactivated successfully. Mar 17 19:11:24.048925 kubelet[2396]: I0317 19:11:24.048782 2396 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-s46g7" podStartSLOduration=5.048760348 podStartE2EDuration="5.048760348s" podCreationTimestamp="2025-03-17 19:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 19:11:24.047931739 +0000 UTC m=+220.703048536" watchObservedRunningTime="2025-03-17 19:11:24.048760348 +0000 UTC m=+220.703877145" Mar 17 19:11:24.983689 systemd[1]: run-containerd-runc-k8s.io-67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685-runc.r6oGvo.mount: Deactivated successfully. Mar 17 19:11:26.355562 systemd-networkd[1571]: lxc_health: Link UP Mar 17 19:11:26.368980 systemd-networkd[1571]: lxc_health: Gained carrier Mar 17 19:11:26.369481 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Mar 17 19:11:26.514400 kubelet[2396]: W0317 19:11:26.512255 2396 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82b2df0a_1f7d_4c9a_b123_172657502d38.slice/cri-containerd-8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501.scope WatchSource:0}: task 8c0442f5e459209f7f56bc463151d46377a49a816604b0fd8e38ad17434e9501 not found: not found Mar 17 19:11:27.162540 systemd[1]: run-containerd-runc-k8s.io-67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685-runc.mn2PH7.mount: Deactivated successfully. Mar 17 19:11:27.546550 systemd-networkd[1571]: lxc_health: Gained IPv6LL Mar 17 19:11:29.624365 kubelet[2396]: W0317 19:11:29.624307 2396 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82b2df0a_1f7d_4c9a_b123_172657502d38.slice/cri-containerd-fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1.scope WatchSource:0}: task fa9aa1740c93005f573d8810ed97fc6323f6d7db45f6f00dc77910c5284731a1 not found: not found Mar 17 19:11:31.538945 systemd[1]: run-containerd-runc-k8s.io-67ff892dde0a6c2e02dfdcb70fade58d8d9338fb4c162e77ed15befa4d641685-runc.BDEqK3.mount: Deactivated successfully. Mar 17 19:11:31.708703 sshd[4224]: pam_unix(sshd:session): session closed for user core Mar 17 19:11:31.712525 systemd[1]: sshd@23-10.200.8.15:22-10.200.16.10:48294.service: Deactivated successfully. Mar 17 19:11:31.713619 systemd[1]: session-26.scope: Deactivated successfully. Mar 17 19:11:31.714329 systemd-logind[1401]: Session 26 logged out. Waiting for processes to exit. Mar 17 19:11:31.715207 systemd-logind[1401]: Removed session 26. Mar 17 19:11:32.734477 kubelet[2396]: W0317 19:11:32.734409 2396 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82b2df0a_1f7d_4c9a_b123_172657502d38.slice/cri-containerd-c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a.scope WatchSource:0}: task c209f45d660abf83371c4fb23800c77e841b0a015ed9c0674299e6106c5ac47a not found: not found