May 17 00:35:06.043288 kernel: Linux version 5.15.182-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Fri May 16 23:09:52 -00 2025 May 17 00:35:06.043326 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=4aad7caeadb0359f379975532748a0b4ae6bb9b229507353e0f5ae84cb9335a0 May 17 00:35:06.043343 kernel: BIOS-provided physical RAM map: May 17 00:35:06.043352 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable May 17 00:35:06.043361 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable May 17 00:35:06.043371 kernel: BIOS-e820: [mem 0x00000000786ce000-0x000000007894dfff] reserved May 17 00:35:06.043382 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data May 17 00:35:06.043392 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS May 17 00:35:06.043408 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable May 17 00:35:06.043418 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved May 17 00:35:06.043429 kernel: NX (Execute Disable) protection: active May 17 00:35:06.043441 kernel: e820: update [mem 0x76813018-0x7681be57] usable ==> usable May 17 00:35:06.043451 kernel: e820: update [mem 0x76813018-0x7681be57] usable ==> usable May 17 00:35:06.043464 kernel: extended physical RAM map: May 17 00:35:06.043482 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable May 17 00:35:06.043493 kernel: reserve setup_data: [mem 0x0000000000100000-0x0000000076813017] usable May 17 00:35:06.043506 kernel: reserve setup_data: [mem 0x0000000076813018-0x000000007681be57] usable May 17 00:35:06.043519 kernel: reserve setup_data: [mem 0x000000007681be58-0x00000000786cdfff] usable May 17 00:35:06.043533 kernel: reserve setup_data: [mem 0x00000000786ce000-0x000000007894dfff] reserved May 17 00:35:06.043546 kernel: reserve setup_data: [mem 0x000000007894e000-0x000000007895dfff] ACPI data May 17 00:35:06.043559 kernel: reserve setup_data: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS May 17 00:35:06.043572 kernel: reserve setup_data: [mem 0x00000000789de000-0x000000007c97bfff] usable May 17 00:35:06.043585 kernel: reserve setup_data: [mem 0x000000007c97c000-0x000000007c9fffff] reserved May 17 00:35:06.043599 kernel: efi: EFI v2.70 by EDK II May 17 00:35:06.043614 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77004a98 May 17 00:35:06.043626 kernel: SMBIOS 2.7 present. May 17 00:35:06.043657 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 May 17 00:35:06.043669 kernel: Hypervisor detected: KVM May 17 00:35:06.043680 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 17 00:35:06.043693 kernel: kvm-clock: cpu 0, msr 5519a001, primary cpu clock May 17 00:35:06.043705 kernel: kvm-clock: using sched offset of 5268701541 cycles May 17 00:35:06.043717 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 17 00:35:06.043728 kernel: tsc: Detected 2499.996 MHz processor May 17 00:35:06.043739 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 17 00:35:06.043751 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 17 00:35:06.043765 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 May 17 00:35:06.043777 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 17 00:35:06.043787 kernel: Using GB pages for direct mapping May 17 00:35:06.043799 kernel: Secure boot disabled May 17 00:35:06.043810 kernel: ACPI: Early table checksum verification disabled May 17 00:35:06.043825 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) May 17 00:35:06.050672 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) May 17 00:35:06.050723 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) May 17 00:35:06.050737 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) May 17 00:35:06.050751 kernel: ACPI: FACS 0x00000000789D0000 000040 May 17 00:35:06.050765 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) May 17 00:35:06.050779 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) May 17 00:35:06.050793 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) May 17 00:35:06.050806 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) May 17 00:35:06.050822 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) May 17 00:35:06.050836 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) May 17 00:35:06.050850 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) May 17 00:35:06.050864 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) May 17 00:35:06.050878 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] May 17 00:35:06.050892 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] May 17 00:35:06.050905 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] May 17 00:35:06.050919 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] May 17 00:35:06.050932 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] May 17 00:35:06.050948 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] May 17 00:35:06.050961 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] May 17 00:35:06.050975 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] May 17 00:35:06.050989 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] May 17 00:35:06.051002 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] May 17 00:35:06.051016 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] May 17 00:35:06.051029 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 May 17 00:35:06.051042 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 May 17 00:35:06.051056 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] May 17 00:35:06.051072 kernel: NUMA: Initialized distance table, cnt=1 May 17 00:35:06.051086 kernel: NODE_DATA(0) allocated [mem 0x7a8ef000-0x7a8f4fff] May 17 00:35:06.051100 kernel: Zone ranges: May 17 00:35:06.051113 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 17 00:35:06.051127 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] May 17 00:35:06.051140 kernel: Normal empty May 17 00:35:06.051154 kernel: Movable zone start for each node May 17 00:35:06.051167 kernel: Early memory node ranges May 17 00:35:06.051181 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] May 17 00:35:06.051197 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] May 17 00:35:06.051211 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] May 17 00:35:06.051224 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] May 17 00:35:06.051237 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 17 00:35:06.051251 kernel: On node 0, zone DMA: 96 pages in unavailable ranges May 17 00:35:06.051264 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges May 17 00:35:06.051278 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges May 17 00:35:06.051291 kernel: ACPI: PM-Timer IO Port: 0xb008 May 17 00:35:06.051305 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 17 00:35:06.051321 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 May 17 00:35:06.051335 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 17 00:35:06.051349 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 17 00:35:06.051362 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 17 00:35:06.051376 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 17 00:35:06.051389 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 17 00:35:06.051402 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 17 00:35:06.051416 kernel: TSC deadline timer available May 17 00:35:06.051429 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs May 17 00:35:06.051445 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices May 17 00:35:06.051459 kernel: Booting paravirtualized kernel on KVM May 17 00:35:06.051473 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 17 00:35:06.051487 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 May 17 00:35:06.051500 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u1048576 May 17 00:35:06.051514 kernel: pcpu-alloc: s188696 r8192 d32488 u1048576 alloc=1*2097152 May 17 00:35:06.051527 kernel: pcpu-alloc: [0] 0 1 May 17 00:35:06.051541 kernel: kvm-guest: stealtime: cpu 0, msr 7a41c0c0 May 17 00:35:06.051554 kernel: kvm-guest: PV spinlocks enabled May 17 00:35:06.051571 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) May 17 00:35:06.051584 kernel: Built 1 zonelists, mobility grouping on. Total pages: 501318 May 17 00:35:06.051598 kernel: Policy zone: DMA32 May 17 00:35:06.051613 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=4aad7caeadb0359f379975532748a0b4ae6bb9b229507353e0f5ae84cb9335a0 May 17 00:35:06.051628 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 17 00:35:06.051651 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 17 00:35:06.051665 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 17 00:35:06.051679 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 17 00:35:06.051696 kernel: Memory: 1876640K/2037804K available (12294K kernel code, 2276K rwdata, 13724K rodata, 47472K init, 4108K bss, 160904K reserved, 0K cma-reserved) May 17 00:35:06.051710 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 17 00:35:06.051723 kernel: Kernel/User page tables isolation: enabled May 17 00:35:06.051737 kernel: ftrace: allocating 34585 entries in 136 pages May 17 00:35:06.051751 kernel: ftrace: allocated 136 pages with 2 groups May 17 00:35:06.051765 kernel: rcu: Hierarchical RCU implementation. May 17 00:35:06.051779 kernel: rcu: RCU event tracing is enabled. May 17 00:35:06.051807 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 17 00:35:06.051822 kernel: Rude variant of Tasks RCU enabled. May 17 00:35:06.051836 kernel: Tracing variant of Tasks RCU enabled. May 17 00:35:06.051851 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 17 00:35:06.051865 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 17 00:35:06.051882 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 17 00:35:06.051897 kernel: random: crng init done May 17 00:35:06.051912 kernel: Console: colour dummy device 80x25 May 17 00:35:06.051926 kernel: printk: console [tty0] enabled May 17 00:35:06.051940 kernel: printk: console [ttyS0] enabled May 17 00:35:06.051954 kernel: ACPI: Core revision 20210730 May 17 00:35:06.051969 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns May 17 00:35:06.051986 kernel: APIC: Switch to symmetric I/O mode setup May 17 00:35:06.052000 kernel: x2apic enabled May 17 00:35:06.052015 kernel: Switched APIC routing to physical x2apic. May 17 00:35:06.052031 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns May 17 00:35:06.052045 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499996) May 17 00:35:06.052060 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 May 17 00:35:06.052074 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 May 17 00:35:06.052091 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 17 00:35:06.052105 kernel: Spectre V2 : Mitigation: Retpolines May 17 00:35:06.052119 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT May 17 00:35:06.052133 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! May 17 00:35:06.052148 kernel: RETBleed: Vulnerable May 17 00:35:06.052162 kernel: Speculative Store Bypass: Vulnerable May 17 00:35:06.052176 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode May 17 00:35:06.052191 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 17 00:35:06.052205 kernel: GDS: Unknown: Dependent on hypervisor status May 17 00:35:06.052219 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 17 00:35:06.052233 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 17 00:35:06.052250 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 17 00:35:06.052264 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' May 17 00:35:06.052279 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' May 17 00:35:06.052293 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' May 17 00:35:06.052307 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' May 17 00:35:06.052321 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' May 17 00:35:06.052335 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' May 17 00:35:06.052350 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 17 00:35:06.052364 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 May 17 00:35:06.052378 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 May 17 00:35:06.052392 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 May 17 00:35:06.052409 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 May 17 00:35:06.052422 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 May 17 00:35:06.052434 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 May 17 00:35:06.052445 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. May 17 00:35:06.052456 kernel: Freeing SMP alternatives memory: 32K May 17 00:35:06.052466 kernel: pid_max: default: 32768 minimum: 301 May 17 00:35:06.052477 kernel: LSM: Security Framework initializing May 17 00:35:06.052488 kernel: SELinux: Initializing. May 17 00:35:06.052510 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 17 00:35:06.052523 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 17 00:35:06.052537 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) May 17 00:35:06.052554 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. May 17 00:35:06.052569 kernel: signal: max sigframe size: 3632 May 17 00:35:06.052584 kernel: rcu: Hierarchical SRCU implementation. May 17 00:35:06.052598 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 17 00:35:06.052613 kernel: smp: Bringing up secondary CPUs ... May 17 00:35:06.052628 kernel: x86: Booting SMP configuration: May 17 00:35:06.052654 kernel: .... node #0, CPUs: #1 May 17 00:35:06.052669 kernel: kvm-clock: cpu 1, msr 5519a041, secondary cpu clock May 17 00:35:06.052684 kernel: kvm-guest: stealtime: cpu 1, msr 7a51c0c0 May 17 00:35:06.052702 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. May 17 00:35:06.052719 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. May 17 00:35:06.052734 kernel: smp: Brought up 1 node, 2 CPUs May 17 00:35:06.052749 kernel: smpboot: Max logical packages: 1 May 17 00:35:06.052764 kernel: smpboot: Total of 2 processors activated (9999.98 BogoMIPS) May 17 00:35:06.052779 kernel: devtmpfs: initialized May 17 00:35:06.052794 kernel: x86/mm: Memory block size: 128MB May 17 00:35:06.052809 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) May 17 00:35:06.052823 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 17 00:35:06.052841 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 17 00:35:06.052856 kernel: pinctrl core: initialized pinctrl subsystem May 17 00:35:06.052871 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 17 00:35:06.052886 kernel: audit: initializing netlink subsys (disabled) May 17 00:35:06.052901 kernel: audit: type=2000 audit(1747442106.184:1): state=initialized audit_enabled=0 res=1 May 17 00:35:06.052916 kernel: thermal_sys: Registered thermal governor 'step_wise' May 17 00:35:06.052930 kernel: thermal_sys: Registered thermal governor 'user_space' May 17 00:35:06.052946 kernel: cpuidle: using governor menu May 17 00:35:06.052960 kernel: ACPI: bus type PCI registered May 17 00:35:06.052977 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 17 00:35:06.052992 kernel: dca service started, version 1.12.1 May 17 00:35:06.053007 kernel: PCI: Using configuration type 1 for base access May 17 00:35:06.053022 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 17 00:35:06.053037 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages May 17 00:35:06.053052 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages May 17 00:35:06.053067 kernel: ACPI: Added _OSI(Module Device) May 17 00:35:06.053082 kernel: ACPI: Added _OSI(Processor Device) May 17 00:35:06.053097 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 17 00:35:06.053114 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 17 00:35:06.053129 kernel: ACPI: Added _OSI(Linux-Dell-Video) May 17 00:35:06.053143 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) May 17 00:35:06.053159 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) May 17 00:35:06.053173 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded May 17 00:35:06.053189 kernel: ACPI: Interpreter enabled May 17 00:35:06.053203 kernel: ACPI: PM: (supports S0 S5) May 17 00:35:06.053218 kernel: ACPI: Using IOAPIC for interrupt routing May 17 00:35:06.053234 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 17 00:35:06.053252 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F May 17 00:35:06.053266 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 17 00:35:06.053503 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 17 00:35:06.059043 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. May 17 00:35:06.059089 kernel: acpiphp: Slot [3] registered May 17 00:35:06.059106 kernel: acpiphp: Slot [4] registered May 17 00:35:06.059122 kernel: acpiphp: Slot [5] registered May 17 00:35:06.059148 kernel: acpiphp: Slot [6] registered May 17 00:35:06.059163 kernel: acpiphp: Slot [7] registered May 17 00:35:06.059178 kernel: acpiphp: Slot [8] registered May 17 00:35:06.059193 kernel: acpiphp: Slot [9] registered May 17 00:35:06.059208 kernel: acpiphp: Slot [10] registered May 17 00:35:06.059223 kernel: acpiphp: Slot [11] registered May 17 00:35:06.059238 kernel: acpiphp: Slot [12] registered May 17 00:35:06.059252 kernel: acpiphp: Slot [13] registered May 17 00:35:06.059267 kernel: acpiphp: Slot [14] registered May 17 00:35:06.059282 kernel: acpiphp: Slot [15] registered May 17 00:35:06.059300 kernel: acpiphp: Slot [16] registered May 17 00:35:06.059314 kernel: acpiphp: Slot [17] registered May 17 00:35:06.059329 kernel: acpiphp: Slot [18] registered May 17 00:35:06.059344 kernel: acpiphp: Slot [19] registered May 17 00:35:06.059359 kernel: acpiphp: Slot [20] registered May 17 00:35:06.059373 kernel: acpiphp: Slot [21] registered May 17 00:35:06.059388 kernel: acpiphp: Slot [22] registered May 17 00:35:06.059403 kernel: acpiphp: Slot [23] registered May 17 00:35:06.059417 kernel: acpiphp: Slot [24] registered May 17 00:35:06.059434 kernel: acpiphp: Slot [25] registered May 17 00:35:06.059449 kernel: acpiphp: Slot [26] registered May 17 00:35:06.059464 kernel: acpiphp: Slot [27] registered May 17 00:35:06.059479 kernel: acpiphp: Slot [28] registered May 17 00:35:06.059496 kernel: acpiphp: Slot [29] registered May 17 00:35:06.059511 kernel: acpiphp: Slot [30] registered May 17 00:35:06.059525 kernel: acpiphp: Slot [31] registered May 17 00:35:06.059540 kernel: PCI host bridge to bus 0000:00 May 17 00:35:06.059725 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 17 00:35:06.059846 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 17 00:35:06.059949 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 17 00:35:06.060052 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] May 17 00:35:06.060153 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] May 17 00:35:06.060252 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 17 00:35:06.060390 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 May 17 00:35:06.060528 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 May 17 00:35:06.060668 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 May 17 00:35:06.060784 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI May 17 00:35:06.060897 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff May 17 00:35:06.061012 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff May 17 00:35:06.061129 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff May 17 00:35:06.061242 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff May 17 00:35:06.061361 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff May 17 00:35:06.061474 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff May 17 00:35:06.061594 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 May 17 00:35:06.061723 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x80000000-0x803fffff pref] May 17 00:35:06.061849 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] May 17 00:35:06.061968 kernel: pci 0000:00:03.0: BAR 0: assigned to efifb May 17 00:35:06.062082 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 17 00:35:06.062217 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 May 17 00:35:06.062337 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80404000-0x80407fff] May 17 00:35:06.062467 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 May 17 00:35:06.062585 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80400000-0x80403fff] May 17 00:35:06.062603 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 17 00:35:06.062619 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 17 00:35:06.062658 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 17 00:35:06.062677 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 17 00:35:06.062691 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 17 00:35:06.062706 kernel: iommu: Default domain type: Translated May 17 00:35:06.062721 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 17 00:35:06.062839 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device May 17 00:35:06.062956 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 17 00:35:06.063071 kernel: pci 0000:00:03.0: vgaarb: bridge control possible May 17 00:35:06.063089 kernel: vgaarb: loaded May 17 00:35:06.063107 kernel: pps_core: LinuxPPS API ver. 1 registered May 17 00:35:06.063121 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti May 17 00:35:06.063136 kernel: PTP clock support registered May 17 00:35:06.063150 kernel: Registered efivars operations May 17 00:35:06.063164 kernel: PCI: Using ACPI for IRQ routing May 17 00:35:06.063179 kernel: PCI: pci_cache_line_size set to 64 bytes May 17 00:35:06.063194 kernel: e820: reserve RAM buffer [mem 0x76813018-0x77ffffff] May 17 00:35:06.063208 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] May 17 00:35:06.063222 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] May 17 00:35:06.063236 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 May 17 00:35:06.063254 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter May 17 00:35:06.063269 kernel: clocksource: Switched to clocksource kvm-clock May 17 00:35:06.063283 kernel: VFS: Disk quotas dquot_6.6.0 May 17 00:35:06.063297 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 17 00:35:06.063312 kernel: pnp: PnP ACPI init May 17 00:35:06.063326 kernel: pnp: PnP ACPI: found 5 devices May 17 00:35:06.063341 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 17 00:35:06.063356 kernel: NET: Registered PF_INET protocol family May 17 00:35:06.063370 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) May 17 00:35:06.063387 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) May 17 00:35:06.063401 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 17 00:35:06.063416 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) May 17 00:35:06.063431 kernel: TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear) May 17 00:35:06.063445 kernel: TCP: Hash tables configured (established 16384 bind 16384) May 17 00:35:06.063459 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) May 17 00:35:06.063474 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) May 17 00:35:06.063489 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 17 00:35:06.063506 kernel: NET: Registered PF_XDP protocol family May 17 00:35:06.063617 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 17 00:35:06.063732 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 17 00:35:06.063874 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 17 00:35:06.065927 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] May 17 00:35:06.066086 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] May 17 00:35:06.066235 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 17 00:35:06.066379 kernel: pci 0000:00:01.0: Activating ISA DMA hang workarounds May 17 00:35:06.066409 kernel: PCI: CLS 0 bytes, default 64 May 17 00:35:06.066425 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 17 00:35:06.066441 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns May 17 00:35:06.066455 kernel: clocksource: Switched to clocksource tsc May 17 00:35:06.066470 kernel: Initialise system trusted keyrings May 17 00:35:06.066484 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 May 17 00:35:06.066500 kernel: Key type asymmetric registered May 17 00:35:06.066515 kernel: Asymmetric key parser 'x509' registered May 17 00:35:06.066530 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) May 17 00:35:06.066547 kernel: io scheduler mq-deadline registered May 17 00:35:06.066563 kernel: io scheduler kyber registered May 17 00:35:06.066578 kernel: io scheduler bfq registered May 17 00:35:06.066593 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 17 00:35:06.066608 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 17 00:35:06.066624 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 17 00:35:06.066659 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 17 00:35:06.066674 kernel: i8042: Warning: Keylock active May 17 00:35:06.066688 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 17 00:35:06.066707 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 17 00:35:06.066858 kernel: rtc_cmos 00:00: RTC can wake from S4 May 17 00:35:06.066983 kernel: rtc_cmos 00:00: registered as rtc0 May 17 00:35:06.067107 kernel: rtc_cmos 00:00: setting system clock to 2025-05-17T00:35:05 UTC (1747442105) May 17 00:35:06.067228 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram May 17 00:35:06.067246 kernel: intel_pstate: CPU model not supported May 17 00:35:06.067261 kernel: efifb: probing for efifb May 17 00:35:06.067277 kernel: efifb: framebuffer at 0x80000000, using 1876k, total 1875k May 17 00:35:06.067296 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 May 17 00:35:06.067310 kernel: efifb: scrolling: redraw May 17 00:35:06.067326 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 May 17 00:35:06.067341 kernel: Console: switching to colour frame buffer device 100x37 May 17 00:35:06.067357 kernel: fb0: EFI VGA frame buffer device May 17 00:35:06.067373 kernel: pstore: Registered efi as persistent store backend May 17 00:35:06.067414 kernel: NET: Registered PF_INET6 protocol family May 17 00:35:06.067432 kernel: Segment Routing with IPv6 May 17 00:35:06.067448 kernel: In-situ OAM (IOAM) with IPv6 May 17 00:35:06.067466 kernel: NET: Registered PF_PACKET protocol family May 17 00:35:06.067482 kernel: Key type dns_resolver registered May 17 00:35:06.067498 kernel: IPI shorthand broadcast: enabled May 17 00:35:06.067514 kernel: sched_clock: Marking stable (384005573, 133117276)->(595457296, -78334447) May 17 00:35:06.067530 kernel: registered taskstats version 1 May 17 00:35:06.067546 kernel: Loading compiled-in X.509 certificates May 17 00:35:06.067561 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.182-flatcar: 01ca23caa8e5879327538f9287e5164b3e97ac0c' May 17 00:35:06.067577 kernel: Key type .fscrypt registered May 17 00:35:06.067592 kernel: Key type fscrypt-provisioning registered May 17 00:35:06.067612 kernel: pstore: Using crash dump compression: deflate May 17 00:35:06.067627 kernel: ima: No TPM chip found, activating TPM-bypass! May 17 00:35:06.067658 kernel: ima: Allocated hash algorithm: sha1 May 17 00:35:06.067674 kernel: ima: No architecture policies found May 17 00:35:06.067689 kernel: clk: Disabling unused clocks May 17 00:35:06.067705 kernel: Freeing unused kernel image (initmem) memory: 47472K May 17 00:35:06.067721 kernel: Write protecting the kernel read-only data: 28672k May 17 00:35:06.067737 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K May 17 00:35:06.067753 kernel: Freeing unused kernel image (rodata/data gap) memory: 612K May 17 00:35:06.067771 kernel: Run /init as init process May 17 00:35:06.067787 kernel: with arguments: May 17 00:35:06.067802 kernel: /init May 17 00:35:06.067818 kernel: with environment: May 17 00:35:06.067834 kernel: HOME=/ May 17 00:35:06.067850 kernel: TERM=linux May 17 00:35:06.067865 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 17 00:35:06.067886 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 17 00:35:06.067909 systemd[1]: Detected virtualization amazon. May 17 00:35:06.067926 systemd[1]: Detected architecture x86-64. May 17 00:35:06.067941 systemd[1]: Running in initrd. May 17 00:35:06.067958 systemd[1]: No hostname configured, using default hostname. May 17 00:35:06.067971 systemd[1]: Hostname set to . May 17 00:35:06.067985 systemd[1]: Initializing machine ID from VM UUID. May 17 00:35:06.067997 systemd[1]: Queued start job for default target initrd.target. May 17 00:35:06.068014 systemd[1]: Started systemd-ask-password-console.path. May 17 00:35:06.068030 systemd[1]: Reached target cryptsetup.target. May 17 00:35:06.068044 systemd[1]: Reached target paths.target. May 17 00:35:06.068058 systemd[1]: Reached target slices.target. May 17 00:35:06.068072 systemd[1]: Reached target swap.target. May 17 00:35:06.068085 systemd[1]: Reached target timers.target. May 17 00:35:06.068103 systemd[1]: Listening on iscsid.socket. May 17 00:35:06.068117 systemd[1]: Listening on iscsiuio.socket. May 17 00:35:06.068131 systemd[1]: Listening on systemd-journald-audit.socket. May 17 00:35:06.068145 systemd[1]: Listening on systemd-journald-dev-log.socket. May 17 00:35:06.068158 systemd[1]: Listening on systemd-journald.socket. May 17 00:35:06.068172 systemd[1]: Listening on systemd-networkd.socket. May 17 00:35:06.068186 systemd[1]: Listening on systemd-udevd-control.socket. May 17 00:35:06.068202 systemd[1]: Listening on systemd-udevd-kernel.socket. May 17 00:35:06.068215 systemd[1]: Reached target sockets.target. May 17 00:35:06.068229 systemd[1]: Starting kmod-static-nodes.service... May 17 00:35:06.068243 systemd[1]: Finished network-cleanup.service. May 17 00:35:06.068258 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 17 00:35:06.068272 systemd[1]: Starting systemd-fsck-usr.service... May 17 00:35:06.068285 systemd[1]: Starting systemd-journald.service... May 17 00:35:06.068299 systemd[1]: Starting systemd-modules-load.service... May 17 00:35:06.068312 systemd[1]: Starting systemd-resolved.service... May 17 00:35:06.068330 systemd[1]: Starting systemd-vconsole-setup.service... May 17 00:35:06.068344 systemd[1]: Finished kmod-static-nodes.service. May 17 00:35:06.068358 kernel: audit: type=1130 audit(1747442106.051:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.068372 systemd[1]: Finished systemd-fsck-usr.service. May 17 00:35:06.068386 kernel: audit: type=1130 audit(1747442106.063:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.068408 systemd-journald[185]: Journal started May 17 00:35:06.068495 systemd-journald[185]: Runtime Journal (/run/log/journal/ec26a0d5e5b36a0451d17934fa6aaf3c) is 4.8M, max 38.3M, 33.5M free. May 17 00:35:06.051000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.063000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.083297 systemd[1]: Finished systemd-vconsole-setup.service. May 17 00:35:06.083385 systemd[1]: Started systemd-journald.service. May 17 00:35:06.083404 kernel: audit: type=1130 audit(1747442106.072:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.072000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.073255 systemd-modules-load[186]: Inserted module 'overlay' May 17 00:35:06.094783 kernel: audit: type=1130 audit(1747442106.082:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.082000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.085036 systemd[1]: Starting dracut-cmdline-ask.service... May 17 00:35:06.093528 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 17 00:35:06.109000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.110355 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 17 00:35:06.119501 kernel: audit: type=1130 audit(1747442106.109:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.126122 systemd[1]: Finished dracut-cmdline-ask.service. May 17 00:35:06.136269 kernel: audit: type=1130 audit(1747442106.125:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.125000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.128600 systemd[1]: Starting dracut-cmdline.service... May 17 00:35:06.146110 systemd-resolved[187]: Positive Trust Anchors: May 17 00:35:06.146128 systemd-resolved[187]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 17 00:35:06.146182 systemd-resolved[187]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 17 00:35:06.176138 kernel: audit: type=1130 audit(1747442106.154:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.176179 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 17 00:35:06.154000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.176297 dracut-cmdline[203]: dracut-dracut-053 May 17 00:35:06.180761 kernel: Bridge firewalling registered May 17 00:35:06.154091 systemd-resolved[187]: Defaulting to hostname 'linux'. May 17 00:35:06.182455 dracut-cmdline[203]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=4aad7caeadb0359f379975532748a0b4ae6bb9b229507353e0f5ae84cb9335a0 May 17 00:35:06.155366 systemd[1]: Started systemd-resolved.service. May 17 00:35:06.156793 systemd[1]: Reached target nss-lookup.target. May 17 00:35:06.178699 systemd-modules-load[186]: Inserted module 'br_netfilter' May 17 00:35:06.212659 kernel: SCSI subsystem initialized May 17 00:35:06.233895 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 17 00:35:06.233979 kernel: device-mapper: uevent: version 1.0.3 May 17 00:35:06.236437 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com May 17 00:35:06.241094 systemd-modules-load[186]: Inserted module 'dm_multipath' May 17 00:35:06.243167 systemd[1]: Finished systemd-modules-load.service. May 17 00:35:06.252165 kernel: audit: type=1130 audit(1747442106.242:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.242000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.251196 systemd[1]: Starting systemd-sysctl.service... May 17 00:35:06.263009 systemd[1]: Finished systemd-sysctl.service. May 17 00:35:06.262000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.272862 kernel: audit: type=1130 audit(1747442106.262:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.278666 kernel: Loading iSCSI transport class v2.0-870. May 17 00:35:06.298671 kernel: iscsi: registered transport (tcp) May 17 00:35:06.325322 kernel: iscsi: registered transport (qla4xxx) May 17 00:35:06.325420 kernel: QLogic iSCSI HBA Driver May 17 00:35:06.358689 systemd[1]: Finished dracut-cmdline.service. May 17 00:35:06.357000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.360878 systemd[1]: Starting dracut-pre-udev.service... May 17 00:35:06.417713 kernel: raid6: avx512x4 gen() 16782 MB/s May 17 00:35:06.435709 kernel: raid6: avx512x4 xor() 7017 MB/s May 17 00:35:06.453701 kernel: raid6: avx512x2 gen() 17890 MB/s May 17 00:35:06.472710 kernel: raid6: avx512x2 xor() 22349 MB/s May 17 00:35:06.490704 kernel: raid6: avx512x1 gen() 17099 MB/s May 17 00:35:06.508745 kernel: raid6: avx512x1 xor() 21561 MB/s May 17 00:35:06.526709 kernel: raid6: avx2x4 gen() 16558 MB/s May 17 00:35:06.544727 kernel: raid6: avx2x4 xor() 6528 MB/s May 17 00:35:06.562711 kernel: raid6: avx2x2 gen() 18062 MB/s May 17 00:35:06.580701 kernel: raid6: avx2x2 xor() 17986 MB/s May 17 00:35:06.598698 kernel: raid6: avx2x1 gen() 13787 MB/s May 17 00:35:06.616695 kernel: raid6: avx2x1 xor() 15848 MB/s May 17 00:35:06.634689 kernel: raid6: sse2x4 gen() 9563 MB/s May 17 00:35:06.652690 kernel: raid6: sse2x4 xor() 5365 MB/s May 17 00:35:06.670702 kernel: raid6: sse2x2 gen() 9725 MB/s May 17 00:35:06.688690 kernel: raid6: sse2x2 xor() 6273 MB/s May 17 00:35:06.706692 kernel: raid6: sse2x1 gen() 9442 MB/s May 17 00:35:06.724794 kernel: raid6: sse2x1 xor() 4878 MB/s May 17 00:35:06.724856 kernel: raid6: using algorithm avx2x2 gen() 18062 MB/s May 17 00:35:06.724889 kernel: raid6: .... xor() 17986 MB/s, rmw enabled May 17 00:35:06.725872 kernel: raid6: using avx512x2 recovery algorithm May 17 00:35:06.740671 kernel: xor: automatically using best checksumming function avx May 17 00:35:06.844674 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no May 17 00:35:06.853957 systemd[1]: Finished dracut-pre-udev.service. May 17 00:35:06.852000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.853000 audit: BPF prog-id=7 op=LOAD May 17 00:35:06.853000 audit: BPF prog-id=8 op=LOAD May 17 00:35:06.855609 systemd[1]: Starting systemd-udevd.service... May 17 00:35:06.869505 systemd-udevd[385]: Using default interface naming scheme 'v252'. May 17 00:35:06.874000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.874942 systemd[1]: Started systemd-udevd.service. May 17 00:35:06.876892 systemd[1]: Starting dracut-pre-trigger.service... May 17 00:35:06.895030 dracut-pre-trigger[391]: rd.md=0: removing MD RAID activation May 17 00:35:06.925000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.927119 systemd[1]: Finished dracut-pre-trigger.service. May 17 00:35:06.928479 systemd[1]: Starting systemd-udev-trigger.service... May 17 00:35:06.972000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:06.973011 systemd[1]: Finished systemd-udev-trigger.service. May 17 00:35:07.036663 kernel: cryptd: max_cpu_qlen set to 1000 May 17 00:35:07.052944 kernel: ena 0000:00:05.0: ENA device version: 0.10 May 17 00:35:07.075182 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 May 17 00:35:07.075353 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. May 17 00:35:07.075487 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:03:0f:6b:5d:e9 May 17 00:35:07.078214 (udev-worker)[428]: Network interface NamePolicy= disabled on kernel command line. May 17 00:35:07.082755 kernel: AVX2 version of gcm_enc/dec engaged. May 17 00:35:07.082791 kernel: AES CTR mode by8 optimization enabled May 17 00:35:07.091754 kernel: nvme nvme0: pci function 0000:00:04.0 May 17 00:35:07.092027 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 May 17 00:35:07.104668 kernel: nvme nvme0: 2/0/0 default/read/poll queues May 17 00:35:07.113678 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 17 00:35:07.113777 kernel: GPT:9289727 != 16777215 May 17 00:35:07.113795 kernel: GPT:Alternate GPT header not at the end of the disk. May 17 00:35:07.113967 kernel: GPT:9289727 != 16777215 May 17 00:35:07.117099 kernel: GPT: Use GNU Parted to correct GPT errors. May 17 00:35:07.117177 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 17 00:35:07.184717 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 scanned by (udev-worker) (430) May 17 00:35:07.205979 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. May 17 00:35:07.271231 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 17 00:35:07.276933 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. May 17 00:35:07.282014 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. May 17 00:35:07.282791 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. May 17 00:35:07.286106 systemd[1]: Starting disk-uuid.service... May 17 00:35:07.293144 disk-uuid[589]: Primary Header is updated. May 17 00:35:07.293144 disk-uuid[589]: Secondary Entries is updated. May 17 00:35:07.293144 disk-uuid[589]: Secondary Header is updated. May 17 00:35:07.301660 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 17 00:35:08.315791 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 17 00:35:08.315860 disk-uuid[590]: The operation has completed successfully. May 17 00:35:08.447646 systemd[1]: disk-uuid.service: Deactivated successfully. May 17 00:35:08.447773 systemd[1]: Finished disk-uuid.service. May 17 00:35:08.447000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.447000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.458800 systemd[1]: Starting verity-setup.service... May 17 00:35:08.487230 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" May 17 00:35:08.595493 systemd[1]: Found device dev-mapper-usr.device. May 17 00:35:08.597689 systemd[1]: Mounting sysusr-usr.mount... May 17 00:35:08.604000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.600401 systemd[1]: Finished verity-setup.service. May 17 00:35:08.703658 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. May 17 00:35:08.704895 systemd[1]: Mounted sysusr-usr.mount. May 17 00:35:08.705929 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. May 17 00:35:08.707151 systemd[1]: Starting ignition-setup.service... May 17 00:35:08.712632 systemd[1]: Starting parse-ip-for-networkd.service... May 17 00:35:08.736027 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm May 17 00:35:08.736105 kernel: BTRFS info (device nvme0n1p6): using free space tree May 17 00:35:08.736124 kernel: BTRFS info (device nvme0n1p6): has skinny extents May 17 00:35:08.756691 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations May 17 00:35:08.770038 systemd[1]: mnt-oem.mount: Deactivated successfully. May 17 00:35:08.782280 systemd[1]: Finished ignition-setup.service. May 17 00:35:08.781000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.784186 systemd[1]: Starting ignition-fetch-offline.service... May 17 00:35:08.804407 systemd[1]: Finished parse-ip-for-networkd.service. May 17 00:35:08.803000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.804000 audit: BPF prog-id=9 op=LOAD May 17 00:35:08.807378 systemd[1]: Starting systemd-networkd.service... May 17 00:35:08.832269 systemd-networkd[1103]: lo: Link UP May 17 00:35:08.832685 systemd-networkd[1103]: lo: Gained carrier May 17 00:35:08.833761 systemd-networkd[1103]: Enumeration completed May 17 00:35:08.834000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.833890 systemd[1]: Started systemd-networkd.service. May 17 00:35:08.834439 systemd-networkd[1103]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 17 00:35:08.836444 systemd[1]: Reached target network.target. May 17 00:35:08.838817 systemd[1]: Starting iscsiuio.service... May 17 00:35:08.840028 systemd-networkd[1103]: eth0: Link UP May 17 00:35:08.840034 systemd-networkd[1103]: eth0: Gained carrier May 17 00:35:08.846000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.846758 systemd[1]: Started iscsiuio.service. May 17 00:35:08.849067 systemd[1]: Starting iscsid.service... May 17 00:35:08.854347 iscsid[1108]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi May 17 00:35:08.854347 iscsid[1108]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. May 17 00:35:08.854347 iscsid[1108]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. May 17 00:35:08.854347 iscsid[1108]: If using hardware iscsi like qla4xxx this message can be ignored. May 17 00:35:08.854347 iscsid[1108]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi May 17 00:35:08.856000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.867279 iscsid[1108]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf May 17 00:35:08.856399 systemd[1]: Started iscsid.service. May 17 00:35:08.859105 systemd[1]: Starting dracut-initqueue.service... May 17 00:35:08.867322 systemd-networkd[1103]: eth0: DHCPv4 address 172.31.23.93/20, gateway 172.31.16.1 acquired from 172.31.16.1 May 17 00:35:08.878551 systemd[1]: Finished dracut-initqueue.service. May 17 00:35:08.877000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:08.879403 systemd[1]: Reached target remote-fs-pre.target. May 17 00:35:08.880809 systemd[1]: Reached target remote-cryptsetup.target. May 17 00:35:08.882117 systemd[1]: Reached target remote-fs.target. May 17 00:35:08.884729 systemd[1]: Starting dracut-pre-mount.service... May 17 00:35:08.894517 systemd[1]: Finished dracut-pre-mount.service. May 17 00:35:08.893000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.387507 ignition[1081]: Ignition 2.14.0 May 17 00:35:09.387529 ignition[1081]: Stage: fetch-offline May 17 00:35:09.387723 ignition[1081]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:35:09.387770 ignition[1081]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b May 17 00:35:09.406971 ignition[1081]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 17 00:35:09.407375 ignition[1081]: Ignition finished successfully May 17 00:35:09.408646 systemd[1]: Finished ignition-fetch-offline.service. May 17 00:35:09.407000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.410265 systemd[1]: Starting ignition-fetch.service... May 17 00:35:09.419923 ignition[1127]: Ignition 2.14.0 May 17 00:35:09.419936 ignition[1127]: Stage: fetch May 17 00:35:09.420151 ignition[1127]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:35:09.420186 ignition[1127]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b May 17 00:35:09.428964 ignition[1127]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 17 00:35:09.429940 ignition[1127]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 17 00:35:09.469532 ignition[1127]: INFO : PUT result: OK May 17 00:35:09.473698 ignition[1127]: DEBUG : parsed url from cmdline: "" May 17 00:35:09.474451 ignition[1127]: INFO : no config URL provided May 17 00:35:09.474451 ignition[1127]: INFO : reading system config file "/usr/lib/ignition/user.ign" May 17 00:35:09.475798 ignition[1127]: INFO : no config at "/usr/lib/ignition/user.ign" May 17 00:35:09.475798 ignition[1127]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 17 00:35:09.475798 ignition[1127]: INFO : PUT result: OK May 17 00:35:09.475798 ignition[1127]: INFO : GET http://169.254.169.254/2019-10-01/user-data: attempt #1 May 17 00:35:09.478847 ignition[1127]: INFO : GET result: OK May 17 00:35:09.478847 ignition[1127]: DEBUG : parsing config with SHA512: 5285ee215bca712239e98c6ff05c0ad54183f900f70ac235518ab1ee5b9dd742cf2ddda8c3b893af1b438499c7df00122ac2a8ba6b52df6344c2a61fe66b0da5 May 17 00:35:09.478403 unknown[1127]: fetched base config from "system" May 17 00:35:09.478411 unknown[1127]: fetched base config from "system" May 17 00:35:09.480720 ignition[1127]: fetch: fetch complete May 17 00:35:09.478416 unknown[1127]: fetched user config from "aws" May 17 00:35:09.480726 ignition[1127]: fetch: fetch passed May 17 00:35:09.482286 systemd[1]: Finished ignition-fetch.service. May 17 00:35:09.481000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.480896 ignition[1127]: Ignition finished successfully May 17 00:35:09.483793 systemd[1]: Starting ignition-kargs.service... May 17 00:35:09.494554 ignition[1133]: Ignition 2.14.0 May 17 00:35:09.494570 ignition[1133]: Stage: kargs May 17 00:35:09.494811 ignition[1133]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:35:09.494848 ignition[1133]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b May 17 00:35:09.502448 ignition[1133]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 17 00:35:09.503492 ignition[1133]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 17 00:35:09.504279 ignition[1133]: INFO : PUT result: OK May 17 00:35:09.507590 ignition[1133]: kargs: kargs passed May 17 00:35:09.507662 ignition[1133]: Ignition finished successfully May 17 00:35:09.508844 systemd[1]: Finished ignition-kargs.service. May 17 00:35:09.508000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.510911 systemd[1]: Starting ignition-disks.service... May 17 00:35:09.520508 ignition[1139]: Ignition 2.14.0 May 17 00:35:09.520577 ignition[1139]: Stage: disks May 17 00:35:09.520826 ignition[1139]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:35:09.520859 ignition[1139]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b May 17 00:35:09.528467 ignition[1139]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 17 00:35:09.529665 ignition[1139]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 17 00:35:09.530847 ignition[1139]: INFO : PUT result: OK May 17 00:35:09.533610 ignition[1139]: disks: disks passed May 17 00:35:09.533711 ignition[1139]: Ignition finished successfully May 17 00:35:09.535383 systemd[1]: Finished ignition-disks.service. May 17 00:35:09.534000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.536310 systemd[1]: Reached target initrd-root-device.target. May 17 00:35:09.537409 systemd[1]: Reached target local-fs-pre.target. May 17 00:35:09.538384 systemd[1]: Reached target local-fs.target. May 17 00:35:09.539351 systemd[1]: Reached target sysinit.target. May 17 00:35:09.540294 systemd[1]: Reached target basic.target. May 17 00:35:09.542702 systemd[1]: Starting systemd-fsck-root.service... May 17 00:35:09.580841 systemd-fsck[1147]: ROOT: clean, 619/553520 files, 56023/553472 blocks May 17 00:35:09.584061 systemd[1]: Finished systemd-fsck-root.service. May 17 00:35:09.583000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.585910 systemd[1]: Mounting sysroot.mount... May 17 00:35:09.603665 kernel: EXT4-fs (nvme0n1p9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. May 17 00:35:09.604831 systemd[1]: Mounted sysroot.mount. May 17 00:35:09.606156 systemd[1]: Reached target initrd-root-fs.target. May 17 00:35:09.616990 systemd[1]: Mounting sysroot-usr.mount... May 17 00:35:09.618946 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. May 17 00:35:09.619757 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 17 00:35:09.619787 systemd[1]: Reached target ignition-diskful.target. May 17 00:35:09.621787 systemd[1]: Mounted sysroot-usr.mount. May 17 00:35:09.625728 systemd[1]: Starting initrd-setup-root.service... May 17 00:35:09.642293 initrd-setup-root[1168]: cut: /sysroot/etc/passwd: No such file or directory May 17 00:35:09.647698 systemd[1]: Mounting sysroot-usr-share-oem.mount... May 17 00:35:09.658754 initrd-setup-root[1177]: cut: /sysroot/etc/group: No such file or directory May 17 00:35:09.668697 initrd-setup-root[1185]: cut: /sysroot/etc/shadow: No such file or directory May 17 00:35:09.670437 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by mount (1175) May 17 00:35:09.675222 initrd-setup-root[1193]: cut: /sysroot/etc/gshadow: No such file or directory May 17 00:35:09.679336 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm May 17 00:35:09.679379 kernel: BTRFS info (device nvme0n1p6): using free space tree May 17 00:35:09.679399 kernel: BTRFS info (device nvme0n1p6): has skinny extents May 17 00:35:09.688800 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations May 17 00:35:09.699126 systemd[1]: Mounted sysroot-usr-share-oem.mount. May 17 00:35:09.903167 systemd[1]: Finished initrd-setup-root.service. May 17 00:35:09.902000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.905474 systemd[1]: Starting ignition-mount.service... May 17 00:35:09.910442 systemd[1]: Starting sysroot-boot.service... May 17 00:35:09.917370 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. May 17 00:35:09.917506 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. May 17 00:35:09.935657 ignition[1229]: INFO : Ignition 2.14.0 May 17 00:35:09.937190 ignition[1229]: INFO : Stage: mount May 17 00:35:09.938677 ignition[1229]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:35:09.940805 ignition[1229]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b May 17 00:35:09.954376 ignition[1229]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 17 00:35:09.955892 ignition[1229]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 17 00:35:09.959407 ignition[1229]: INFO : PUT result: OK May 17 00:35:09.962437 systemd[1]: Finished sysroot-boot.service. May 17 00:35:09.961000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.966842 ignition[1229]: INFO : mount: mount passed May 17 00:35:09.967577 ignition[1229]: INFO : Ignition finished successfully May 17 00:35:09.969560 systemd[1]: Finished ignition-mount.service. May 17 00:35:09.968000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:09.971809 systemd[1]: Starting ignition-files.service... May 17 00:35:09.980794 systemd[1]: Mounting sysroot-usr-share-oem.mount... May 17 00:35:09.999694 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 scanned by mount (1239) May 17 00:35:10.003564 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm May 17 00:35:10.003656 kernel: BTRFS info (device nvme0n1p6): using free space tree May 17 00:35:10.003670 kernel: BTRFS info (device nvme0n1p6): has skinny extents May 17 00:35:10.042669 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations May 17 00:35:10.045740 systemd[1]: Mounted sysroot-usr-share-oem.mount. May 17 00:35:10.056714 ignition[1258]: INFO : Ignition 2.14.0 May 17 00:35:10.056714 ignition[1258]: INFO : Stage: files May 17 00:35:10.058218 ignition[1258]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:35:10.058218 ignition[1258]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b May 17 00:35:10.066159 ignition[1258]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 17 00:35:10.066985 ignition[1258]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 17 00:35:10.067711 ignition[1258]: INFO : PUT result: OK May 17 00:35:10.070737 ignition[1258]: DEBUG : files: compiled without relabeling support, skipping May 17 00:35:10.077218 ignition[1258]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 17 00:35:10.077218 ignition[1258]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 17 00:35:10.100960 ignition[1258]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 17 00:35:10.102277 ignition[1258]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 17 00:35:10.104435 unknown[1258]: wrote ssh authorized keys file for user: core May 17 00:35:10.105711 ignition[1258]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 17 00:35:10.107695 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/eks/bootstrap.sh" May 17 00:35:10.107695 ignition[1258]: INFO : oem config not found in "/usr/share/oem", looking on oem partition May 17 00:35:10.113514 ignition[1258]: INFO : op(1): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1050791613" May 17 00:35:10.113514 ignition[1258]: CRITICAL : op(1): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1050791613": device or resource busy May 17 00:35:10.113514 ignition[1258]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem1050791613", trying btrfs: device or resource busy May 17 00:35:10.113514 ignition[1258]: INFO : op(2): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1050791613" May 17 00:35:10.113514 ignition[1258]: INFO : op(2): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1050791613" May 17 00:35:10.129218 ignition[1258]: INFO : op(3): [started] unmounting "/mnt/oem1050791613" May 17 00:35:10.130210 ignition[1258]: INFO : op(3): [finished] unmounting "/mnt/oem1050791613" May 17 00:35:10.130210 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/eks/bootstrap.sh" May 17 00:35:10.130210 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" May 17 00:35:10.130210 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" May 17 00:35:10.130210 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/flatcar/update.conf" May 17 00:35:10.130210 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 17 00:35:10.130210 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" May 17 00:35:10.146230 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" May 17 00:35:10.146230 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/etc/amazon/ssm/amazon-ssm-agent.json" May 17 00:35:10.146230 ignition[1258]: INFO : oem config not found in "/usr/share/oem", looking on oem partition May 17 00:35:10.146230 ignition[1258]: INFO : op(4): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem456481901" May 17 00:35:10.146230 ignition[1258]: CRITICAL : op(4): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem456481901": device or resource busy May 17 00:35:10.146230 ignition[1258]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem456481901", trying btrfs: device or resource busy May 17 00:35:10.146230 ignition[1258]: INFO : op(5): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem456481901" May 17 00:35:10.146230 ignition[1258]: INFO : op(5): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem456481901" May 17 00:35:10.146230 ignition[1258]: INFO : op(6): [started] unmounting "/mnt/oem456481901" May 17 00:35:10.146230 ignition[1258]: INFO : op(6): [finished] unmounting "/mnt/oem456481901" May 17 00:35:10.146230 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/etc/amazon/ssm/amazon-ssm-agent.json" May 17 00:35:10.146230 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/amazon/ssm/seelog.xml" May 17 00:35:10.146230 ignition[1258]: INFO : oem config not found in "/usr/share/oem", looking on oem partition May 17 00:35:10.174460 ignition[1258]: INFO : op(7): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3089138759" May 17 00:35:10.174460 ignition[1258]: CRITICAL : op(7): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3089138759": device or resource busy May 17 00:35:10.174460 ignition[1258]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem3089138759", trying btrfs: device or resource busy May 17 00:35:10.174460 ignition[1258]: INFO : op(8): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3089138759" May 17 00:35:10.174460 ignition[1258]: INFO : op(8): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3089138759" May 17 00:35:10.174460 ignition[1258]: INFO : op(9): [started] unmounting "/mnt/oem3089138759" May 17 00:35:10.174460 ignition[1258]: INFO : op(9): [finished] unmounting "/mnt/oem3089138759" May 17 00:35:10.174460 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/amazon/ssm/seelog.xml" May 17 00:35:10.174460 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/systemd/system/nvidia.service" May 17 00:35:10.174460 ignition[1258]: INFO : oem config not found in "/usr/share/oem", looking on oem partition May 17 00:35:10.174460 ignition[1258]: INFO : op(a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2112213914" May 17 00:35:10.174460 ignition[1258]: CRITICAL : op(a): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2112213914": device or resource busy May 17 00:35:10.174460 ignition[1258]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem2112213914", trying btrfs: device or resource busy May 17 00:35:10.174460 ignition[1258]: INFO : op(b): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2112213914" May 17 00:35:10.174460 ignition[1258]: INFO : op(b): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2112213914" May 17 00:35:10.174460 ignition[1258]: INFO : op(c): [started] unmounting "/mnt/oem2112213914" May 17 00:35:10.174460 ignition[1258]: INFO : op(c): [finished] unmounting "/mnt/oem2112213914" May 17 00:35:10.174460 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/systemd/system/nvidia.service" May 17 00:35:10.174460 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" May 17 00:35:10.174460 ignition[1258]: INFO : GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 May 17 00:35:10.201467 systemd-networkd[1103]: eth0: Gained IPv6LL May 17 00:35:10.867090 ignition[1258]: INFO : GET result: OK May 17 00:35:11.238330 ignition[1258]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" May 17 00:35:11.238330 ignition[1258]: INFO : files: op(b): [started] processing unit "coreos-metadata-sshkeys@.service" May 17 00:35:11.238330 ignition[1258]: INFO : files: op(b): [finished] processing unit "coreos-metadata-sshkeys@.service" May 17 00:35:11.238330 ignition[1258]: INFO : files: op(c): [started] processing unit "amazon-ssm-agent.service" May 17 00:35:11.238330 ignition[1258]: INFO : files: op(c): op(d): [started] writing unit "amazon-ssm-agent.service" at "/sysroot/etc/systemd/system/amazon-ssm-agent.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(c): op(d): [finished] writing unit "amazon-ssm-agent.service" at "/sysroot/etc/systemd/system/amazon-ssm-agent.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(c): [finished] processing unit "amazon-ssm-agent.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(e): [started] processing unit "nvidia.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(e): [finished] processing unit "nvidia.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(f): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " May 17 00:35:11.246729 ignition[1258]: INFO : files: op(f): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " May 17 00:35:11.246729 ignition[1258]: INFO : files: op(10): [started] setting preset to enabled for "amazon-ssm-agent.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(10): [finished] setting preset to enabled for "amazon-ssm-agent.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(11): [started] setting preset to enabled for "nvidia.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: op(11): [finished] setting preset to enabled for "nvidia.service" May 17 00:35:11.246729 ignition[1258]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" May 17 00:35:11.246729 ignition[1258]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" May 17 00:35:11.246729 ignition[1258]: INFO : files: files passed May 17 00:35:11.246729 ignition[1258]: INFO : Ignition finished successfully May 17 00:35:11.298338 kernel: kauditd_printk_skb: 26 callbacks suppressed May 17 00:35:11.298373 kernel: audit: type=1130 audit(1747442111.252:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.298396 kernel: audit: type=1130 audit(1747442111.276:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.298414 kernel: audit: type=1131 audit(1747442111.276:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.298432 kernel: audit: type=1130 audit(1747442111.288:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.252000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.276000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.276000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.288000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.251719 systemd[1]: Finished ignition-files.service. May 17 00:35:11.257258 systemd[1]: Starting initrd-setup-root-after-ignition.service... May 17 00:35:11.302704 initrd-setup-root-after-ignition[1283]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 17 00:35:11.264332 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). May 17 00:35:11.265903 systemd[1]: Starting ignition-quench.service... May 17 00:35:11.276777 systemd[1]: ignition-quench.service: Deactivated successfully. May 17 00:35:11.276920 systemd[1]: Finished ignition-quench.service. May 17 00:35:11.278797 systemd[1]: Finished initrd-setup-root-after-ignition.service. May 17 00:35:11.290864 systemd[1]: Reached target ignition-complete.target. May 17 00:35:11.298590 systemd[1]: Starting initrd-parse-etc.service... May 17 00:35:11.320727 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 17 00:35:11.320870 systemd[1]: Finished initrd-parse-etc.service. May 17 00:35:11.332784 kernel: audit: type=1130 audit(1747442111.320:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.332825 kernel: audit: type=1131 audit(1747442111.320:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.320000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.320000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.322464 systemd[1]: Reached target initrd-fs.target. May 17 00:35:11.333527 systemd[1]: Reached target initrd.target. May 17 00:35:11.334930 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. May 17 00:35:11.336370 systemd[1]: Starting dracut-pre-pivot.service... May 17 00:35:11.349891 systemd[1]: Finished dracut-pre-pivot.service. May 17 00:35:11.358829 kernel: audit: type=1130 audit(1747442111.349:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.349000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.351956 systemd[1]: Starting initrd-cleanup.service... May 17 00:35:11.365887 systemd[1]: Stopped target nss-lookup.target. May 17 00:35:11.366857 systemd[1]: Stopped target remote-cryptsetup.target. May 17 00:35:11.368297 systemd[1]: Stopped target timers.target. May 17 00:35:11.369767 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 17 00:35:11.376616 kernel: audit: type=1131 audit(1747442111.369:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.369000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.369992 systemd[1]: Stopped dracut-pre-pivot.service. May 17 00:35:11.371455 systemd[1]: Stopped target initrd.target. May 17 00:35:11.377582 systemd[1]: Stopped target basic.target. May 17 00:35:11.378898 systemd[1]: Stopped target ignition-complete.target. May 17 00:35:11.380156 systemd[1]: Stopped target ignition-diskful.target. May 17 00:35:11.381527 systemd[1]: Stopped target initrd-root-device.target. May 17 00:35:11.382765 systemd[1]: Stopped target remote-fs.target. May 17 00:35:11.383922 systemd[1]: Stopped target remote-fs-pre.target. May 17 00:35:11.385239 systemd[1]: Stopped target sysinit.target. May 17 00:35:11.386432 systemd[1]: Stopped target local-fs.target. May 17 00:35:11.387580 systemd[1]: Stopped target local-fs-pre.target. May 17 00:35:11.388988 systemd[1]: Stopped target swap.target. May 17 00:35:11.396272 kernel: audit: type=1131 audit(1747442111.389:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.389000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.390085 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 17 00:35:11.390292 systemd[1]: Stopped dracut-pre-mount.service. May 17 00:35:11.403588 kernel: audit: type=1131 audit(1747442111.396:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.396000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.391488 systemd[1]: Stopped target cryptsetup.target. May 17 00:35:11.402000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.397200 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 17 00:35:11.404000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.397422 systemd[1]: Stopped dracut-initqueue.service. May 17 00:35:11.398661 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 17 00:35:11.398881 systemd[1]: Stopped initrd-setup-root-after-ignition.service. May 17 00:35:11.404669 systemd[1]: ignition-files.service: Deactivated successfully. May 17 00:35:11.404901 systemd[1]: Stopped ignition-files.service. May 17 00:35:11.407372 systemd[1]: Stopping ignition-mount.service... May 17 00:35:11.419111 systemd[1]: Stopping iscsiuio.service... May 17 00:35:11.420260 ignition[1296]: INFO : Ignition 2.14.0 May 17 00:35:11.420260 ignition[1296]: INFO : Stage: umount May 17 00:35:11.420260 ignition[1296]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:35:11.420260 ignition[1296]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b May 17 00:35:11.426964 systemd[1]: Stopping sysroot-boot.service... May 17 00:35:11.429396 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 17 00:35:11.432000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.434000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.429756 systemd[1]: Stopped systemd-udev-trigger.service. May 17 00:35:11.434603 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 17 00:35:11.434868 systemd[1]: Stopped dracut-pre-trigger.service. May 17 00:35:11.444000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.448030 ignition[1296]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 17 00:35:11.448030 ignition[1296]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 17 00:35:11.449000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.449000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.440077 systemd[1]: iscsiuio.service: Deactivated successfully. May 17 00:35:11.456090 ignition[1296]: INFO : PUT result: OK May 17 00:35:11.456090 ignition[1296]: INFO : umount: umount passed May 17 00:35:11.456090 ignition[1296]: INFO : Ignition finished successfully May 17 00:35:11.442862 systemd[1]: Stopped iscsiuio.service. May 17 00:35:11.458000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.447758 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 17 00:35:11.459000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.449073 systemd[1]: Finished initrd-cleanup.service. May 17 00:35:11.460000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.457763 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 17 00:35:11.462000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.458570 systemd[1]: ignition-mount.service: Deactivated successfully. May 17 00:35:11.458767 systemd[1]: Stopped ignition-mount.service. May 17 00:35:11.464000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.460166 systemd[1]: ignition-disks.service: Deactivated successfully. May 17 00:35:11.460228 systemd[1]: Stopped ignition-disks.service. May 17 00:35:11.461474 systemd[1]: ignition-kargs.service: Deactivated successfully. May 17 00:35:11.461531 systemd[1]: Stopped ignition-kargs.service. May 17 00:35:11.462717 systemd[1]: ignition-fetch.service: Deactivated successfully. May 17 00:35:11.462775 systemd[1]: Stopped ignition-fetch.service. May 17 00:35:11.463895 systemd[1]: Stopped target network.target. May 17 00:35:11.465105 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 17 00:35:11.465170 systemd[1]: Stopped ignition-fetch-offline.service. May 17 00:35:11.466327 systemd[1]: Stopped target paths.target. May 17 00:35:11.467456 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 17 00:35:11.470709 systemd[1]: Stopped systemd-ask-password-console.path. May 17 00:35:11.477000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.471812 systemd[1]: Stopped target slices.target. May 17 00:35:11.473008 systemd[1]: Stopped target sockets.target. May 17 00:35:11.474161 systemd[1]: iscsid.socket: Deactivated successfully. May 17 00:35:11.474206 systemd[1]: Closed iscsid.socket. May 17 00:35:11.476020 systemd[1]: iscsiuio.socket: Deactivated successfully. May 17 00:35:11.476078 systemd[1]: Closed iscsiuio.socket. May 17 00:35:11.477264 systemd[1]: ignition-setup.service: Deactivated successfully. May 17 00:35:11.477342 systemd[1]: Stopped ignition-setup.service. May 17 00:35:11.479111 systemd[1]: Stopping systemd-networkd.service... May 17 00:35:11.480310 systemd[1]: Stopping systemd-resolved.service... May 17 00:35:11.486000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.483717 systemd-networkd[1103]: eth0: DHCPv6 lease lost May 17 00:35:11.487000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.486754 systemd[1]: systemd-resolved.service: Deactivated successfully. May 17 00:35:11.486898 systemd[1]: Stopped systemd-resolved.service. May 17 00:35:11.490000 audit: BPF prog-id=6 op=UNLOAD May 17 00:35:11.492000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.488446 systemd[1]: systemd-networkd.service: Deactivated successfully. May 17 00:35:11.493000 audit: BPF prog-id=9 op=UNLOAD May 17 00:35:11.493000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.488753 systemd[1]: Stopped systemd-networkd.service. May 17 00:35:11.494000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.490088 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 17 00:35:11.490134 systemd[1]: Closed systemd-networkd.socket. May 17 00:35:11.492010 systemd[1]: Stopping network-cleanup.service... May 17 00:35:11.493484 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 17 00:35:11.493559 systemd[1]: Stopped parse-ip-for-networkd.service. May 17 00:35:11.507000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.494346 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 17 00:35:11.494394 systemd[1]: Stopped systemd-sysctl.service. May 17 00:35:11.495161 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 17 00:35:11.495217 systemd[1]: Stopped systemd-modules-load.service. May 17 00:35:11.511000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.496370 systemd[1]: Stopping systemd-udevd.service... May 17 00:35:11.501578 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 17 00:35:11.507726 systemd[1]: network-cleanup.service: Deactivated successfully. May 17 00:35:11.515000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.507885 systemd[1]: Stopped network-cleanup.service. May 17 00:35:11.516000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.511766 systemd[1]: systemd-udevd.service: Deactivated successfully. May 17 00:35:11.517000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.511954 systemd[1]: Stopped systemd-udevd.service. May 17 00:35:11.523000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.524000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.525000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.513370 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 17 00:35:11.513426 systemd[1]: Closed systemd-udevd-control.socket. May 17 00:35:11.514306 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 17 00:35:11.514357 systemd[1]: Closed systemd-udevd-kernel.socket. May 17 00:35:11.531000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.531000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.515409 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 17 00:35:11.515472 systemd[1]: Stopped dracut-pre-udev.service. May 17 00:35:11.517338 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 17 00:35:11.517397 systemd[1]: Stopped dracut-cmdline.service. May 17 00:35:11.518502 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 17 00:35:11.518556 systemd[1]: Stopped dracut-cmdline-ask.service. May 17 00:35:11.520943 systemd[1]: Starting initrd-udevadm-cleanup-db.service... May 17 00:35:11.523592 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 17 00:35:11.523686 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. May 17 00:35:11.524817 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 17 00:35:11.524875 systemd[1]: Stopped kmod-static-nodes.service. May 17 00:35:11.525828 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 17 00:35:11.525881 systemd[1]: Stopped systemd-vconsole-setup.service. May 17 00:35:11.529462 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 17 00:35:11.531762 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 17 00:35:11.531871 systemd[1]: Finished initrd-udevadm-cleanup-db.service. May 17 00:35:11.577447 systemd[1]: sysroot-boot.service: Deactivated successfully. May 17 00:35:11.576000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.577557 systemd[1]: Stopped sysroot-boot.service. May 17 00:35:11.578454 systemd[1]: Reached target initrd-switch-root.target. May 17 00:35:11.579000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:11.579511 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 17 00:35:11.579583 systemd[1]: Stopped initrd-setup-root.service. May 17 00:35:11.583922 systemd[1]: Starting initrd-switch-root.service... May 17 00:35:11.597371 systemd[1]: Switching root. May 17 00:35:11.620062 iscsid[1108]: iscsid shutting down. May 17 00:35:11.621369 systemd-journald[185]: Received SIGTERM from PID 1 (n/a). May 17 00:35:11.621453 systemd-journald[185]: Journal stopped May 17 00:35:17.259561 kernel: SELinux: Class mctp_socket not defined in policy. May 17 00:35:17.259664 kernel: SELinux: Class anon_inode not defined in policy. May 17 00:35:17.259684 kernel: SELinux: the above unknown classes and permissions will be allowed May 17 00:35:17.259701 kernel: SELinux: policy capability network_peer_controls=1 May 17 00:35:17.259718 kernel: SELinux: policy capability open_perms=1 May 17 00:35:17.259741 kernel: SELinux: policy capability extended_socket_class=1 May 17 00:35:17.259763 kernel: SELinux: policy capability always_check_network=0 May 17 00:35:17.259779 kernel: SELinux: policy capability cgroup_seclabel=1 May 17 00:35:17.259798 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 17 00:35:17.259814 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 17 00:35:17.259830 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 17 00:35:17.259848 systemd[1]: Successfully loaded SELinux policy in 91.274ms. May 17 00:35:17.259887 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 8.769ms. May 17 00:35:17.259905 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 17 00:35:17.259924 systemd[1]: Detected virtualization amazon. May 17 00:35:17.259942 systemd[1]: Detected architecture x86-64. May 17 00:35:17.259960 systemd[1]: Detected first boot. May 17 00:35:17.259982 systemd[1]: Initializing machine ID from VM UUID. May 17 00:35:17.259999 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). May 17 00:35:17.260018 systemd[1]: Populated /etc with preset unit settings. May 17 00:35:17.260036 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 17 00:35:17.260057 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 17 00:35:17.260077 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:35:17.260098 kernel: kauditd_printk_skb: 48 callbacks suppressed May 17 00:35:17.260114 kernel: audit: type=1334 audit(1747442116.960:88): prog-id=12 op=LOAD May 17 00:35:17.260132 kernel: audit: type=1334 audit(1747442116.960:89): prog-id=3 op=UNLOAD May 17 00:35:17.260153 kernel: audit: type=1334 audit(1747442116.962:90): prog-id=13 op=LOAD May 17 00:35:17.260169 kernel: audit: type=1334 audit(1747442116.967:91): prog-id=14 op=LOAD May 17 00:35:17.260185 kernel: audit: type=1334 audit(1747442116.967:92): prog-id=4 op=UNLOAD May 17 00:35:17.260202 kernel: audit: type=1334 audit(1747442116.967:93): prog-id=5 op=UNLOAD May 17 00:35:17.260218 kernel: audit: type=1131 audit(1747442116.969:94): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.260239 systemd[1]: iscsid.service: Deactivated successfully. May 17 00:35:17.260256 systemd[1]: Stopped iscsid.service. May 17 00:35:17.260273 kernel: audit: type=1131 audit(1747442116.986:95): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.260290 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 17 00:35:17.260307 systemd[1]: Stopped initrd-switch-root.service. May 17 00:35:17.260329 kernel: audit: type=1334 audit(1747442116.997:96): prog-id=12 op=UNLOAD May 17 00:35:17.260347 kernel: audit: type=1130 audit(1747442117.000:97): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.260384 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 17 00:35:17.260405 systemd[1]: Created slice system-addon\x2dconfig.slice. May 17 00:35:17.260426 systemd[1]: Created slice system-addon\x2drun.slice. May 17 00:35:17.260446 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. May 17 00:35:17.260468 systemd[1]: Created slice system-getty.slice. May 17 00:35:17.260557 systemd[1]: Created slice system-modprobe.slice. May 17 00:35:17.260578 systemd[1]: Created slice system-serial\x2dgetty.slice. May 17 00:35:17.260598 systemd[1]: Created slice system-system\x2dcloudinit.slice. May 17 00:35:17.260624 systemd[1]: Created slice system-systemd\x2dfsck.slice. May 17 00:35:17.260665 systemd[1]: Created slice user.slice. May 17 00:35:17.260686 systemd[1]: Started systemd-ask-password-console.path. May 17 00:35:17.260707 systemd[1]: Started systemd-ask-password-wall.path. May 17 00:35:17.260727 systemd[1]: Set up automount boot.automount. May 17 00:35:17.260747 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. May 17 00:35:17.260768 systemd[1]: Stopped target initrd-switch-root.target. May 17 00:35:17.260789 systemd[1]: Stopped target initrd-fs.target. May 17 00:35:17.260808 systemd[1]: Stopped target initrd-root-fs.target. May 17 00:35:17.260833 systemd[1]: Reached target integritysetup.target. May 17 00:35:17.260854 systemd[1]: Reached target remote-cryptsetup.target. May 17 00:35:17.260875 systemd[1]: Reached target remote-fs.target. May 17 00:35:17.260895 systemd[1]: Reached target slices.target. May 17 00:35:17.260915 systemd[1]: Reached target swap.target. May 17 00:35:17.260935 systemd[1]: Reached target torcx.target. May 17 00:35:17.260957 systemd[1]: Reached target veritysetup.target. May 17 00:35:17.260978 systemd[1]: Listening on systemd-coredump.socket. May 17 00:35:17.261000 systemd[1]: Listening on systemd-initctl.socket. May 17 00:35:17.261023 systemd[1]: Listening on systemd-networkd.socket. May 17 00:35:17.261043 systemd[1]: Listening on systemd-udevd-control.socket. May 17 00:35:17.261063 systemd[1]: Listening on systemd-udevd-kernel.socket. May 17 00:35:17.261084 systemd[1]: Listening on systemd-userdbd.socket. May 17 00:35:17.261104 systemd[1]: Mounting dev-hugepages.mount... May 17 00:35:17.261124 systemd[1]: Mounting dev-mqueue.mount... May 17 00:35:17.261144 systemd[1]: Mounting media.mount... May 17 00:35:17.261165 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:17.261186 systemd[1]: Mounting sys-kernel-debug.mount... May 17 00:35:17.261217 systemd[1]: Mounting sys-kernel-tracing.mount... May 17 00:35:17.261241 systemd[1]: Mounting tmp.mount... May 17 00:35:17.261262 systemd[1]: Starting flatcar-tmpfiles.service... May 17 00:35:17.261282 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:35:17.261300 systemd[1]: Starting kmod-static-nodes.service... May 17 00:35:17.261316 systemd[1]: Starting modprobe@configfs.service... May 17 00:35:17.261335 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:35:17.261353 systemd[1]: Starting modprobe@drm.service... May 17 00:35:17.261371 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:35:17.261391 systemd[1]: Starting modprobe@fuse.service... May 17 00:35:17.261409 systemd[1]: Starting modprobe@loop.service... May 17 00:35:17.261426 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 17 00:35:17.261444 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 17 00:35:17.261464 systemd[1]: Stopped systemd-fsck-root.service. May 17 00:35:17.261486 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 17 00:35:17.261507 systemd[1]: Stopped systemd-fsck-usr.service. May 17 00:35:17.261527 systemd[1]: Stopped systemd-journald.service. May 17 00:35:17.261547 systemd[1]: Starting systemd-journald.service... May 17 00:35:17.261566 systemd[1]: Starting systemd-modules-load.service... May 17 00:35:17.261585 systemd[1]: Starting systemd-network-generator.service... May 17 00:35:17.261606 systemd[1]: Starting systemd-remount-fs.service... May 17 00:35:17.261627 systemd[1]: Starting systemd-udev-trigger.service... May 17 00:35:17.261667 systemd[1]: verity-setup.service: Deactivated successfully. May 17 00:35:17.261691 systemd[1]: Stopped verity-setup.service. May 17 00:35:17.261718 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:17.261742 kernel: fuse: init (API version 7.34) May 17 00:35:17.261761 systemd[1]: Mounted dev-hugepages.mount. May 17 00:35:17.261781 kernel: loop: module loaded May 17 00:35:17.261801 systemd[1]: Mounted dev-mqueue.mount. May 17 00:35:17.261822 systemd[1]: Mounted media.mount. May 17 00:35:17.261844 systemd[1]: Mounted sys-kernel-debug.mount. May 17 00:35:17.261865 systemd[1]: Mounted sys-kernel-tracing.mount. May 17 00:35:17.261885 systemd[1]: Mounted tmp.mount. May 17 00:35:17.261908 systemd[1]: Finished kmod-static-nodes.service. May 17 00:35:17.261929 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 17 00:35:17.261956 systemd-journald[1404]: Journal started May 17 00:35:17.262035 systemd-journald[1404]: Runtime Journal (/run/log/journal/ec26a0d5e5b36a0451d17934fa6aaf3c) is 4.8M, max 38.3M, 33.5M free. May 17 00:35:12.356000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 May 17 00:35:12.542000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 May 17 00:35:12.542000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 May 17 00:35:12.542000 audit: BPF prog-id=10 op=LOAD May 17 00:35:12.542000 audit: BPF prog-id=10 op=UNLOAD May 17 00:35:12.542000 audit: BPF prog-id=11 op=LOAD May 17 00:35:12.542000 audit: BPF prog-id=11 op=UNLOAD May 17 00:35:12.833000 audit[1329]: AVC avc: denied { associate } for pid=1329 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" May 17 00:35:12.833000 audit[1329]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c00018a2d2 a1=c000194378 a2=c000196800 a3=32 items=0 ppid=1312 pid=1329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:35:12.833000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 May 17 00:35:12.836000 audit[1329]: AVC avc: denied { associate } for pid=1329 comm="torcx-generator" name="usr" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 May 17 00:35:12.836000 audit[1329]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c00018a3a9 a2=1ed a3=0 items=2 ppid=1312 pid=1329 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:35:12.836000 audit: CWD cwd="/" May 17 00:35:12.836000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:12.836000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:12.836000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 May 17 00:35:16.960000 audit: BPF prog-id=12 op=LOAD May 17 00:35:16.960000 audit: BPF prog-id=3 op=UNLOAD May 17 00:35:16.962000 audit: BPF prog-id=13 op=LOAD May 17 00:35:16.967000 audit: BPF prog-id=14 op=LOAD May 17 00:35:16.967000 audit: BPF prog-id=4 op=UNLOAD May 17 00:35:16.967000 audit: BPF prog-id=5 op=UNLOAD May 17 00:35:16.969000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:16.986000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:16.997000 audit: BPF prog-id=12 op=UNLOAD May 17 00:35:17.000000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.000000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.163000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.169000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.172000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.172000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.173000 audit: BPF prog-id=15 op=LOAD May 17 00:35:17.173000 audit: BPF prog-id=16 op=LOAD May 17 00:35:17.174000 audit: BPF prog-id=17 op=LOAD May 17 00:35:17.174000 audit: BPF prog-id=13 op=UNLOAD May 17 00:35:17.174000 audit: BPF prog-id=14 op=UNLOAD May 17 00:35:17.218000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.256000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 May 17 00:35:17.256000 audit[1404]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=3 a1=7fffa6934270 a2=4000 a3=7fffa693430c items=0 ppid=1 pid=1404 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:35:17.256000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" May 17 00:35:17.259000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:16.959496 systemd[1]: Queued start job for default target multi-user.target. May 17 00:35:17.272860 systemd[1]: Finished modprobe@configfs.service. May 17 00:35:17.272906 systemd[1]: Started systemd-journald.service. May 17 00:35:17.268000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.268000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:12.806834 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 17 00:35:16.959510 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device. May 17 00:35:12.815866 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json May 17 00:35:16.970900 systemd[1]: systemd-journald.service: Deactivated successfully. May 17 00:35:12.815890 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json May 17 00:35:12.815938 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" May 17 00:35:12.815957 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="skipped missing lower profile" missing profile=oem May 17 00:35:12.816009 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" May 17 00:35:12.816024 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= May 17 00:35:12.816240 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack May 17 00:35:12.816283 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json May 17 00:35:17.274000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:12.816296 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json May 17 00:35:17.276590 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:35:12.826206 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 May 17 00:35:17.276851 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:35:12.826265 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl May 17 00:35:12.826292 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.7: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.7 May 17 00:35:12.826309 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store May 17 00:35:12.826333 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.7: no such file or directory" path=/var/lib/torcx/store/3510.3.7 May 17 00:35:12.826347 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:12Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store May 17 00:35:16.377328 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:16Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl May 17 00:35:16.377633 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:16Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl May 17 00:35:16.377772 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:16Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl May 17 00:35:16.378090 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:16Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl May 17 00:35:16.378157 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:16Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= May 17 00:35:16.378224 /usr/lib/systemd/system-generators/torcx-generator[1329]: time="2025-05-17T00:35:16Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx May 17 00:35:17.278000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.278000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.280629 systemd[1]: modprobe@drm.service: Deactivated successfully. May 17 00:35:17.280954 systemd[1]: Finished modprobe@drm.service. May 17 00:35:17.281000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.281000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.283296 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:35:17.283494 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:35:17.283000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.283000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.285342 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 17 00:35:17.285517 systemd[1]: Finished modprobe@fuse.service. May 17 00:35:17.285000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.285000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.287284 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:35:17.287546 systemd[1]: Finished modprobe@loop.service. May 17 00:35:17.286000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.287000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.290000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.290929 systemd[1]: Finished systemd-modules-load.service. May 17 00:35:17.292000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.292804 systemd[1]: Finished systemd-network-generator.service. May 17 00:35:17.294199 systemd[1]: Finished systemd-remount-fs.service. May 17 00:35:17.293000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.297003 systemd[1]: Reached target network-pre.target. May 17 00:35:17.299691 systemd[1]: Mounting sys-fs-fuse-connections.mount... May 17 00:35:17.303467 systemd[1]: Mounting sys-kernel-config.mount... May 17 00:35:17.309128 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 17 00:35:17.314631 systemd[1]: Starting systemd-hwdb-update.service... May 17 00:35:17.318327 systemd[1]: Starting systemd-journal-flush.service... May 17 00:35:17.319412 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:35:17.321081 systemd[1]: Starting systemd-random-seed.service... May 17 00:35:17.323025 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:35:17.325812 systemd[1]: Starting systemd-sysctl.service... May 17 00:35:17.328451 systemd[1]: Mounted sys-fs-fuse-connections.mount. May 17 00:35:17.334676 systemd-journald[1404]: Time spent on flushing to /var/log/journal/ec26a0d5e5b36a0451d17934fa6aaf3c is 60.422ms for 1199 entries. May 17 00:35:17.334676 systemd-journald[1404]: System Journal (/var/log/journal/ec26a0d5e5b36a0451d17934fa6aaf3c) is 8.0M, max 195.6M, 187.6M free. May 17 00:35:17.413608 systemd-journald[1404]: Received client request to flush runtime journal. May 17 00:35:17.359000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.387000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.395000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.333597 systemd[1]: Mounted sys-kernel-config.mount. May 17 00:35:17.359885 systemd[1]: Finished systemd-random-seed.service. May 17 00:35:17.414000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.361044 systemd[1]: Reached target first-boot-complete.target. May 17 00:35:17.388012 systemd[1]: Finished systemd-sysctl.service. May 17 00:35:17.396233 systemd[1]: Finished flatcar-tmpfiles.service. May 17 00:35:17.398937 systemd[1]: Starting systemd-sysusers.service... May 17 00:35:17.414857 systemd[1]: Finished systemd-journal-flush.service. May 17 00:35:17.416515 systemd[1]: Finished systemd-udev-trigger.service. May 17 00:35:17.415000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.418700 systemd[1]: Starting systemd-udev-settle.service... May 17 00:35:17.431376 udevadm[1448]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. May 17 00:35:17.575000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.576366 systemd[1]: Finished systemd-sysusers.service. May 17 00:35:17.578152 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 17 00:35:17.710000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:17.711180 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 17 00:35:18.014880 systemd[1]: Finished systemd-hwdb-update.service. May 17 00:35:18.013000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.014000 audit: BPF prog-id=18 op=LOAD May 17 00:35:18.014000 audit: BPF prog-id=19 op=LOAD May 17 00:35:18.014000 audit: BPF prog-id=7 op=UNLOAD May 17 00:35:18.014000 audit: BPF prog-id=8 op=UNLOAD May 17 00:35:18.016919 systemd[1]: Starting systemd-udevd.service... May 17 00:35:18.035380 systemd-udevd[1451]: Using default interface naming scheme 'v252'. May 17 00:35:18.105000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.107000 audit: BPF prog-id=20 op=LOAD May 17 00:35:18.107014 systemd[1]: Started systemd-udevd.service. May 17 00:35:18.109450 systemd[1]: Starting systemd-networkd.service... May 17 00:35:18.137000 audit: BPF prog-id=21 op=LOAD May 17 00:35:18.137000 audit: BPF prog-id=22 op=LOAD May 17 00:35:18.137000 audit: BPF prog-id=23 op=LOAD May 17 00:35:18.139855 systemd[1]: Starting systemd-userdbd.service... May 17 00:35:18.141523 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. May 17 00:35:18.166629 (udev-worker)[1462]: Network interface NamePolicy= disabled on kernel command line. May 17 00:35:18.171000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.173117 systemd[1]: Started systemd-userdbd.service. May 17 00:35:18.222000 audit[1461]: AVC avc: denied { confidentiality } for pid=1461 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 May 17 00:35:18.239672 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 May 17 00:35:18.272662 kernel: ACPI: button: Power Button [PWRF] May 17 00:35:18.222000 audit[1461]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=555c267e1e50 a1=338ac a2=7fc2ac422bc5 a3=5 items=110 ppid=1451 pid=1461 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:35:18.222000 audit: CWD cwd="/" May 17 00:35:18.222000 audit: PATH item=0 name=(null) inode=29 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=1 name=(null) inode=14156 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=2 name=(null) inode=14156 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=3 name=(null) inode=14157 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=4 name=(null) inode=14156 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=5 name=(null) inode=14158 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=6 name=(null) inode=14156 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=7 name=(null) inode=14159 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=8 name=(null) inode=14159 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=9 name=(null) inode=14160 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=10 name=(null) inode=14159 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=11 name=(null) inode=14161 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.273000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.222000 audit: PATH item=12 name=(null) inode=14159 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=13 name=(null) inode=14162 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=14 name=(null) inode=14159 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=15 name=(null) inode=14163 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=16 name=(null) inode=14159 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=17 name=(null) inode=14164 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=18 name=(null) inode=14156 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=19 name=(null) inode=14165 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=20 name=(null) inode=14165 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=21 name=(null) inode=14166 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=22 name=(null) inode=14165 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=23 name=(null) inode=14167 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=24 name=(null) inode=14165 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=25 name=(null) inode=14168 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=26 name=(null) inode=14165 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=27 name=(null) inode=14169 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.274331 systemd-networkd[1459]: lo: Link UP May 17 00:35:18.274346 systemd-networkd[1459]: lo: Gained carrier May 17 00:35:18.222000 audit: PATH item=28 name=(null) inode=14165 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=29 name=(null) inode=14170 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=30 name=(null) inode=14156 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=31 name=(null) inode=14171 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=32 name=(null) inode=14171 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=33 name=(null) inode=14172 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=34 name=(null) inode=14171 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=35 name=(null) inode=14173 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=36 name=(null) inode=14171 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=37 name=(null) inode=14174 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=38 name=(null) inode=14171 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=39 name=(null) inode=14175 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=40 name=(null) inode=14171 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=41 name=(null) inode=14176 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=42 name=(null) inode=14156 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=43 name=(null) inode=14177 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=44 name=(null) inode=14177 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=45 name=(null) inode=14178 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=46 name=(null) inode=14177 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=47 name=(null) inode=14179 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=48 name=(null) inode=14177 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=49 name=(null) inode=14180 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=50 name=(null) inode=14177 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=51 name=(null) inode=14181 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=52 name=(null) inode=14177 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=53 name=(null) inode=14182 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=54 name=(null) inode=29 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=55 name=(null) inode=14183 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=56 name=(null) inode=14183 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=57 name=(null) inode=14184 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=58 name=(null) inode=14183 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=59 name=(null) inode=14185 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=60 name=(null) inode=14183 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=61 name=(null) inode=14186 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=62 name=(null) inode=14186 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=63 name=(null) inode=14187 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=64 name=(null) inode=14186 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=65 name=(null) inode=14188 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=66 name=(null) inode=14186 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=67 name=(null) inode=14189 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=68 name=(null) inode=14186 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=69 name=(null) inode=14190 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=70 name=(null) inode=14186 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=71 name=(null) inode=14191 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=72 name=(null) inode=14183 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=73 name=(null) inode=14192 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=74 name=(null) inode=14192 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=75 name=(null) inode=14193 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=76 name=(null) inode=14192 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=77 name=(null) inode=14194 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=78 name=(null) inode=14192 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=79 name=(null) inode=14195 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=80 name=(null) inode=14192 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=81 name=(null) inode=14196 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=82 name=(null) inode=14192 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=83 name=(null) inode=14197 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=84 name=(null) inode=14183 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=85 name=(null) inode=14198 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=86 name=(null) inode=14198 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=87 name=(null) inode=14199 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=88 name=(null) inode=14198 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=89 name=(null) inode=14200 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=90 name=(null) inode=14198 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=91 name=(null) inode=14201 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=92 name=(null) inode=14198 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=93 name=(null) inode=14202 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=94 name=(null) inode=14198 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=95 name=(null) inode=14203 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=96 name=(null) inode=14183 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=97 name=(null) inode=14204 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=98 name=(null) inode=14204 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=99 name=(null) inode=14205 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=100 name=(null) inode=14204 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=101 name=(null) inode=14206 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=102 name=(null) inode=14204 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=103 name=(null) inode=14207 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=104 name=(null) inode=14204 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=105 name=(null) inode=14208 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=106 name=(null) inode=14204 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=107 name=(null) inode=14209 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PATH item=109 name=(null) inode=14210 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:35:18.222000 audit: PROCTITLE proctitle="(udev-worker)" May 17 00:35:18.274798 systemd-networkd[1459]: Enumeration completed May 17 00:35:18.274908 systemd[1]: Started systemd-networkd.service. May 17 00:35:18.276664 systemd[1]: Starting systemd-networkd-wait-online.service... May 17 00:35:18.279841 systemd-networkd[1459]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 17 00:35:18.288442 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 17 00:35:18.288053 systemd-networkd[1459]: eth0: Link UP May 17 00:35:18.288179 systemd-networkd[1459]: eth0: Gained carrier May 17 00:35:18.296710 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input3 May 17 00:35:18.296939 systemd-networkd[1459]: eth0: DHCPv4 address 172.31.23.93/20, gateway 172.31.16.1 acquired from 172.31.16.1 May 17 00:35:18.302678 kernel: ACPI: button: Sleep Button [SLPF] May 17 00:35:18.309694 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input4 May 17 00:35:18.313672 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr May 17 00:35:18.323659 kernel: mousedev: PS/2 mouse device common for all mice May 17 00:35:18.443519 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 17 00:35:18.446196 systemd[1]: Finished systemd-udev-settle.service. May 17 00:35:18.445000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.448419 systemd[1]: Starting lvm2-activation-early.service... May 17 00:35:18.526348 lvm[1566]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 17 00:35:18.553316 systemd[1]: Finished lvm2-activation-early.service. May 17 00:35:18.552000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.554113 systemd[1]: Reached target cryptsetup.target. May 17 00:35:18.556055 systemd[1]: Starting lvm2-activation.service... May 17 00:35:18.561707 lvm[1567]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 17 00:35:18.584157 systemd[1]: Finished lvm2-activation.service. May 17 00:35:18.583000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.585176 systemd[1]: Reached target local-fs-pre.target. May 17 00:35:18.585785 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 17 00:35:18.585826 systemd[1]: Reached target local-fs.target. May 17 00:35:18.586359 systemd[1]: Reached target machines.target. May 17 00:35:18.588267 systemd[1]: Starting ldconfig.service... May 17 00:35:18.590320 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:35:18.590388 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:35:18.591610 systemd[1]: Starting systemd-boot-update.service... May 17 00:35:18.593801 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... May 17 00:35:18.595964 systemd[1]: Starting systemd-machine-id-commit.service... May 17 00:35:18.598437 systemd[1]: Starting systemd-sysext.service... May 17 00:35:18.611049 systemd[1]: Unmounting usr-share-oem.mount... May 17 00:35:18.616322 systemd[1]: usr-share-oem.mount: Deactivated successfully. May 17 00:35:18.616763 systemd[1]: Unmounted usr-share-oem.mount. May 17 00:35:18.623290 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1569 (bootctl) May 17 00:35:18.624753 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... May 17 00:35:18.630670 kernel: loop0: detected capacity change from 0 to 224512 May 17 00:35:18.647407 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. May 17 00:35:18.646000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.793697 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 17 00:35:18.813592 systemd-fsck[1581]: fsck.fat 4.2 (2021-01-31) May 17 00:35:18.813592 systemd-fsck[1581]: /dev/nvme0n1p1: 790 files, 120726/258078 clusters May 17 00:35:18.814720 kernel: loop1: detected capacity change from 0 to 224512 May 17 00:35:18.817620 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. May 17 00:35:18.816000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.820876 systemd[1]: Mounting boot.mount... May 17 00:35:18.856839 systemd[1]: Mounted boot.mount. May 17 00:35:18.868174 (sd-sysext)[1584]: Using extensions 'kubernetes'. May 17 00:35:18.868792 (sd-sysext)[1584]: Merged extensions into '/usr'. May 17 00:35:18.887748 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 17 00:35:18.897208 systemd[1]: Finished systemd-machine-id-commit.service. May 17 00:35:18.897000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.917891 systemd[1]: Finished systemd-boot-update.service. May 17 00:35:18.917000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.919211 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:18.921535 systemd[1]: Mounting usr-share-oem.mount... May 17 00:35:18.923208 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:35:18.926101 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:35:18.929170 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:35:18.932523 systemd[1]: Starting modprobe@loop.service... May 17 00:35:18.933473 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:35:18.933699 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:35:18.933878 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:18.935372 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:35:18.935585 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:35:18.935000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.935000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.939449 systemd[1]: Mounted usr-share-oem.mount. May 17 00:35:18.940879 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:35:18.941053 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:35:18.940000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.940000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.942090 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:35:18.942253 systemd[1]: Finished modprobe@loop.service. May 17 00:35:18.941000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.941000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.943373 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:35:18.943530 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:35:18.944831 systemd[1]: Finished systemd-sysext.service. May 17 00:35:18.943000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:18.947038 systemd[1]: Starting ensure-sysext.service... May 17 00:35:18.949286 systemd[1]: Starting systemd-tmpfiles-setup.service... May 17 00:35:18.961356 systemd[1]: Reloading. May 17 00:35:18.986042 systemd-tmpfiles[1603]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. May 17 00:35:18.990404 systemd-tmpfiles[1603]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 17 00:35:18.995330 systemd-tmpfiles[1603]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 17 00:35:19.038204 /usr/lib/systemd/system-generators/torcx-generator[1622]: time="2025-05-17T00:35:19Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 17 00:35:19.038977 /usr/lib/systemd/system-generators/torcx-generator[1622]: time="2025-05-17T00:35:19Z" level=info msg="torcx already run" May 17 00:35:19.200197 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 17 00:35:19.200230 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 17 00:35:19.238536 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:35:19.316000 audit: BPF prog-id=24 op=LOAD May 17 00:35:19.316000 audit: BPF prog-id=21 op=UNLOAD May 17 00:35:19.316000 audit: BPF prog-id=25 op=LOAD May 17 00:35:19.316000 audit: BPF prog-id=26 op=LOAD May 17 00:35:19.316000 audit: BPF prog-id=22 op=UNLOAD May 17 00:35:19.316000 audit: BPF prog-id=23 op=UNLOAD May 17 00:35:19.318000 audit: BPF prog-id=27 op=LOAD May 17 00:35:19.318000 audit: BPF prog-id=15 op=UNLOAD May 17 00:35:19.318000 audit: BPF prog-id=28 op=LOAD May 17 00:35:19.318000 audit: BPF prog-id=29 op=LOAD May 17 00:35:19.318000 audit: BPF prog-id=16 op=UNLOAD May 17 00:35:19.318000 audit: BPF prog-id=17 op=UNLOAD May 17 00:35:19.320000 audit: BPF prog-id=30 op=LOAD May 17 00:35:19.320000 audit: BPF prog-id=31 op=LOAD May 17 00:35:19.320000 audit: BPF prog-id=18 op=UNLOAD May 17 00:35:19.320000 audit: BPF prog-id=19 op=UNLOAD May 17 00:35:19.321000 audit: BPF prog-id=32 op=LOAD May 17 00:35:19.321000 audit: BPF prog-id=20 op=UNLOAD May 17 00:35:19.327093 systemd[1]: Finished systemd-tmpfiles-setup.service. May 17 00:35:19.326000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.333395 systemd[1]: Starting audit-rules.service... May 17 00:35:19.335764 systemd[1]: Starting clean-ca-certificates.service... May 17 00:35:19.338389 systemd[1]: Starting systemd-journal-catalog-update.service... May 17 00:35:19.340000 audit: BPF prog-id=33 op=LOAD May 17 00:35:19.346000 audit: BPF prog-id=34 op=LOAD May 17 00:35:19.354000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.346027 systemd[1]: Starting systemd-resolved.service... May 17 00:35:19.349499 systemd[1]: Starting systemd-timesyncd.service... May 17 00:35:19.352044 systemd[1]: Starting systemd-update-utmp.service... May 17 00:35:19.354065 systemd[1]: Finished clean-ca-certificates.service. May 17 00:35:19.355008 systemd-networkd[1459]: eth0: Gained IPv6LL May 17 00:35:19.356128 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:35:19.366895 systemd[1]: Finished systemd-networkd-wait-online.service. May 17 00:35:19.367000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.369469 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:19.369898 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:35:19.375221 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:35:19.379295 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:35:19.383812 systemd[1]: Starting modprobe@loop.service... May 17 00:35:19.385390 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:35:19.385593 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:35:19.385813 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:35:19.385940 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:19.387772 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:35:19.387000 audit[1685]: SYSTEM_BOOT pid=1685 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' May 17 00:35:19.389092 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:35:19.388000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.388000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.392320 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:35:19.392538 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:35:19.391000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.391000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.394286 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:35:19.394728 systemd[1]: Finished modprobe@loop.service. May 17 00:35:19.394000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.394000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.401312 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:19.401783 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:35:19.404577 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:35:19.407321 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:35:19.412517 systemd[1]: Starting modprobe@loop.service... May 17 00:35:19.413538 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:35:19.413875 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:35:19.414161 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:35:19.414388 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:19.418810 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:35:19.419006 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:35:19.418000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.418000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.420418 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:35:19.420609 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:35:19.419000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.419000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.421896 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:35:19.423743 systemd[1]: Finished systemd-update-utmp.service. May 17 00:35:19.422000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.430171 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:19.430644 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:35:19.433595 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:35:19.436217 systemd[1]: Starting modprobe@drm.service... May 17 00:35:19.439622 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:35:19.440439 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:35:19.440679 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:35:19.440905 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:35:19.441062 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:35:19.443654 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:35:19.443848 systemd[1]: Finished modprobe@loop.service. May 17 00:35:19.442000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.443000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.445119 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:35:19.445321 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:35:19.445000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.445000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.447523 systemd[1]: modprobe@drm.service: Deactivated successfully. May 17 00:35:19.447921 systemd[1]: Finished modprobe@drm.service. May 17 00:35:19.446000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.446000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.449020 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:35:19.452553 systemd[1]: Finished ensure-sysext.service. May 17 00:35:19.451000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.456170 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:35:19.456354 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:35:19.455000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.455000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.457163 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:35:19.490794 systemd[1]: Finished systemd-journal-catalog-update.service. May 17 00:35:19.489000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:35:19.536906 systemd-resolved[1683]: Positive Trust Anchors: May 17 00:35:19.536929 systemd-resolved[1683]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 17 00:35:19.536985 systemd-resolved[1683]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 17 00:35:19.549839 systemd[1]: Started systemd-timesyncd.service. May 17 00:35:19.550997 augenrules[1708]: No rules May 17 00:35:19.550595 systemd[1]: Reached target time-set.target. May 17 00:35:19.551430 systemd[1]: Finished audit-rules.service. May 17 00:35:19.548000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 May 17 00:35:19.548000 audit[1708]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fff94d7fe90 a2=420 a3=0 items=0 ppid=1679 pid=1708 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:35:19.548000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 May 17 00:35:19.576287 systemd-resolved[1683]: Defaulting to hostname 'linux'. May 17 00:35:19.578660 systemd[1]: Started systemd-resolved.service. May 17 00:35:19.579138 systemd[1]: Reached target network.target. May 17 00:35:19.579463 systemd[1]: Reached target network-online.target. May 17 00:35:19.579806 systemd[1]: Reached target nss-lookup.target. May 17 00:35:21.002716 systemd-timesyncd[1684]: Contacted time server 24.229.44.105:123 (0.flatcar.pool.ntp.org). May 17 00:35:21.002916 systemd-timesyncd[1684]: Initial clock synchronization to Sat 2025-05-17 00:35:21.002374 UTC. May 17 00:35:21.002990 systemd-resolved[1683]: Clock change detected. Flushing caches. May 17 00:35:21.150317 ldconfig[1568]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 17 00:35:21.159066 systemd[1]: Finished ldconfig.service. May 17 00:35:21.161468 systemd[1]: Starting systemd-update-done.service... May 17 00:35:21.170388 systemd[1]: Finished systemd-update-done.service. May 17 00:35:21.170976 systemd[1]: Reached target sysinit.target. May 17 00:35:21.171482 systemd[1]: Started motdgen.path. May 17 00:35:21.171851 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. May 17 00:35:21.172359 systemd[1]: Started logrotate.timer. May 17 00:35:21.173712 systemd[1]: Started mdadm.timer. May 17 00:35:21.174120 systemd[1]: Started systemd-tmpfiles-clean.timer. May 17 00:35:21.174517 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 17 00:35:21.174573 systemd[1]: Reached target paths.target. May 17 00:35:21.174958 systemd[1]: Reached target timers.target. May 17 00:35:21.175734 systemd[1]: Listening on dbus.socket. May 17 00:35:21.177938 systemd[1]: Starting docker.socket... May 17 00:35:21.182403 systemd[1]: Listening on sshd.socket. May 17 00:35:21.182993 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:35:21.183612 systemd[1]: Listening on docker.socket. May 17 00:35:21.184112 systemd[1]: Reached target sockets.target. May 17 00:35:21.184538 systemd[1]: Reached target basic.target. May 17 00:35:21.184943 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. May 17 00:35:21.184981 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. May 17 00:35:21.186657 systemd[1]: Started amazon-ssm-agent.service. May 17 00:35:21.188786 systemd[1]: Starting containerd.service... May 17 00:35:21.191873 systemd[1]: Starting coreos-metadata-sshkeys@core.service... May 17 00:35:21.194169 systemd[1]: Starting dbus.service... May 17 00:35:21.195881 systemd[1]: Starting enable-oem-cloudinit.service... May 17 00:35:21.198553 systemd[1]: Starting extend-filesystems.service... May 17 00:35:21.200445 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). May 17 00:35:21.202839 systemd[1]: Starting kubelet.service... May 17 00:35:21.206500 systemd[1]: Starting motdgen.service... May 17 00:35:21.210666 systemd[1]: Started nvidia.service. May 17 00:35:21.213076 systemd[1]: Starting ssh-key-proc-cmdline.service... May 17 00:35:21.216372 systemd[1]: Starting sshd-keygen.service... May 17 00:35:21.225439 systemd[1]: Starting systemd-logind.service... May 17 00:35:21.226137 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:35:21.226218 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 17 00:35:21.227037 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 17 00:35:21.230507 systemd[1]: Starting update-engine.service... May 17 00:35:21.235446 systemd[1]: Starting update-ssh-keys-after-ignition.service... May 17 00:35:21.318124 jq[1721]: false May 17 00:35:21.355956 jq[1730]: true May 17 00:35:21.276224 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 17 00:35:21.276482 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. May 17 00:35:21.309082 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 17 00:35:21.309356 systemd[1]: Finished ssh-key-proc-cmdline.service. May 17 00:35:21.416320 extend-filesystems[1722]: Found loop1 May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1 May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1p1 May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1p2 May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1p3 May 17 00:35:21.417877 extend-filesystems[1722]: Found usr May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1p4 May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1p6 May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1p7 May 17 00:35:21.417877 extend-filesystems[1722]: Found nvme0n1p9 May 17 00:35:21.417877 extend-filesystems[1722]: Checking size of /dev/nvme0n1p9 May 17 00:35:21.459313 jq[1737]: true May 17 00:35:21.473895 systemd[1]: motdgen.service: Deactivated successfully. May 17 00:35:21.474147 systemd[1]: Finished motdgen.service. May 17 00:35:21.481505 dbus-daemon[1720]: [system] SELinux support is enabled May 17 00:35:21.481731 systemd[1]: Started dbus.service. May 17 00:35:21.485489 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 17 00:35:21.485526 systemd[1]: Reached target system-config.target. May 17 00:35:21.486123 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 17 00:35:21.486150 systemd[1]: Reached target user-config.target. May 17 00:35:21.494003 dbus-daemon[1720]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1459 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") May 17 00:35:21.499675 systemd[1]: Starting systemd-hostnamed.service... May 17 00:35:21.514170 extend-filesystems[1722]: Resized partition /dev/nvme0n1p9 May 17 00:35:21.523130 extend-filesystems[1769]: resize2fs 1.46.5 (30-Dec-2021) May 17 00:35:21.524138 amazon-ssm-agent[1717]: 2025/05/17 00:35:21 Failed to load instance info from vault. RegistrationKey does not exist. May 17 00:35:21.535599 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks May 17 00:35:21.543301 amazon-ssm-agent[1717]: Initializing new seelog logger May 17 00:35:21.543510 amazon-ssm-agent[1717]: New Seelog Logger Creation Complete May 17 00:35:21.543629 amazon-ssm-agent[1717]: 2025/05/17 00:35:21 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 17 00:35:21.543629 amazon-ssm-agent[1717]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 17 00:35:21.543946 amazon-ssm-agent[1717]: 2025/05/17 00:35:21 processing appconfig overrides May 17 00:35:21.652600 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 May 17 00:35:21.652720 update_engine[1729]: I0517 00:35:21.635514 1729 main.cc:92] Flatcar Update Engine starting May 17 00:35:21.653773 systemd[1]: Started update-engine.service. May 17 00:35:21.659258 update_engine[1729]: I0517 00:35:21.658311 1729 update_check_scheduler.cc:74] Next update check in 9m10s May 17 00:35:21.656937 systemd[1]: Started locksmithd.service. May 17 00:35:21.663782 extend-filesystems[1769]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required May 17 00:35:21.663782 extend-filesystems[1769]: old_desc_blocks = 1, new_desc_blocks = 1 May 17 00:35:21.663782 extend-filesystems[1769]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. May 17 00:35:21.660526 systemd[1]: extend-filesystems.service: Deactivated successfully. May 17 00:35:21.670670 bash[1791]: Updated "/home/core/.ssh/authorized_keys" May 17 00:35:21.670800 extend-filesystems[1722]: Resized filesystem in /dev/nvme0n1p9 May 17 00:35:21.660764 systemd[1]: Finished extend-filesystems.service. May 17 00:35:21.668381 systemd[1]: Finished update-ssh-keys-after-ignition.service. May 17 00:35:21.685236 env[1734]: time="2025-05-17T00:35:21.685166076Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 May 17 00:35:21.700783 systemd[1]: nvidia.service: Deactivated successfully. May 17 00:35:21.767083 systemd-logind[1728]: Watching system buttons on /dev/input/event1 (Power Button) May 17 00:35:21.767123 systemd-logind[1728]: Watching system buttons on /dev/input/event2 (Sleep Button) May 17 00:35:21.767147 systemd-logind[1728]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 17 00:35:21.769919 systemd-logind[1728]: New seat seat0. May 17 00:35:21.776650 systemd[1]: Started systemd-logind.service. May 17 00:35:21.803752 env[1734]: time="2025-05-17T00:35:21.803661846Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 May 17 00:35:21.804091 env[1734]: time="2025-05-17T00:35:21.804062342Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 May 17 00:35:21.806053 env[1734]: time="2025-05-17T00:35:21.806007204Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.182-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 May 17 00:35:21.806177 env[1734]: time="2025-05-17T00:35:21.806159634Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 May 17 00:35:21.806563 env[1734]: time="2025-05-17T00:35:21.806536362Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:35:21.807305 env[1734]: time="2025-05-17T00:35:21.807260482Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 May 17 00:35:21.807415 env[1734]: time="2025-05-17T00:35:21.807395945Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" May 17 00:35:21.807502 env[1734]: time="2025-05-17T00:35:21.807488226Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 May 17 00:35:21.807669 env[1734]: time="2025-05-17T00:35:21.807653370Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 May 17 00:35:21.808037 env[1734]: time="2025-05-17T00:35:21.808010909Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 May 17 00:35:21.813668 env[1734]: time="2025-05-17T00:35:21.813620719Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:35:21.814564 env[1734]: time="2025-05-17T00:35:21.814529940Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 May 17 00:35:21.814781 env[1734]: time="2025-05-17T00:35:21.814759839Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" May 17 00:35:21.815343 env[1734]: time="2025-05-17T00:35:21.815320748Z" level=info msg="metadata content store policy set" policy=shared May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825107271Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825181935Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825206263Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825269735Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825351701Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825379145Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825400658Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825421254Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825443167Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825463324Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825482442Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825503858Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825663036Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 May 17 00:35:21.827778 env[1734]: time="2025-05-17T00:35:21.825776215Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826229475Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826272865Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826338060Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826409796Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826432608Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826513416Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826531523Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826551339Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826569726Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826586488Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826606040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826625339Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826773553Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826794950Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828389 env[1734]: time="2025-05-17T00:35:21.826817067Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 May 17 00:35:21.828936 env[1734]: time="2025-05-17T00:35:21.826835469Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 May 17 00:35:21.828936 env[1734]: time="2025-05-17T00:35:21.826858880Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 May 17 00:35:21.828936 env[1734]: time="2025-05-17T00:35:21.826875990Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 May 17 00:35:21.828936 env[1734]: time="2025-05-17T00:35:21.826901404Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" May 17 00:35:21.828936 env[1734]: time="2025-05-17T00:35:21.826949766Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 May 17 00:35:21.829133 env[1734]: time="2025-05-17T00:35:21.827242485Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" May 17 00:35:21.829133 env[1734]: time="2025-05-17T00:35:21.827344229Z" level=info msg="Connect containerd service" May 17 00:35:21.829133 env[1734]: time="2025-05-17T00:35:21.827384287Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" May 17 00:35:21.834201 env[1734]: time="2025-05-17T00:35:21.829710445Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 17 00:35:21.834201 env[1734]: time="2025-05-17T00:35:21.830039791Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 17 00:35:21.834201 env[1734]: time="2025-05-17T00:35:21.830089593Z" level=info msg=serving... address=/run/containerd/containerd.sock May 17 00:35:21.834201 env[1734]: time="2025-05-17T00:35:21.830153468Z" level=info msg="containerd successfully booted in 0.161581s" May 17 00:35:21.830253 systemd[1]: Started containerd.service. May 17 00:35:21.846264 env[1734]: time="2025-05-17T00:35:21.846195515Z" level=info msg="Start subscribing containerd event" May 17 00:35:21.849995 env[1734]: time="2025-05-17T00:35:21.849949765Z" level=info msg="Start recovering state" May 17 00:35:21.850243 env[1734]: time="2025-05-17T00:35:21.850223500Z" level=info msg="Start event monitor" May 17 00:35:21.850679 env[1734]: time="2025-05-17T00:35:21.850656627Z" level=info msg="Start snapshots syncer" May 17 00:35:21.850795 env[1734]: time="2025-05-17T00:35:21.850781203Z" level=info msg="Start cni network conf syncer for default" May 17 00:35:21.850858 env[1734]: time="2025-05-17T00:35:21.850846178Z" level=info msg="Start streaming server" May 17 00:35:21.895036 dbus-daemon[1720]: [system] Successfully activated service 'org.freedesktop.hostname1' May 17 00:35:21.895228 systemd[1]: Started systemd-hostnamed.service. May 17 00:35:21.899698 dbus-daemon[1720]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.6' (uid=0 pid=1765 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") May 17 00:35:21.903881 systemd[1]: Starting polkit.service... May 17 00:35:21.939394 polkitd[1843]: Started polkitd version 121 May 17 00:35:21.971431 polkitd[1843]: Loading rules from directory /etc/polkit-1/rules.d May 17 00:35:21.979755 polkitd[1843]: Loading rules from directory /usr/share/polkit-1/rules.d May 17 00:35:21.989971 polkitd[1843]: Finished loading, compiling and executing 2 rules May 17 00:35:21.991610 dbus-daemon[1720]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' May 17 00:35:21.991816 systemd[1]: Started polkit.service. May 17 00:35:21.993313 polkitd[1843]: Acquired the name org.freedesktop.PolicyKit1 on the system bus May 17 00:35:22.017253 systemd-hostnamed[1765]: Hostname set to (transient) May 17 00:35:22.017385 systemd-resolved[1683]: System hostname changed to 'ip-172-31-23-93'. May 17 00:35:22.074120 coreos-metadata[1719]: May 17 00:35:22.074 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 May 17 00:35:22.081985 coreos-metadata[1719]: May 17 00:35:22.081 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-keys: Attempt #1 May 17 00:35:22.084509 coreos-metadata[1719]: May 17 00:35:22.084 INFO Fetch successful May 17 00:35:22.084509 coreos-metadata[1719]: May 17 00:35:22.084 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-keys/0/openssh-key: Attempt #1 May 17 00:35:22.085341 coreos-metadata[1719]: May 17 00:35:22.085 INFO Fetch successful May 17 00:35:22.090450 unknown[1719]: wrote ssh authorized keys file for user: core May 17 00:35:22.117895 update-ssh-keys[1879]: Updated "/home/core/.ssh/authorized_keys" May 17 00:35:22.118338 systemd[1]: Finished coreos-metadata-sshkeys@core.service. May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Create new startup processor May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [LongRunningPluginsManager] registered plugins: {} May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing bookkeeping folders May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO removing the completed state files May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing bookkeeping folders for long running plugins May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing replies folder for MDS reply requests that couldn't reach the service May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing healthcheck folders for long running plugins May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing locations for inventory plugin May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing default location for custom inventory May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing default location for file inventory May 17 00:35:22.216501 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Initializing default location for role inventory May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Init the cloudwatchlogs publisher May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:updateSsmAgent May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:configureDocker May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:configurePackage May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:downloadContent May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:runDocument May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:runPowerShellScript May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:runDockerAction May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:refreshAssociation May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform independent plugin aws:softwareInventory May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Successfully loaded platform dependent plugin aws:runShellScript May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO Starting Agent: amazon-ssm-agent - v2.3.1319.0 May 17 00:35:22.216987 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO OS: linux, Arch: amd64 May 17 00:35:22.226954 amazon-ssm-agent[1717]: datastore file /var/lib/amazon/ssm/i-0a9e69d024c1fced0/longrunningplugins/datastore/store doesn't exist - no long running plugins to execute May 17 00:35:22.234155 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] Starting document processing engine... May 17 00:35:22.337696 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] [EngineProcessor] Starting May 17 00:35:22.432338 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] [EngineProcessor] Initial processing May 17 00:35:22.527336 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] Starting message polling May 17 00:35:22.547843 locksmithd[1799]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 17 00:35:22.590072 sshd_keygen[1753]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 17 00:35:22.618386 systemd[1]: Finished sshd-keygen.service. May 17 00:35:22.621069 systemd[1]: Starting issuegen.service... May 17 00:35:22.622903 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] Starting send replies to MDS May 17 00:35:22.628507 systemd[1]: issuegen.service: Deactivated successfully. May 17 00:35:22.628748 systemd[1]: Finished issuegen.service. May 17 00:35:22.631888 systemd[1]: Starting systemd-user-sessions.service... May 17 00:35:22.642863 systemd[1]: Finished systemd-user-sessions.service. May 17 00:35:22.646173 systemd[1]: Started getty@tty1.service. May 17 00:35:22.648818 systemd[1]: Started serial-getty@ttyS0.service. May 17 00:35:22.649888 systemd[1]: Reached target getty.target. May 17 00:35:22.717767 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [instanceID=i-0a9e69d024c1fced0] Starting association polling May 17 00:35:22.812978 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] [Association] [EngineProcessor] Starting May 17 00:35:22.908190 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] [Association] Launching response handler May 17 00:35:23.003710 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] [Association] [EngineProcessor] Initial processing May 17 00:35:23.099561 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] [Association] Initializing association scheduling service May 17 00:35:23.195390 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessagingDeliveryService] [Association] Association scheduling service initialized May 17 00:35:23.291425 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] Starting session document processing engine... May 17 00:35:23.388002 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] [EngineProcessor] Starting May 17 00:35:23.484431 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] SSM Agent is trying to setup control channel for Session Manager module. May 17 00:35:23.581307 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] Setting up websocket for controlchannel for instance: i-0a9e69d024c1fced0, requestId: 744811df-52d6-46c4-9133-b0355f543a78 May 17 00:35:23.678432 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [OfflineService] Starting document processing engine... May 17 00:35:23.775468 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [OfflineService] [EngineProcessor] Starting May 17 00:35:23.871694 systemd[1]: Started kubelet.service. May 17 00:35:23.872848 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [OfflineService] [EngineProcessor] Initial processing May 17 00:35:23.873464 systemd[1]: Reached target multi-user.target. May 17 00:35:23.876002 systemd[1]: Starting systemd-update-utmp-runlevel.service... May 17 00:35:23.885540 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. May 17 00:35:23.885731 systemd[1]: Finished systemd-update-utmp-runlevel.service. May 17 00:35:23.886588 systemd[1]: Startup finished in 661ms (kernel) + 6.457s (initrd) + 10.446s (userspace) = 17.565s. May 17 00:35:23.970381 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [OfflineService] Starting message polling May 17 00:35:24.068158 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [OfflineService] Starting send replies to MDS May 17 00:35:24.165937 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [LongRunningPluginsManager] starting long running plugin manager May 17 00:35:24.264090 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [LongRunningPluginsManager] there aren't any long running plugin to execute May 17 00:35:24.362303 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [HealthCheck] HealthCheck reporting agent health. May 17 00:35:24.460809 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] listening reply. May 17 00:35:24.559522 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [LongRunningPluginsManager] There are no long running plugins currently getting executed - skipping their healthcheck May 17 00:35:24.658252 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [StartupProcessor] Executing startup processor tasks May 17 00:35:24.757374 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [StartupProcessor] Write to serial port: Amazon SSM Agent v2.3.1319.0 is running May 17 00:35:24.856661 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [StartupProcessor] Write to serial port: OsProductName: Flatcar Container Linux by Kinvolk May 17 00:35:24.956055 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [StartupProcessor] Write to serial port: OsVersion: 3510.3.7 May 17 00:35:25.017062 kubelet[1924]: E0517 00:35:25.016874 1924 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 17 00:35:25.020546 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 17 00:35:25.020747 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 17 00:35:25.021096 systemd[1]: kubelet.service: Consumed 1.174s CPU time. May 17 00:35:25.055691 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] Opening websocket connection to: wss://ssmmessages.us-west-2.amazonaws.com/v1/control-channel/i-0a9e69d024c1fced0?role=subscribe&stream=input May 17 00:35:25.155578 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] Successfully opened websocket connection to: wss://ssmmessages.us-west-2.amazonaws.com/v1/control-channel/i-0a9e69d024c1fced0?role=subscribe&stream=input May 17 00:35:25.255631 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] Starting receiving message from control channel May 17 00:35:25.355902 amazon-ssm-agent[1717]: 2025-05-17 00:35:22 INFO [MessageGatewayService] [EngineProcessor] Initial processing May 17 00:35:30.399454 systemd[1]: Created slice system-sshd.slice. May 17 00:35:30.401571 systemd[1]: Started sshd@0-172.31.23.93:22-139.178.68.195:42736.service. May 17 00:35:30.610646 sshd[1932]: Accepted publickey for core from 139.178.68.195 port 42736 ssh2: RSA SHA256:I5cGDzOOPhNK8a4J4SFPiuUQivu3TK8ocBzhX4AkN30 May 17 00:35:30.613627 sshd[1932]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:35:30.628725 systemd[1]: Created slice user-500.slice. May 17 00:35:30.630121 systemd[1]: Starting user-runtime-dir@500.service... May 17 00:35:30.633369 systemd-logind[1728]: New session 1 of user core. May 17 00:35:30.641733 systemd[1]: Finished user-runtime-dir@500.service. May 17 00:35:30.643402 systemd[1]: Starting user@500.service... May 17 00:35:30.648135 (systemd)[1935]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 17 00:35:30.752121 systemd[1935]: Queued start job for default target default.target. May 17 00:35:30.752740 systemd[1935]: Reached target paths.target. May 17 00:35:30.752765 systemd[1935]: Reached target sockets.target. May 17 00:35:30.752778 systemd[1935]: Reached target timers.target. May 17 00:35:30.752796 systemd[1935]: Reached target basic.target. May 17 00:35:30.752912 systemd[1]: Started user@500.service. May 17 00:35:30.754002 systemd[1]: Started session-1.scope. May 17 00:35:30.754986 systemd[1935]: Reached target default.target. May 17 00:35:30.755040 systemd[1935]: Startup finished in 99ms. May 17 00:35:30.904531 systemd[1]: Started sshd@1-172.31.23.93:22-139.178.68.195:42742.service. May 17 00:35:31.070746 sshd[1944]: Accepted publickey for core from 139.178.68.195 port 42742 ssh2: RSA SHA256:I5cGDzOOPhNK8a4J4SFPiuUQivu3TK8ocBzhX4AkN30 May 17 00:35:31.072216 sshd[1944]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:35:31.077775 systemd-logind[1728]: New session 2 of user core. May 17 00:35:31.078334 systemd[1]: Started session-2.scope. May 17 00:35:31.208901 sshd[1944]: pam_unix(sshd:session): session closed for user core May 17 00:35:31.212132 systemd[1]: sshd@1-172.31.23.93:22-139.178.68.195:42742.service: Deactivated successfully. May 17 00:35:31.212938 systemd[1]: session-2.scope: Deactivated successfully. May 17 00:35:31.213800 systemd-logind[1728]: Session 2 logged out. Waiting for processes to exit. May 17 00:35:31.214798 systemd-logind[1728]: Removed session 2. May 17 00:35:31.234254 systemd[1]: Started sshd@2-172.31.23.93:22-139.178.68.195:42744.service. May 17 00:35:31.392139 sshd[1950]: Accepted publickey for core from 139.178.68.195 port 42744 ssh2: RSA SHA256:I5cGDzOOPhNK8a4J4SFPiuUQivu3TK8ocBzhX4AkN30 May 17 00:35:31.393421 sshd[1950]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:35:31.399360 systemd-logind[1728]: New session 3 of user core. May 17 00:35:31.400043 systemd[1]: Started session-3.scope. May 17 00:35:31.519779 sshd[1950]: pam_unix(sshd:session): session closed for user core May 17 00:35:31.523444 systemd[1]: sshd@2-172.31.23.93:22-139.178.68.195:42744.service: Deactivated successfully. May 17 00:35:31.524398 systemd[1]: session-3.scope: Deactivated successfully. May 17 00:35:31.525062 systemd-logind[1728]: Session 3 logged out. Waiting for processes to exit. May 17 00:35:31.526129 systemd-logind[1728]: Removed session 3. May 17 00:35:31.544995 systemd[1]: Started sshd@3-172.31.23.93:22-139.178.68.195:42750.service. May 17 00:35:31.703855 sshd[1956]: Accepted publickey for core from 139.178.68.195 port 42750 ssh2: RSA SHA256:I5cGDzOOPhNK8a4J4SFPiuUQivu3TK8ocBzhX4AkN30 May 17 00:35:31.705802 sshd[1956]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:35:31.710503 systemd-logind[1728]: New session 4 of user core. May 17 00:35:31.710986 systemd[1]: Started session-4.scope. May 17 00:35:31.839438 sshd[1956]: pam_unix(sshd:session): session closed for user core May 17 00:35:31.842381 systemd[1]: sshd@3-172.31.23.93:22-139.178.68.195:42750.service: Deactivated successfully. May 17 00:35:31.843120 systemd[1]: session-4.scope: Deactivated successfully. May 17 00:35:31.843720 systemd-logind[1728]: Session 4 logged out. Waiting for processes to exit. May 17 00:35:31.844528 systemd-logind[1728]: Removed session 4. May 17 00:35:31.865906 systemd[1]: Started sshd@4-172.31.23.93:22-139.178.68.195:42756.service. May 17 00:35:32.027191 sshd[1962]: Accepted publickey for core from 139.178.68.195 port 42756 ssh2: RSA SHA256:I5cGDzOOPhNK8a4J4SFPiuUQivu3TK8ocBzhX4AkN30 May 17 00:35:32.029051 sshd[1962]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:35:32.035541 systemd[1]: Started session-5.scope. May 17 00:35:32.035990 systemd-logind[1728]: New session 5 of user core. May 17 00:35:32.189885 sudo[1965]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 17 00:35:32.190229 sudo[1965]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 17 00:35:32.206337 systemd[1]: Starting coreos-metadata.service... May 17 00:35:32.286504 coreos-metadata[1969]: May 17 00:35:32.286 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 May 17 00:35:32.287470 coreos-metadata[1969]: May 17 00:35:32.287 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/instance-id: Attempt #1 May 17 00:35:32.288382 coreos-metadata[1969]: May 17 00:35:32.288 INFO Fetch successful May 17 00:35:32.288382 coreos-metadata[1969]: May 17 00:35:32.288 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/instance-type: Attempt #1 May 17 00:35:32.289810 coreos-metadata[1969]: May 17 00:35:32.289 INFO Fetch successful May 17 00:35:32.289810 coreos-metadata[1969]: May 17 00:35:32.289 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/local-ipv4: Attempt #1 May 17 00:35:32.290709 coreos-metadata[1969]: May 17 00:35:32.290 INFO Fetch successful May 17 00:35:32.290879 coreos-metadata[1969]: May 17 00:35:32.290 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-ipv4: Attempt #1 May 17 00:35:32.291732 coreos-metadata[1969]: May 17 00:35:32.291 INFO Fetch successful May 17 00:35:32.291825 coreos-metadata[1969]: May 17 00:35:32.291 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/placement/availability-zone: Attempt #1 May 17 00:35:32.292808 coreos-metadata[1969]: May 17 00:35:32.292 INFO Fetch successful May 17 00:35:32.292922 coreos-metadata[1969]: May 17 00:35:32.292 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/hostname: Attempt #1 May 17 00:35:32.293874 coreos-metadata[1969]: May 17 00:35:32.293 INFO Fetch successful May 17 00:35:32.293874 coreos-metadata[1969]: May 17 00:35:32.293 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-hostname: Attempt #1 May 17 00:35:32.294815 coreos-metadata[1969]: May 17 00:35:32.294 INFO Fetch successful May 17 00:35:32.294887 coreos-metadata[1969]: May 17 00:35:32.294 INFO Fetching http://169.254.169.254/2019-10-01/dynamic/instance-identity/document: Attempt #1 May 17 00:35:32.295617 coreos-metadata[1969]: May 17 00:35:32.295 INFO Fetch successful May 17 00:35:32.304840 systemd[1]: Finished coreos-metadata.service. May 17 00:35:33.302413 systemd[1]: Stopped kubelet.service. May 17 00:35:33.303064 systemd[1]: kubelet.service: Consumed 1.174s CPU time. May 17 00:35:33.306313 systemd[1]: Starting kubelet.service... May 17 00:35:33.345139 systemd[1]: Reloading. May 17 00:35:33.464580 /usr/lib/systemd/system-generators/torcx-generator[2022]: time="2025-05-17T00:35:33Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 17 00:35:33.474417 /usr/lib/systemd/system-generators/torcx-generator[2022]: time="2025-05-17T00:35:33Z" level=info msg="torcx already run" May 17 00:35:33.601587 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 17 00:35:33.601615 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 17 00:35:33.621297 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:35:33.730246 systemd[1]: Started kubelet.service. May 17 00:35:33.733822 systemd[1]: Stopping kubelet.service... May 17 00:35:33.734481 systemd[1]: kubelet.service: Deactivated successfully. May 17 00:35:33.734714 systemd[1]: Stopped kubelet.service. May 17 00:35:33.736634 systemd[1]: Starting kubelet.service... May 17 00:35:33.962125 systemd[1]: Started kubelet.service. May 17 00:35:34.012754 kubelet[2085]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:35:34.012754 kubelet[2085]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 17 00:35:34.012754 kubelet[2085]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:35:34.012754 kubelet[2085]: I0517 00:35:34.012698 2085 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 17 00:35:34.584586 kubelet[2085]: I0517 00:35:34.584537 2085 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" May 17 00:35:34.584795 kubelet[2085]: I0517 00:35:34.584782 2085 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 17 00:35:34.585599 kubelet[2085]: I0517 00:35:34.585572 2085 server.go:954] "Client rotation is on, will bootstrap in background" May 17 00:35:34.630961 kubelet[2085]: I0517 00:35:34.630925 2085 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 17 00:35:34.656957 kubelet[2085]: E0517 00:35:34.656917 2085 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" May 17 00:35:34.656957 kubelet[2085]: I0517 00:35:34.656958 2085 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." May 17 00:35:34.660123 kubelet[2085]: I0517 00:35:34.660092 2085 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 17 00:35:34.660409 kubelet[2085]: I0517 00:35:34.660369 2085 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 17 00:35:34.660609 kubelet[2085]: I0517 00:35:34.660408 2085 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"172.31.23.93","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 17 00:35:34.662014 kubelet[2085]: I0517 00:35:34.661958 2085 topology_manager.go:138] "Creating topology manager with none policy" May 17 00:35:34.662014 kubelet[2085]: I0517 00:35:34.661997 2085 container_manager_linux.go:304] "Creating device plugin manager" May 17 00:35:34.662184 kubelet[2085]: I0517 00:35:34.662136 2085 state_mem.go:36] "Initialized new in-memory state store" May 17 00:35:34.668342 kubelet[2085]: I0517 00:35:34.668303 2085 kubelet.go:446] "Attempting to sync node with API server" May 17 00:35:34.668342 kubelet[2085]: I0517 00:35:34.668348 2085 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 17 00:35:34.668579 kubelet[2085]: I0517 00:35:34.668382 2085 kubelet.go:352] "Adding apiserver pod source" May 17 00:35:34.668579 kubelet[2085]: I0517 00:35:34.668396 2085 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 17 00:35:34.671524 kubelet[2085]: E0517 00:35:34.671486 2085 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:34.671692 kubelet[2085]: E0517 00:35:34.671544 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:34.680139 kubelet[2085]: I0517 00:35:34.680108 2085 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 17 00:35:34.680658 kubelet[2085]: I0517 00:35:34.680640 2085 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 17 00:35:34.680768 kubelet[2085]: W0517 00:35:34.680706 2085 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 17 00:35:34.683523 kubelet[2085]: I0517 00:35:34.683489 2085 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 17 00:35:34.683637 kubelet[2085]: I0517 00:35:34.683553 2085 server.go:1287] "Started kubelet" May 17 00:35:34.686718 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). May 17 00:35:34.686891 kubelet[2085]: I0517 00:35:34.686869 2085 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 17 00:35:34.699079 kubelet[2085]: W0517 00:35:34.699038 2085 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "172.31.23.93" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope May 17 00:35:34.699215 kubelet[2085]: E0517 00:35:34.699093 2085 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"172.31.23.93\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" May 17 00:35:34.699215 kubelet[2085]: W0517 00:35:34.699210 2085 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope May 17 00:35:34.699338 kubelet[2085]: E0517 00:35:34.699224 2085 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" May 17 00:35:34.699941 kubelet[2085]: I0517 00:35:34.699876 2085 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 17 00:35:34.707898 kubelet[2085]: I0517 00:35:34.707864 2085 server.go:479] "Adding debug handlers to kubelet server" May 17 00:35:34.710998 kubelet[2085]: I0517 00:35:34.700509 2085 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 17 00:35:34.712005 kubelet[2085]: E0517 00:35:34.700848 2085 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{172.31.23.93.1840295cf75c68e7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:172.31.23.93,UID:172.31.23.93,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:172.31.23.93,},FirstTimestamp:2025-05-17 00:35:34.683511015 +0000 UTC m=+0.716439233,LastTimestamp:2025-05-17 00:35:34.683511015 +0000 UTC m=+0.716439233,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.31.23.93,}" May 17 00:35:34.712262 kubelet[2085]: I0517 00:35:34.702051 2085 desired_state_of_world_populator.go:150] "Desired state populator starts to run" May 17 00:35:34.713990 kubelet[2085]: I0517 00:35:34.700177 2085 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 17 00:35:34.714391 kubelet[2085]: I0517 00:35:34.714374 2085 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 17 00:35:34.715307 kubelet[2085]: E0517 00:35:34.715071 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:34.715678 kubelet[2085]: I0517 00:35:34.702031 2085 volume_manager.go:297] "Starting Kubelet Volume Manager" May 17 00:35:34.716526 kubelet[2085]: I0517 00:35:34.716511 2085 reconciler.go:26] "Reconciler: start to sync state" May 17 00:35:34.717228 kubelet[2085]: I0517 00:35:34.717205 2085 factory.go:221] Registration of the systemd container factory successfully May 17 00:35:34.717403 kubelet[2085]: I0517 00:35:34.717381 2085 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 17 00:35:34.720529 kubelet[2085]: E0517 00:35:34.720481 2085 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 17 00:35:34.726147 kubelet[2085]: I0517 00:35:34.726096 2085 factory.go:221] Registration of the containerd container factory successfully May 17 00:35:34.736808 kubelet[2085]: E0517 00:35:34.736779 2085 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"172.31.23.93\" not found" node="172.31.23.93" May 17 00:35:34.742666 kubelet[2085]: I0517 00:35:34.741209 2085 cpu_manager.go:221] "Starting CPU manager" policy="none" May 17 00:35:34.742666 kubelet[2085]: I0517 00:35:34.741228 2085 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 17 00:35:34.742666 kubelet[2085]: I0517 00:35:34.741251 2085 state_mem.go:36] "Initialized new in-memory state store" May 17 00:35:34.744685 kubelet[2085]: I0517 00:35:34.744661 2085 policy_none.go:49] "None policy: Start" May 17 00:35:34.744858 kubelet[2085]: I0517 00:35:34.744846 2085 memory_manager.go:186] "Starting memorymanager" policy="None" May 17 00:35:34.744931 kubelet[2085]: I0517 00:35:34.744922 2085 state_mem.go:35] "Initializing new in-memory state store" May 17 00:35:34.752990 systemd[1]: Created slice kubepods.slice. May 17 00:35:34.760166 systemd[1]: Created slice kubepods-burstable.slice. May 17 00:35:34.764622 systemd[1]: Created slice kubepods-besteffort.slice. May 17 00:35:34.774313 kubelet[2085]: I0517 00:35:34.774251 2085 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 17 00:35:34.774474 kubelet[2085]: I0517 00:35:34.774459 2085 eviction_manager.go:189] "Eviction manager: starting control loop" May 17 00:35:34.774526 kubelet[2085]: I0517 00:35:34.774477 2085 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 17 00:35:34.775873 kubelet[2085]: I0517 00:35:34.775544 2085 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 17 00:35:34.777444 kubelet[2085]: E0517 00:35:34.777419 2085 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 17 00:35:34.777566 kubelet[2085]: E0517 00:35:34.777467 2085 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"172.31.23.93\" not found" May 17 00:35:34.877006 kubelet[2085]: I0517 00:35:34.876001 2085 kubelet_node_status.go:75] "Attempting to register node" node="172.31.23.93" May 17 00:35:34.881386 kubelet[2085]: I0517 00:35:34.881071 2085 kubelet_node_status.go:78] "Successfully registered node" node="172.31.23.93" May 17 00:35:34.881386 kubelet[2085]: E0517 00:35:34.881389 2085 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"172.31.23.93\": node \"172.31.23.93\" not found" May 17 00:35:34.904362 kubelet[2085]: E0517 00:35:34.904329 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:34.909509 kubelet[2085]: I0517 00:35:34.909464 2085 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 17 00:35:34.911412 kubelet[2085]: I0517 00:35:34.911368 2085 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 17 00:35:34.911412 kubelet[2085]: I0517 00:35:34.911402 2085 status_manager.go:227] "Starting to sync pod status with apiserver" May 17 00:35:34.911570 kubelet[2085]: I0517 00:35:34.911428 2085 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 17 00:35:34.911570 kubelet[2085]: I0517 00:35:34.911437 2085 kubelet.go:2382] "Starting kubelet main sync loop" May 17 00:35:34.911570 kubelet[2085]: E0517 00:35:34.911494 2085 kubelet.go:2406] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" May 17 00:35:34.997660 sudo[1965]: pam_unix(sudo:session): session closed for user root May 17 00:35:35.005514 kubelet[2085]: E0517 00:35:35.005464 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.023412 sshd[1962]: pam_unix(sshd:session): session closed for user core May 17 00:35:35.026612 systemd[1]: sshd@4-172.31.23.93:22-139.178.68.195:42756.service: Deactivated successfully. May 17 00:35:35.027392 systemd[1]: session-5.scope: Deactivated successfully. May 17 00:35:35.028003 systemd-logind[1728]: Session 5 logged out. Waiting for processes to exit. May 17 00:35:35.029020 systemd-logind[1728]: Removed session 5. May 17 00:35:35.106453 kubelet[2085]: E0517 00:35:35.106390 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.206920 kubelet[2085]: E0517 00:35:35.206740 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.307786 kubelet[2085]: E0517 00:35:35.307707 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.408944 kubelet[2085]: E0517 00:35:35.408780 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.509871 kubelet[2085]: E0517 00:35:35.509729 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.589261 kubelet[2085]: I0517 00:35:35.588912 2085 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" May 17 00:35:35.589955 kubelet[2085]: W0517 00:35:35.589906 2085 reflector.go:492] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received May 17 00:35:35.590111 kubelet[2085]: W0517 00:35:35.590013 2085 reflector.go:492] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received May 17 00:35:35.610269 kubelet[2085]: E0517 00:35:35.610207 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.671760 kubelet[2085]: E0517 00:35:35.671697 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:35.710557 kubelet[2085]: E0517 00:35:35.710523 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.811148 kubelet[2085]: E0517 00:35:35.811000 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:35.912199 kubelet[2085]: E0517 00:35:35.912119 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:36.013401 kubelet[2085]: E0517 00:35:36.013351 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:36.114615 kubelet[2085]: E0517 00:35:36.114487 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:36.215052 kubelet[2085]: E0517 00:35:36.215000 2085 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"172.31.23.93\" not found" May 17 00:35:36.316234 kubelet[2085]: I0517 00:35:36.316202 2085 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" May 17 00:35:36.316788 env[1734]: time="2025-05-17T00:35:36.316748287Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 17 00:35:36.317507 kubelet[2085]: I0517 00:35:36.317462 2085 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" May 17 00:35:36.671166 kubelet[2085]: I0517 00:35:36.671114 2085 apiserver.go:52] "Watching apiserver" May 17 00:35:36.672301 kubelet[2085]: E0517 00:35:36.672249 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:36.679434 systemd[1]: Created slice kubepods-burstable-pod2a104fbd_8f68_4e88_91cf_90e3dd90a9f4.slice. May 17 00:35:36.689046 systemd[1]: Created slice kubepods-besteffort-pod140d453b_3e23_4b1e_85c6_4ec9be4a3b8a.slice. May 17 00:35:36.716862 kubelet[2085]: I0517 00:35:36.716827 2085 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" May 17 00:35:36.729058 kubelet[2085]: I0517 00:35:36.728986 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/140d453b-3e23-4b1e-85c6-4ec9be4a3b8a-lib-modules\") pod \"kube-proxy-7l9cb\" (UID: \"140d453b-3e23-4b1e-85c6-4ec9be4a3b8a\") " pod="kube-system/kube-proxy-7l9cb" May 17 00:35:36.729058 kubelet[2085]: I0517 00:35:36.729027 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-cgroup\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729058 kubelet[2085]: I0517 00:35:36.729045 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-etc-cni-netd\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729058 kubelet[2085]: I0517 00:35:36.729063 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-net\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729476 kubelet[2085]: I0517 00:35:36.729261 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/140d453b-3e23-4b1e-85c6-4ec9be4a3b8a-kube-proxy\") pod \"kube-proxy-7l9cb\" (UID: \"140d453b-3e23-4b1e-85c6-4ec9be4a3b8a\") " pod="kube-system/kube-proxy-7l9cb" May 17 00:35:36.729476 kubelet[2085]: I0517 00:35:36.729321 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q59dw\" (UniqueName: \"kubernetes.io/projected/140d453b-3e23-4b1e-85c6-4ec9be4a3b8a-kube-api-access-q59dw\") pod \"kube-proxy-7l9cb\" (UID: \"140d453b-3e23-4b1e-85c6-4ec9be4a3b8a\") " pod="kube-system/kube-proxy-7l9cb" May 17 00:35:36.729476 kubelet[2085]: I0517 00:35:36.729343 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-bpf-maps\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729476 kubelet[2085]: I0517 00:35:36.729373 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hostproc\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729476 kubelet[2085]: I0517 00:35:36.729391 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-clustermesh-secrets\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729609 kubelet[2085]: I0517 00:35:36.729408 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lzj6\" (UniqueName: \"kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-kube-api-access-4lzj6\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729609 kubelet[2085]: I0517 00:35:36.729423 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/140d453b-3e23-4b1e-85c6-4ec9be4a3b8a-xtables-lock\") pod \"kube-proxy-7l9cb\" (UID: \"140d453b-3e23-4b1e-85c6-4ec9be4a3b8a\") " pod="kube-system/kube-proxy-7l9cb" May 17 00:35:36.729799 kubelet[2085]: I0517 00:35:36.729440 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cni-path\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729844 kubelet[2085]: I0517 00:35:36.729813 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-lib-modules\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729874 kubelet[2085]: I0517 00:35:36.729853 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-xtables-lock\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729902 kubelet[2085]: I0517 00:35:36.729876 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-config-path\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729902 kubelet[2085]: I0517 00:35:36.729895 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hubble-tls\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729956 kubelet[2085]: I0517 00:35:36.729911 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-run\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.729984 kubelet[2085]: I0517 00:35:36.729931 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-kernel\") pod \"cilium-4wk8n\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " pod="kube-system/cilium-4wk8n" May 17 00:35:36.831473 kubelet[2085]: I0517 00:35:36.831437 2085 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" May 17 00:35:36.988821 env[1734]: time="2025-05-17T00:35:36.987844402Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4wk8n,Uid:2a104fbd-8f68-4e88-91cf-90e3dd90a9f4,Namespace:kube-system,Attempt:0,}" May 17 00:35:36.998031 env[1734]: time="2025-05-17T00:35:36.997978929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7l9cb,Uid:140d453b-3e23-4b1e-85c6-4ec9be4a3b8a,Namespace:kube-system,Attempt:0,}" May 17 00:35:37.495018 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3932317282.mount: Deactivated successfully. May 17 00:35:37.502976 env[1734]: time="2025-05-17T00:35:37.502914720Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.506282 env[1734]: time="2025-05-17T00:35:37.506230183Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.509006 env[1734]: time="2025-05-17T00:35:37.508961010Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.510676 env[1734]: time="2025-05-17T00:35:37.510638675Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.511959 env[1734]: time="2025-05-17T00:35:37.511819442Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.512661 env[1734]: time="2025-05-17T00:35:37.512629124Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.515196 env[1734]: time="2025-05-17T00:35:37.515160634Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.518092 env[1734]: time="2025-05-17T00:35:37.518049225Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:37.542105 env[1734]: time="2025-05-17T00:35:37.538668186Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:35:37.542105 env[1734]: time="2025-05-17T00:35:37.538721805Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:35:37.542105 env[1734]: time="2025-05-17T00:35:37.538733384Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:35:37.542105 env[1734]: time="2025-05-17T00:35:37.539022139Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/04e7a456f17eb03cb1694d157ccc57989fbfdc3fc513becbdb0da1bf33007c00 pid=2139 runtime=io.containerd.runc.v2 May 17 00:35:37.546653 env[1734]: time="2025-05-17T00:35:37.546321880Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:35:37.546939 env[1734]: time="2025-05-17T00:35:37.546889673Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:35:37.547074 env[1734]: time="2025-05-17T00:35:37.547041784Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:35:37.547875 env[1734]: time="2025-05-17T00:35:37.547667418Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7 pid=2157 runtime=io.containerd.runc.v2 May 17 00:35:37.567444 systemd[1]: Started cri-containerd-04e7a456f17eb03cb1694d157ccc57989fbfdc3fc513becbdb0da1bf33007c00.scope. May 17 00:35:37.569885 systemd[1]: Started cri-containerd-22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7.scope. May 17 00:35:37.609792 env[1734]: time="2025-05-17T00:35:37.609742633Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4wk8n,Uid:2a104fbd-8f68-4e88-91cf-90e3dd90a9f4,Namespace:kube-system,Attempt:0,} returns sandbox id \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\"" May 17 00:35:37.615659 env[1734]: time="2025-05-17T00:35:37.615609548Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 17 00:35:37.632321 env[1734]: time="2025-05-17T00:35:37.632250662Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7l9cb,Uid:140d453b-3e23-4b1e-85c6-4ec9be4a3b8a,Namespace:kube-system,Attempt:0,} returns sandbox id \"04e7a456f17eb03cb1694d157ccc57989fbfdc3fc513becbdb0da1bf33007c00\"" May 17 00:35:37.673562 kubelet[2085]: E0517 00:35:37.673504 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:38.674300 kubelet[2085]: E0517 00:35:38.674248 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:39.675347 kubelet[2085]: E0517 00:35:39.675176 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:40.676262 kubelet[2085]: E0517 00:35:40.676179 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:41.676582 kubelet[2085]: E0517 00:35:41.676501 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:42.676759 kubelet[2085]: E0517 00:35:42.676682 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:42.851631 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1983983781.mount: Deactivated successfully. May 17 00:35:43.678053 kubelet[2085]: E0517 00:35:43.677982 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:44.678652 kubelet[2085]: E0517 00:35:44.678571 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:45.679670 kubelet[2085]: E0517 00:35:45.679609 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:45.743080 env[1734]: time="2025-05-17T00:35:45.743023346Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:45.746633 env[1734]: time="2025-05-17T00:35:45.746588438Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:45.748885 env[1734]: time="2025-05-17T00:35:45.748848281Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:45.749986 env[1734]: time="2025-05-17T00:35:45.749530340Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" May 17 00:35:45.752411 env[1734]: time="2025-05-17T00:35:45.752352952Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.5\"" May 17 00:35:45.753743 env[1734]: time="2025-05-17T00:35:45.753703045Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:35:45.772002 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4036471395.mount: Deactivated successfully. May 17 00:35:45.780025 env[1734]: time="2025-05-17T00:35:45.779950593Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\"" May 17 00:35:45.782318 env[1734]: time="2025-05-17T00:35:45.782263552Z" level=info msg="StartContainer for \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\"" May 17 00:35:45.809832 systemd[1]: Started cri-containerd-abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c.scope. May 17 00:35:45.840682 env[1734]: time="2025-05-17T00:35:45.840639615Z" level=info msg="StartContainer for \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\" returns successfully" May 17 00:35:45.850788 systemd[1]: cri-containerd-abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c.scope: Deactivated successfully. May 17 00:35:46.029945 env[1734]: time="2025-05-17T00:35:46.029795016Z" level=info msg="shim disconnected" id=abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c May 17 00:35:46.029945 env[1734]: time="2025-05-17T00:35:46.029857995Z" level=warning msg="cleaning up after shim disconnected" id=abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c namespace=k8s.io May 17 00:35:46.029945 env[1734]: time="2025-05-17T00:35:46.029872376Z" level=info msg="cleaning up dead shim" May 17 00:35:46.041461 env[1734]: time="2025-05-17T00:35:46.041408923Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:35:46Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2266 runtime=io.containerd.runc.v2\n" May 17 00:35:46.392918 amazon-ssm-agent[1717]: 2025-05-17 00:35:46 INFO [MessagingDeliveryService] [Association] No associations on boot. Requerying for associations after 30 seconds. May 17 00:35:46.680241 kubelet[2085]: E0517 00:35:46.680109 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:46.770627 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c-rootfs.mount: Deactivated successfully. May 17 00:35:46.961157 env[1734]: time="2025-05-17T00:35:46.961022311Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 17 00:35:46.980057 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount228455563.mount: Deactivated successfully. May 17 00:35:46.992255 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1875150506.mount: Deactivated successfully. May 17 00:35:46.998547 env[1734]: time="2025-05-17T00:35:46.998465070Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\"" May 17 00:35:47.002684 env[1734]: time="2025-05-17T00:35:47.002639483Z" level=info msg="StartContainer for \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\"" May 17 00:35:47.045461 systemd[1]: Started cri-containerd-c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd.scope. May 17 00:35:47.104217 env[1734]: time="2025-05-17T00:35:47.104149766Z" level=info msg="StartContainer for \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\" returns successfully" May 17 00:35:47.124799 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 17 00:35:47.125164 systemd[1]: Stopped systemd-sysctl.service. May 17 00:35:47.125866 systemd[1]: Stopping systemd-sysctl.service... May 17 00:35:47.129609 systemd[1]: Starting systemd-sysctl.service... May 17 00:35:47.141728 systemd[1]: cri-containerd-c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd.scope: Deactivated successfully. May 17 00:35:47.146443 systemd[1]: Finished systemd-sysctl.service. May 17 00:35:47.246574 env[1734]: time="2025-05-17T00:35:47.246428551Z" level=info msg="shim disconnected" id=c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd May 17 00:35:47.246932 env[1734]: time="2025-05-17T00:35:47.246899667Z" level=warning msg="cleaning up after shim disconnected" id=c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd namespace=k8s.io May 17 00:35:47.247040 env[1734]: time="2025-05-17T00:35:47.247023462Z" level=info msg="cleaning up dead shim" May 17 00:35:47.258963 env[1734]: time="2025-05-17T00:35:47.258913543Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:35:47Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2333 runtime=io.containerd.runc.v2\n" May 17 00:35:47.681223 kubelet[2085]: E0517 00:35:47.681142 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:47.769060 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2115970646.mount: Deactivated successfully. May 17 00:35:47.851909 env[1734]: time="2025-05-17T00:35:47.851816548Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.32.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:47.853989 env[1734]: time="2025-05-17T00:35:47.853942824Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:f532b7356fac4d7c4e4f6763bb5a15a43e3bb740c9fb26c85b906a4d971f2363,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:47.856036 env[1734]: time="2025-05-17T00:35:47.856001311Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.32.5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:47.858195 env[1734]: time="2025-05-17T00:35:47.858153807Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:9dc6553459c3319525ba4090a780db1a133d5dee68c08e07f9b9d6ba83b42a0b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:35:47.858668 env[1734]: time="2025-05-17T00:35:47.858627218Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.5\" returns image reference \"sha256:f532b7356fac4d7c4e4f6763bb5a15a43e3bb740c9fb26c85b906a4d971f2363\"" May 17 00:35:47.860888 env[1734]: time="2025-05-17T00:35:47.860842368Z" level=info msg="CreateContainer within sandbox \"04e7a456f17eb03cb1694d157ccc57989fbfdc3fc513becbdb0da1bf33007c00\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 17 00:35:47.874169 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3046833927.mount: Deactivated successfully. May 17 00:35:47.880268 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1323604101.mount: Deactivated successfully. May 17 00:35:47.885725 env[1734]: time="2025-05-17T00:35:47.885663704Z" level=info msg="CreateContainer within sandbox \"04e7a456f17eb03cb1694d157ccc57989fbfdc3fc513becbdb0da1bf33007c00\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"230157f96a9938876f47d03c4c44265bf25ae45a59b0a276fabe821806b206cd\"" May 17 00:35:47.886361 env[1734]: time="2025-05-17T00:35:47.886235208Z" level=info msg="StartContainer for \"230157f96a9938876f47d03c4c44265bf25ae45a59b0a276fabe821806b206cd\"" May 17 00:35:47.904528 systemd[1]: Started cri-containerd-230157f96a9938876f47d03c4c44265bf25ae45a59b0a276fabe821806b206cd.scope. May 17 00:35:47.946132 env[1734]: time="2025-05-17T00:35:47.943935857Z" level=info msg="StartContainer for \"230157f96a9938876f47d03c4c44265bf25ae45a59b0a276fabe821806b206cd\" returns successfully" May 17 00:35:47.966193 env[1734]: time="2025-05-17T00:35:47.966147377Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 17 00:35:47.987238 env[1734]: time="2025-05-17T00:35:47.987178340Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\"" May 17 00:35:47.988453 env[1734]: time="2025-05-17T00:35:47.988398199Z" level=info msg="StartContainer for \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\"" May 17 00:35:47.994811 kubelet[2085]: I0517 00:35:47.994271 2085 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-7l9cb" podStartSLOduration=3.7686368420000003 podStartE2EDuration="13.994247857s" podCreationTimestamp="2025-05-17 00:35:34 +0000 UTC" firstStartedPulling="2025-05-17 00:35:37.633899937 +0000 UTC m=+3.666828142" lastFinishedPulling="2025-05-17 00:35:47.859510949 +0000 UTC m=+13.892439157" observedRunningTime="2025-05-17 00:35:47.993103272 +0000 UTC m=+14.026031499" watchObservedRunningTime="2025-05-17 00:35:47.994247857 +0000 UTC m=+14.027176117" May 17 00:35:48.029892 systemd[1]: Started cri-containerd-68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85.scope. May 17 00:35:48.082407 env[1734]: time="2025-05-17T00:35:48.082343567Z" level=info msg="StartContainer for \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\" returns successfully" May 17 00:35:48.087621 systemd[1]: cri-containerd-68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85.scope: Deactivated successfully. May 17 00:35:48.213731 env[1734]: time="2025-05-17T00:35:48.212988582Z" level=info msg="shim disconnected" id=68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85 May 17 00:35:48.213731 env[1734]: time="2025-05-17T00:35:48.213211060Z" level=warning msg="cleaning up after shim disconnected" id=68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85 namespace=k8s.io May 17 00:35:48.213731 env[1734]: time="2025-05-17T00:35:48.213230839Z" level=info msg="cleaning up dead shim" May 17 00:35:48.227155 env[1734]: time="2025-05-17T00:35:48.227097373Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:35:48Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2468 runtime=io.containerd.runc.v2\n" May 17 00:35:48.681590 kubelet[2085]: E0517 00:35:48.681507 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:48.979376 env[1734]: time="2025-05-17T00:35:48.979092391Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 17 00:35:48.998254 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1354085919.mount: Deactivated successfully. May 17 00:35:49.016945 env[1734]: time="2025-05-17T00:35:49.016875768Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\"" May 17 00:35:49.017863 env[1734]: time="2025-05-17T00:35:49.017820776Z" level=info msg="StartContainer for \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\"" May 17 00:35:49.042720 systemd[1]: Started cri-containerd-098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05.scope. May 17 00:35:49.086477 systemd[1]: cri-containerd-098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05.scope: Deactivated successfully. May 17 00:35:49.089461 env[1734]: time="2025-05-17T00:35:49.088866783Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a104fbd_8f68_4e88_91cf_90e3dd90a9f4.slice/cri-containerd-098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05.scope/memory.events\": no such file or directory" May 17 00:35:49.093610 env[1734]: time="2025-05-17T00:35:49.093565463Z" level=info msg="StartContainer for \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\" returns successfully" May 17 00:35:49.127087 env[1734]: time="2025-05-17T00:35:49.127032276Z" level=info msg="shim disconnected" id=098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05 May 17 00:35:49.127087 env[1734]: time="2025-05-17T00:35:49.127084887Z" level=warning msg="cleaning up after shim disconnected" id=098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05 namespace=k8s.io May 17 00:35:49.127087 env[1734]: time="2025-05-17T00:35:49.127096238Z" level=info msg="cleaning up dead shim" May 17 00:35:49.136746 env[1734]: time="2025-05-17T00:35:49.136685245Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:35:49Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2618 runtime=io.containerd.runc.v2\n" May 17 00:35:49.682801 kubelet[2085]: E0517 00:35:49.682640 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:49.770720 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05-rootfs.mount: Deactivated successfully. May 17 00:35:49.985784 env[1734]: time="2025-05-17T00:35:49.985660234Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 17 00:35:50.016561 env[1734]: time="2025-05-17T00:35:50.016466041Z" level=info msg="CreateContainer within sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\"" May 17 00:35:50.018390 env[1734]: time="2025-05-17T00:35:50.018348695Z" level=info msg="StartContainer for \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\"" May 17 00:35:50.051029 systemd[1]: run-containerd-runc-k8s.io-d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739-runc.1nTJuB.mount: Deactivated successfully. May 17 00:35:50.056051 systemd[1]: Started cri-containerd-d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739.scope. May 17 00:35:50.101490 env[1734]: time="2025-05-17T00:35:50.101431547Z" level=info msg="StartContainer for \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\" returns successfully" May 17 00:35:50.280659 kubelet[2085]: I0517 00:35:50.280274 2085 kubelet_node_status.go:501] "Fast updating node status as it just became ready" May 17 00:35:50.583354 kernel: Initializing XFRM netlink socket May 17 00:35:50.683836 kubelet[2085]: E0517 00:35:50.683752 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:51.684159 kubelet[2085]: E0517 00:35:51.684108 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:52.047786 systemd[1]: systemd-hostnamed.service: Deactivated successfully. May 17 00:35:52.248277 kubelet[2085]: I0517 00:35:52.248210 2085 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-4wk8n" podStartSLOduration=10.109646716 podStartE2EDuration="18.248185661s" podCreationTimestamp="2025-05-17 00:35:34 +0000 UTC" firstStartedPulling="2025-05-17 00:35:37.612828994 +0000 UTC m=+3.645757199" lastFinishedPulling="2025-05-17 00:35:45.751367925 +0000 UTC m=+11.784296144" observedRunningTime="2025-05-17 00:35:51.004668531 +0000 UTC m=+17.037596758" watchObservedRunningTime="2025-05-17 00:35:52.248185661 +0000 UTC m=+18.281113879" May 17 00:35:52.253930 systemd[1]: Created slice kubepods-besteffort-pod1a10d037_d1f1_4f17_be90_c36addb54be1.slice. May 17 00:35:52.261753 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready May 17 00:35:52.261884 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready May 17 00:35:52.261853 systemd-networkd[1459]: cilium_host: Link UP May 17 00:35:52.261964 systemd-networkd[1459]: cilium_net: Link UP May 17 00:35:52.262095 systemd-networkd[1459]: cilium_net: Gained carrier May 17 00:35:52.262218 systemd-networkd[1459]: cilium_host: Gained carrier May 17 00:35:52.265271 (udev-worker)[2465]: Network interface NamePolicy= disabled on kernel command line. May 17 00:35:52.268415 (udev-worker)[2770]: Network interface NamePolicy= disabled on kernel command line. May 17 00:35:52.348845 kubelet[2085]: I0517 00:35:52.348669 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj476\" (UniqueName: \"kubernetes.io/projected/1a10d037-d1f1-4f17-be90-c36addb54be1-kube-api-access-jj476\") pod \"nginx-deployment-7fcdb87857-pppvb\" (UID: \"1a10d037-d1f1-4f17-be90-c36addb54be1\") " pod="default/nginx-deployment-7fcdb87857-pppvb" May 17 00:35:52.406128 systemd-networkd[1459]: cilium_vxlan: Link UP May 17 00:35:52.406136 systemd-networkd[1459]: cilium_vxlan: Gained carrier May 17 00:35:52.558355 env[1734]: time="2025-05-17T00:35:52.557939218Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-7fcdb87857-pppvb,Uid:1a10d037-d1f1-4f17-be90-c36addb54be1,Namespace:default,Attempt:0,}" May 17 00:35:52.665336 kernel: NET: Registered PF_ALG protocol family May 17 00:35:52.684771 kubelet[2085]: E0517 00:35:52.684731 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:52.985463 systemd-networkd[1459]: cilium_host: Gained IPv6LL May 17 00:35:53.178578 systemd-networkd[1459]: cilium_net: Gained IPv6LL May 17 00:35:53.363809 (udev-worker)[2782]: Network interface NamePolicy= disabled on kernel command line. May 17 00:35:53.390397 systemd-networkd[1459]: lxc_health: Link UP May 17 00:35:53.402466 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready May 17 00:35:53.404518 systemd-networkd[1459]: lxc_health: Gained carrier May 17 00:35:53.614589 systemd-networkd[1459]: lxce534c249a408: Link UP May 17 00:35:53.625425 kernel: eth0: renamed from tmp688e4 May 17 00:35:53.632352 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxce534c249a408: link becomes ready May 17 00:35:53.631098 systemd-networkd[1459]: lxce534c249a408: Gained carrier May 17 00:35:53.685373 kubelet[2085]: E0517 00:35:53.685316 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:54.075381 systemd-networkd[1459]: cilium_vxlan: Gained IPv6LL May 17 00:35:54.585790 systemd-networkd[1459]: lxc_health: Gained IPv6LL May 17 00:35:54.669430 kubelet[2085]: E0517 00:35:54.669375 2085 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:54.686232 kubelet[2085]: E0517 00:35:54.686179 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:55.034926 systemd-networkd[1459]: lxce534c249a408: Gained IPv6LL May 17 00:35:55.686906 kubelet[2085]: E0517 00:35:55.686856 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:56.687982 kubelet[2085]: E0517 00:35:56.687901 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:57.688818 kubelet[2085]: E0517 00:35:57.688763 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:58.263168 env[1734]: time="2025-05-17T00:35:58.263060239Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:35:58.263751 env[1734]: time="2025-05-17T00:35:58.263129116Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:35:58.263751 env[1734]: time="2025-05-17T00:35:58.263171670Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:35:58.263751 env[1734]: time="2025-05-17T00:35:58.263393704Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/688e4ad2b5a0e2109b2acbedba09367c68c61d3ba33791a9a62db4a38413fe87 pid=3145 runtime=io.containerd.runc.v2 May 17 00:35:58.284785 systemd[1]: Started cri-containerd-688e4ad2b5a0e2109b2acbedba09367c68c61d3ba33791a9a62db4a38413fe87.scope. May 17 00:35:58.336458 env[1734]: time="2025-05-17T00:35:58.336407421Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-7fcdb87857-pppvb,Uid:1a10d037-d1f1-4f17-be90-c36addb54be1,Namespace:default,Attempt:0,} returns sandbox id \"688e4ad2b5a0e2109b2acbedba09367c68c61d3ba33791a9a62db4a38413fe87\"" May 17 00:35:58.338800 env[1734]: time="2025-05-17T00:35:58.338755103Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" May 17 00:35:58.690247 kubelet[2085]: E0517 00:35:58.690187 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:35:58.880351 kubelet[2085]: I0517 00:35:58.879826 2085 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 17 00:35:59.690957 kubelet[2085]: E0517 00:35:59.690889 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:00.691943 kubelet[2085]: E0517 00:36:00.691891 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:01.104163 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1919968986.mount: Deactivated successfully. May 17 00:36:01.693130 kubelet[2085]: E0517 00:36:01.692938 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:02.693234 kubelet[2085]: E0517 00:36:02.693153 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:03.694224 kubelet[2085]: E0517 00:36:03.694172 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:04.566198 env[1734]: time="2025-05-17T00:36:04.566132335Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:04.568797 env[1734]: time="2025-05-17T00:36:04.568746166Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7e2dd24abce21cd256091445aca4b7eb00774264c2b0a8714701dd7091509efa,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:04.572196 env[1734]: time="2025-05-17T00:36:04.572146168Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:04.574908 env[1734]: time="2025-05-17T00:36:04.574849161Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx@sha256:beabce8f1782671ba500ddff99dd260fbf9c5ec85fb9c3162e35a3c40bafd023,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:04.575813 env[1734]: time="2025-05-17T00:36:04.575771314Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:7e2dd24abce21cd256091445aca4b7eb00774264c2b0a8714701dd7091509efa\"" May 17 00:36:04.579047 env[1734]: time="2025-05-17T00:36:04.579008194Z" level=info msg="CreateContainer within sandbox \"688e4ad2b5a0e2109b2acbedba09367c68c61d3ba33791a9a62db4a38413fe87\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" May 17 00:36:04.607916 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2340986743.mount: Deactivated successfully. May 17 00:36:04.617207 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount408327604.mount: Deactivated successfully. May 17 00:36:04.621833 env[1734]: time="2025-05-17T00:36:04.621763020Z" level=info msg="CreateContainer within sandbox \"688e4ad2b5a0e2109b2acbedba09367c68c61d3ba33791a9a62db4a38413fe87\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"8eca61ae73b007aebde2d17ad824c82821a9a15152257145a57a3c9c0440d8ac\"" May 17 00:36:04.622704 env[1734]: time="2025-05-17T00:36:04.622620915Z" level=info msg="StartContainer for \"8eca61ae73b007aebde2d17ad824c82821a9a15152257145a57a3c9c0440d8ac\"" May 17 00:36:04.648379 systemd[1]: Started cri-containerd-8eca61ae73b007aebde2d17ad824c82821a9a15152257145a57a3c9c0440d8ac.scope. May 17 00:36:04.688149 env[1734]: time="2025-05-17T00:36:04.687645565Z" level=info msg="StartContainer for \"8eca61ae73b007aebde2d17ad824c82821a9a15152257145a57a3c9c0440d8ac\" returns successfully" May 17 00:36:04.696401 kubelet[2085]: E0517 00:36:04.695154 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:05.046531 kubelet[2085]: I0517 00:36:05.046460 2085 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nginx-deployment-7fcdb87857-pppvb" podStartSLOduration=6.807027007 podStartE2EDuration="13.046443791s" podCreationTimestamp="2025-05-17 00:35:52 +0000 UTC" firstStartedPulling="2025-05-17 00:35:58.33810804 +0000 UTC m=+24.371036248" lastFinishedPulling="2025-05-17 00:36:04.577524815 +0000 UTC m=+30.610453032" observedRunningTime="2025-05-17 00:36:05.046131816 +0000 UTC m=+31.079060044" watchObservedRunningTime="2025-05-17 00:36:05.046443791 +0000 UTC m=+31.079372018" May 17 00:36:05.696980 kubelet[2085]: E0517 00:36:05.696935 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:06.698113 kubelet[2085]: E0517 00:36:06.698021 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:07.324309 update_engine[1729]: I0517 00:36:07.324211 1729 update_attempter.cc:509] Updating boot flags... May 17 00:36:07.701383 kubelet[2085]: E0517 00:36:07.699229 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:08.699562 kubelet[2085]: E0517 00:36:08.699509 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:09.700732 kubelet[2085]: E0517 00:36:09.700678 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:10.701440 kubelet[2085]: E0517 00:36:10.701377 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:11.606194 systemd[1]: Created slice kubepods-besteffort-pod5bbda3d1_1330_4a38_979d_b87380a039db.slice. May 17 00:36:11.622834 kubelet[2085]: I0517 00:36:11.622759 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/5bbda3d1-1330-4a38-979d-b87380a039db-data\") pod \"nfs-server-provisioner-0\" (UID: \"5bbda3d1-1330-4a38-979d-b87380a039db\") " pod="default/nfs-server-provisioner-0" May 17 00:36:11.622834 kubelet[2085]: I0517 00:36:11.622811 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4jn7\" (UniqueName: \"kubernetes.io/projected/5bbda3d1-1330-4a38-979d-b87380a039db-kube-api-access-l4jn7\") pod \"nfs-server-provisioner-0\" (UID: \"5bbda3d1-1330-4a38-979d-b87380a039db\") " pod="default/nfs-server-provisioner-0" May 17 00:36:11.701854 kubelet[2085]: E0517 00:36:11.701794 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:11.911184 env[1734]: time="2025-05-17T00:36:11.911068909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:5bbda3d1-1330-4a38-979d-b87380a039db,Namespace:default,Attempt:0,}" May 17 00:36:11.976152 systemd-networkd[1459]: lxc08b6b168c360: Link UP May 17 00:36:11.979591 (udev-worker)[3518]: Network interface NamePolicy= disabled on kernel command line. May 17 00:36:11.983409 kernel: eth0: renamed from tmp0629a May 17 00:36:11.996193 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 17 00:36:11.996324 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc08b6b168c360: link becomes ready May 17 00:36:11.996639 systemd-networkd[1459]: lxc08b6b168c360: Gained carrier May 17 00:36:12.170350 env[1734]: time="2025-05-17T00:36:12.169788853Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:36:12.170350 env[1734]: time="2025-05-17T00:36:12.169874747Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:36:12.170350 env[1734]: time="2025-05-17T00:36:12.169897332Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:36:12.170698 env[1734]: time="2025-05-17T00:36:12.170651622Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0629a40c24a0435da4074739cfd23aed28ae7b6a5752afd077cbe751deb93d8d pid=3534 runtime=io.containerd.runc.v2 May 17 00:36:12.188937 systemd[1]: Started cri-containerd-0629a40c24a0435da4074739cfd23aed28ae7b6a5752afd077cbe751deb93d8d.scope. May 17 00:36:12.254040 env[1734]: time="2025-05-17T00:36:12.253983599Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:5bbda3d1-1330-4a38-979d-b87380a039db,Namespace:default,Attempt:0,} returns sandbox id \"0629a40c24a0435da4074739cfd23aed28ae7b6a5752afd077cbe751deb93d8d\"" May 17 00:36:12.256275 env[1734]: time="2025-05-17T00:36:12.256226963Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" May 17 00:36:12.702650 kubelet[2085]: E0517 00:36:12.702600 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:13.337664 systemd-networkd[1459]: lxc08b6b168c360: Gained IPv6LL May 17 00:36:13.702886 kubelet[2085]: E0517 00:36:13.702815 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:14.636059 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1912115022.mount: Deactivated successfully. May 17 00:36:14.669551 kubelet[2085]: E0517 00:36:14.669497 2085 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:14.703264 kubelet[2085]: E0517 00:36:14.703173 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:15.703636 kubelet[2085]: E0517 00:36:15.703560 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:16.443889 amazon-ssm-agent[1717]: 2025-05-17 00:36:16 INFO [MessagingDeliveryService] [Association] Schedule manager refreshed with 0 associations, 0 new associations associated May 17 00:36:16.704488 kubelet[2085]: E0517 00:36:16.704021 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:16.792175 env[1734]: time="2025-05-17T00:36:16.792098798Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:16.797630 env[1734]: time="2025-05-17T00:36:16.795088735Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:16.803139 env[1734]: time="2025-05-17T00:36:16.803082300Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:16.806269 env[1734]: time="2025-05-17T00:36:16.806210114Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" May 17 00:36:16.810670 env[1734]: time="2025-05-17T00:36:16.810610950Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:16.817662 env[1734]: time="2025-05-17T00:36:16.817604179Z" level=info msg="CreateContainer within sandbox \"0629a40c24a0435da4074739cfd23aed28ae7b6a5752afd077cbe751deb93d8d\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" May 17 00:36:16.835842 env[1734]: time="2025-05-17T00:36:16.835780867Z" level=info msg="CreateContainer within sandbox \"0629a40c24a0435da4074739cfd23aed28ae7b6a5752afd077cbe751deb93d8d\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"2d857567bc19838cafc52bb091da61a0a5675149017495b54a7c6160053dfb30\"" May 17 00:36:16.836696 env[1734]: time="2025-05-17T00:36:16.836611107Z" level=info msg="StartContainer for \"2d857567bc19838cafc52bb091da61a0a5675149017495b54a7c6160053dfb30\"" May 17 00:36:16.870369 systemd[1]: Started cri-containerd-2d857567bc19838cafc52bb091da61a0a5675149017495b54a7c6160053dfb30.scope. May 17 00:36:16.938921 env[1734]: time="2025-05-17T00:36:16.938814969Z" level=info msg="StartContainer for \"2d857567bc19838cafc52bb091da61a0a5675149017495b54a7c6160053dfb30\" returns successfully" May 17 00:36:17.102610 kubelet[2085]: I0517 00:36:17.102472 2085 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=1.5468757260000001 podStartE2EDuration="6.102453209s" podCreationTimestamp="2025-05-17 00:36:11 +0000 UTC" firstStartedPulling="2025-05-17 00:36:12.255606126 +0000 UTC m=+38.288534349" lastFinishedPulling="2025-05-17 00:36:16.811183613 +0000 UTC m=+42.844111832" observedRunningTime="2025-05-17 00:36:17.102120588 +0000 UTC m=+43.135048827" watchObservedRunningTime="2025-05-17 00:36:17.102453209 +0000 UTC m=+43.135381430" May 17 00:36:17.704615 kubelet[2085]: E0517 00:36:17.704536 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:17.826509 systemd[1]: run-containerd-runc-k8s.io-2d857567bc19838cafc52bb091da61a0a5675149017495b54a7c6160053dfb30-runc.vMGkjy.mount: Deactivated successfully. May 17 00:36:18.705258 kubelet[2085]: E0517 00:36:18.704993 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:19.705649 kubelet[2085]: E0517 00:36:19.705572 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:20.706537 kubelet[2085]: E0517 00:36:20.706478 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:21.706907 kubelet[2085]: E0517 00:36:21.706847 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:22.707846 kubelet[2085]: E0517 00:36:22.707799 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:23.708087 kubelet[2085]: E0517 00:36:23.708008 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:24.709024 kubelet[2085]: E0517 00:36:24.708967 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:25.709760 kubelet[2085]: E0517 00:36:25.709680 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:26.710610 kubelet[2085]: E0517 00:36:26.710541 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:27.076611 systemd[1]: Created slice kubepods-besteffort-pode0ddcb22_6695_42a9_9a22_af4bfa0b400b.slice. May 17 00:36:27.220939 kubelet[2085]: I0517 00:36:27.220880 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ddbb884e-b147-4179-a581-d96680d1b104\" (UniqueName: \"kubernetes.io/nfs/e0ddcb22-6695-42a9-9a22-af4bfa0b400b-pvc-ddbb884e-b147-4179-a581-d96680d1b104\") pod \"test-pod-1\" (UID: \"e0ddcb22-6695-42a9-9a22-af4bfa0b400b\") " pod="default/test-pod-1" May 17 00:36:27.220939 kubelet[2085]: I0517 00:36:27.220929 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqj6m\" (UniqueName: \"kubernetes.io/projected/e0ddcb22-6695-42a9-9a22-af4bfa0b400b-kube-api-access-qqj6m\") pod \"test-pod-1\" (UID: \"e0ddcb22-6695-42a9-9a22-af4bfa0b400b\") " pod="default/test-pod-1" May 17 00:36:27.402319 kernel: FS-Cache: Loaded May 17 00:36:27.454977 kernel: RPC: Registered named UNIX socket transport module. May 17 00:36:27.455152 kernel: RPC: Registered udp transport module. May 17 00:36:27.455190 kernel: RPC: Registered tcp transport module. May 17 00:36:27.455219 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. May 17 00:36:27.526351 kernel: FS-Cache: Netfs 'nfs' registered for caching May 17 00:36:27.711669 kubelet[2085]: E0517 00:36:27.711504 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:27.727427 kernel: NFS: Registering the id_resolver key type May 17 00:36:27.727592 kernel: Key type id_resolver registered May 17 00:36:27.727633 kernel: Key type id_legacy registered May 17 00:36:27.775721 nfsidmap[3652]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'us-west-2.compute.internal' May 17 00:36:27.779903 nfsidmap[3653]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'us-west-2.compute.internal' May 17 00:36:27.980646 env[1734]: time="2025-05-17T00:36:27.980501926Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:e0ddcb22-6695-42a9-9a22-af4bfa0b400b,Namespace:default,Attempt:0,}" May 17 00:36:28.017190 (udev-worker)[3640]: Network interface NamePolicy= disabled on kernel command line. May 17 00:36:28.017248 (udev-worker)[3646]: Network interface NamePolicy= disabled on kernel command line. May 17 00:36:28.020173 systemd-networkd[1459]: lxc7421b91834c5: Link UP May 17 00:36:28.027333 kernel: eth0: renamed from tmp01b5b May 17 00:36:28.032562 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready May 17 00:36:28.032668 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc7421b91834c5: link becomes ready May 17 00:36:28.032622 systemd-networkd[1459]: lxc7421b91834c5: Gained carrier May 17 00:36:28.232962 env[1734]: time="2025-05-17T00:36:28.232387094Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:36:28.232962 env[1734]: time="2025-05-17T00:36:28.232421683Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:36:28.232962 env[1734]: time="2025-05-17T00:36:28.232431748Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:36:28.232962 env[1734]: time="2025-05-17T00:36:28.232544489Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/01b5b14d4560fca36084de00073940528feeca351d4b9d1f709623d40e9c380d pid=3679 runtime=io.containerd.runc.v2 May 17 00:36:28.254519 systemd[1]: Started cri-containerd-01b5b14d4560fca36084de00073940528feeca351d4b9d1f709623d40e9c380d.scope. May 17 00:36:28.309925 env[1734]: time="2025-05-17T00:36:28.309873432Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:e0ddcb22-6695-42a9-9a22-af4bfa0b400b,Namespace:default,Attempt:0,} returns sandbox id \"01b5b14d4560fca36084de00073940528feeca351d4b9d1f709623d40e9c380d\"" May 17 00:36:28.312066 env[1734]: time="2025-05-17T00:36:28.312026172Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" May 17 00:36:28.614828 env[1734]: time="2025-05-17T00:36:28.614331884Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:28.618383 env[1734]: time="2025-05-17T00:36:28.618339284Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7e2dd24abce21cd256091445aca4b7eb00774264c2b0a8714701dd7091509efa,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:28.621415 env[1734]: time="2025-05-17T00:36:28.621375669Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:28.624239 env[1734]: time="2025-05-17T00:36:28.624201764Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx@sha256:beabce8f1782671ba500ddff99dd260fbf9c5ec85fb9c3162e35a3c40bafd023,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:28.625209 env[1734]: time="2025-05-17T00:36:28.625127560Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:7e2dd24abce21cd256091445aca4b7eb00774264c2b0a8714701dd7091509efa\"" May 17 00:36:28.627648 env[1734]: time="2025-05-17T00:36:28.627614352Z" level=info msg="CreateContainer within sandbox \"01b5b14d4560fca36084de00073940528feeca351d4b9d1f709623d40e9c380d\" for container &ContainerMetadata{Name:test,Attempt:0,}" May 17 00:36:28.649629 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2533062980.mount: Deactivated successfully. May 17 00:36:28.664649 env[1734]: time="2025-05-17T00:36:28.664588237Z" level=info msg="CreateContainer within sandbox \"01b5b14d4560fca36084de00073940528feeca351d4b9d1f709623d40e9c380d\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"58242395be60a64d5e0c7f45ad939899115fd7d33dab6f7274d7e5ecc1d30eff\"" May 17 00:36:28.665564 env[1734]: time="2025-05-17T00:36:28.665519819Z" level=info msg="StartContainer for \"58242395be60a64d5e0c7f45ad939899115fd7d33dab6f7274d7e5ecc1d30eff\"" May 17 00:36:28.687536 systemd[1]: Started cri-containerd-58242395be60a64d5e0c7f45ad939899115fd7d33dab6f7274d7e5ecc1d30eff.scope. May 17 00:36:28.712817 kubelet[2085]: E0517 00:36:28.711628 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:28.727754 env[1734]: time="2025-05-17T00:36:28.727149482Z" level=info msg="StartContainer for \"58242395be60a64d5e0c7f45ad939899115fd7d33dab6f7274d7e5ecc1d30eff\" returns successfully" May 17 00:36:29.125918 kubelet[2085]: I0517 00:36:29.125866 2085 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=16.810940372 podStartE2EDuration="17.1258507s" podCreationTimestamp="2025-05-17 00:36:12 +0000 UTC" firstStartedPulling="2025-05-17 00:36:28.31135839 +0000 UTC m=+54.344286594" lastFinishedPulling="2025-05-17 00:36:28.626268703 +0000 UTC m=+54.659196922" observedRunningTime="2025-05-17 00:36:29.125585949 +0000 UTC m=+55.158514166" watchObservedRunningTime="2025-05-17 00:36:29.1258507 +0000 UTC m=+55.158778927" May 17 00:36:29.465614 systemd-networkd[1459]: lxc7421b91834c5: Gained IPv6LL May 17 00:36:29.712617 kubelet[2085]: E0517 00:36:29.712557 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:30.712750 kubelet[2085]: E0517 00:36:30.712700 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:31.713305 kubelet[2085]: E0517 00:36:31.713208 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:32.713506 kubelet[2085]: E0517 00:36:32.713427 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:33.714542 kubelet[2085]: E0517 00:36:33.714482 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:34.669591 kubelet[2085]: E0517 00:36:34.669378 2085 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:34.715562 kubelet[2085]: E0517 00:36:34.715500 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:35.715909 kubelet[2085]: E0517 00:36:35.715844 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:36.716606 kubelet[2085]: E0517 00:36:36.716547 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:37.717803 kubelet[2085]: E0517 00:36:37.717746 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:38.718460 kubelet[2085]: E0517 00:36:38.718382 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:39.719407 kubelet[2085]: E0517 00:36:39.719357 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:40.262869 systemd[1]: run-containerd-runc-k8s.io-d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739-runc.hjLfM4.mount: Deactivated successfully. May 17 00:36:40.290203 env[1734]: time="2025-05-17T00:36:40.290122271Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 17 00:36:40.300471 env[1734]: time="2025-05-17T00:36:40.300429321Z" level=info msg="StopContainer for \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\" with timeout 2 (s)" May 17 00:36:40.300795 env[1734]: time="2025-05-17T00:36:40.300755557Z" level=info msg="Stop container \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\" with signal terminated" May 17 00:36:40.309175 systemd-networkd[1459]: lxc_health: Link DOWN May 17 00:36:40.309185 systemd-networkd[1459]: lxc_health: Lost carrier May 17 00:36:40.335801 systemd[1]: cri-containerd-d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739.scope: Deactivated successfully. May 17 00:36:40.336280 systemd[1]: cri-containerd-d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739.scope: Consumed 8.414s CPU time. May 17 00:36:40.363437 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739-rootfs.mount: Deactivated successfully. May 17 00:36:40.396680 env[1734]: time="2025-05-17T00:36:40.396633233Z" level=info msg="shim disconnected" id=d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739 May 17 00:36:40.396888 env[1734]: time="2025-05-17T00:36:40.396694511Z" level=warning msg="cleaning up after shim disconnected" id=d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739 namespace=k8s.io May 17 00:36:40.396888 env[1734]: time="2025-05-17T00:36:40.396712827Z" level=info msg="cleaning up dead shim" May 17 00:36:40.406717 env[1734]: time="2025-05-17T00:36:40.406667049Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:40Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3813 runtime=io.containerd.runc.v2\n" May 17 00:36:40.410436 env[1734]: time="2025-05-17T00:36:40.410372148Z" level=info msg="StopContainer for \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\" returns successfully" May 17 00:36:40.411089 env[1734]: time="2025-05-17T00:36:40.411054690Z" level=info msg="StopPodSandbox for \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\"" May 17 00:36:40.411207 env[1734]: time="2025-05-17T00:36:40.411114100Z" level=info msg="Container to stop \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:36:40.411207 env[1734]: time="2025-05-17T00:36:40.411127654Z" level=info msg="Container to stop \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:36:40.411207 env[1734]: time="2025-05-17T00:36:40.411139755Z" level=info msg="Container to stop \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:36:40.411207 env[1734]: time="2025-05-17T00:36:40.411152418Z" level=info msg="Container to stop \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:36:40.411207 env[1734]: time="2025-05-17T00:36:40.411162251Z" level=info msg="Container to stop \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:36:40.413466 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7-shm.mount: Deactivated successfully. May 17 00:36:40.422730 systemd[1]: cri-containerd-22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7.scope: Deactivated successfully. May 17 00:36:40.464682 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7-rootfs.mount: Deactivated successfully. May 17 00:36:40.474430 env[1734]: time="2025-05-17T00:36:40.474367887Z" level=info msg="shim disconnected" id=22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7 May 17 00:36:40.474762 env[1734]: time="2025-05-17T00:36:40.474434451Z" level=warning msg="cleaning up after shim disconnected" id=22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7 namespace=k8s.io May 17 00:36:40.474762 env[1734]: time="2025-05-17T00:36:40.474449469Z" level=info msg="cleaning up dead shim" May 17 00:36:40.484904 env[1734]: time="2025-05-17T00:36:40.484846262Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:40Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3847 runtime=io.containerd.runc.v2\n" May 17 00:36:40.486087 env[1734]: time="2025-05-17T00:36:40.486013655Z" level=info msg="TearDown network for sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" successfully" May 17 00:36:40.486087 env[1734]: time="2025-05-17T00:36:40.486057142Z" level=info msg="StopPodSandbox for \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" returns successfully" May 17 00:36:40.618076 kubelet[2085]: I0517 00:36:40.616254 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-net\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618076 kubelet[2085]: I0517 00:36:40.616319 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cni-path\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618076 kubelet[2085]: I0517 00:36:40.616340 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-lib-modules\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618076 kubelet[2085]: I0517 00:36:40.616340 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618076 kubelet[2085]: I0517 00:36:40.616382 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618377 kubelet[2085]: I0517 00:36:40.616402 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cni-path" (OuterVolumeSpecName: "cni-path") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618377 kubelet[2085]: I0517 00:36:40.616354 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-etc-cni-netd\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618377 kubelet[2085]: I0517 00:36:40.616417 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618377 kubelet[2085]: I0517 00:36:40.616427 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-kernel\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618377 kubelet[2085]: I0517 00:36:40.616448 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lzj6\" (UniqueName: \"kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-kube-api-access-4lzj6\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618514 kubelet[2085]: I0517 00:36:40.616438 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618514 kubelet[2085]: I0517 00:36:40.616465 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hubble-tls\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618514 kubelet[2085]: I0517 00:36:40.616479 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-run\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618514 kubelet[2085]: I0517 00:36:40.616496 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-cgroup\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618514 kubelet[2085]: I0517 00:36:40.616515 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-bpf-maps\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618514 kubelet[2085]: I0517 00:36:40.616539 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-config-path\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618757 kubelet[2085]: I0517 00:36:40.616558 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-clustermesh-secrets\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618757 kubelet[2085]: I0517 00:36:40.616572 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-xtables-lock\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618757 kubelet[2085]: I0517 00:36:40.616590 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hostproc\") pod \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\" (UID: \"2a104fbd-8f68-4e88-91cf-90e3dd90a9f4\") " May 17 00:36:40.618757 kubelet[2085]: I0517 00:36:40.616622 2085 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cni-path\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.618757 kubelet[2085]: I0517 00:36:40.616632 2085 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-net\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.618757 kubelet[2085]: I0517 00:36:40.616645 2085 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-host-proc-sys-kernel\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.618757 kubelet[2085]: I0517 00:36:40.616653 2085 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-lib-modules\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.618947 kubelet[2085]: I0517 00:36:40.616660 2085 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-etc-cni-netd\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.618947 kubelet[2085]: I0517 00:36:40.616688 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hostproc" (OuterVolumeSpecName: "hostproc") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618947 kubelet[2085]: I0517 00:36:40.616708 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618947 kubelet[2085]: I0517 00:36:40.616725 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.618947 kubelet[2085]: I0517 00:36:40.616748 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.619315 kubelet[2085]: I0517 00:36:40.619119 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:40.620971 kubelet[2085]: I0517 00:36:40.620937 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 17 00:36:40.624135 kubelet[2085]: I0517 00:36:40.624101 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 17 00:36:40.624631 kubelet[2085]: I0517 00:36:40.624455 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-kube-api-access-4lzj6" (OuterVolumeSpecName: "kube-api-access-4lzj6") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "kube-api-access-4lzj6". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 17 00:36:40.624751 kubelet[2085]: I0517 00:36:40.624635 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" (UID: "2a104fbd-8f68-4e88-91cf-90e3dd90a9f4"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 17 00:36:40.719188 kubelet[2085]: I0517 00:36:40.719142 2085 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hostproc\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719188 kubelet[2085]: I0517 00:36:40.719183 2085 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-clustermesh-secrets\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719188 kubelet[2085]: I0517 00:36:40.719194 2085 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-xtables-lock\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719188 kubelet[2085]: I0517 00:36:40.719206 2085 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4lzj6\" (UniqueName: \"kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-kube-api-access-4lzj6\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719188 kubelet[2085]: I0517 00:36:40.719215 2085 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-bpf-maps\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719782 kubelet[2085]: I0517 00:36:40.719223 2085 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-config-path\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719782 kubelet[2085]: I0517 00:36:40.719235 2085 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-hubble-tls\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719782 kubelet[2085]: I0517 00:36:40.719242 2085 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-run\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.719782 kubelet[2085]: I0517 00:36:40.719250 2085 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4-cilium-cgroup\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:40.720395 kubelet[2085]: E0517 00:36:40.720349 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:40.920821 systemd[1]: Removed slice kubepods-burstable-pod2a104fbd_8f68_4e88_91cf_90e3dd90a9f4.slice. May 17 00:36:40.920972 systemd[1]: kubepods-burstable-pod2a104fbd_8f68_4e88_91cf_90e3dd90a9f4.slice: Consumed 8.544s CPU time. May 17 00:36:41.136360 kubelet[2085]: I0517 00:36:41.136328 2085 scope.go:117] "RemoveContainer" containerID="d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739" May 17 00:36:41.139943 env[1734]: time="2025-05-17T00:36:41.139889455Z" level=info msg="RemoveContainer for \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\"" May 17 00:36:41.146913 env[1734]: time="2025-05-17T00:36:41.146352010Z" level=info msg="RemoveContainer for \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\" returns successfully" May 17 00:36:41.150557 kubelet[2085]: I0517 00:36:41.150506 2085 scope.go:117] "RemoveContainer" containerID="098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05" May 17 00:36:41.151932 env[1734]: time="2025-05-17T00:36:41.151893361Z" level=info msg="RemoveContainer for \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\"" May 17 00:36:41.157825 env[1734]: time="2025-05-17T00:36:41.157760985Z" level=info msg="RemoveContainer for \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\" returns successfully" May 17 00:36:41.158065 kubelet[2085]: I0517 00:36:41.158027 2085 scope.go:117] "RemoveContainer" containerID="68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85" May 17 00:36:41.159756 env[1734]: time="2025-05-17T00:36:41.159718028Z" level=info msg="RemoveContainer for \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\"" May 17 00:36:41.165187 env[1734]: time="2025-05-17T00:36:41.164950707Z" level=info msg="RemoveContainer for \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\" returns successfully" May 17 00:36:41.165437 kubelet[2085]: I0517 00:36:41.165408 2085 scope.go:117] "RemoveContainer" containerID="c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd" May 17 00:36:41.166623 env[1734]: time="2025-05-17T00:36:41.166586301Z" level=info msg="RemoveContainer for \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\"" May 17 00:36:41.171861 env[1734]: time="2025-05-17T00:36:41.171800183Z" level=info msg="RemoveContainer for \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\" returns successfully" May 17 00:36:41.172130 kubelet[2085]: I0517 00:36:41.172094 2085 scope.go:117] "RemoveContainer" containerID="abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c" May 17 00:36:41.174312 env[1734]: time="2025-05-17T00:36:41.173643931Z" level=info msg="RemoveContainer for \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\"" May 17 00:36:41.178787 env[1734]: time="2025-05-17T00:36:41.178738550Z" level=info msg="RemoveContainer for \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\" returns successfully" May 17 00:36:41.179075 kubelet[2085]: I0517 00:36:41.179049 2085 scope.go:117] "RemoveContainer" containerID="d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739" May 17 00:36:41.179466 env[1734]: time="2025-05-17T00:36:41.179374534Z" level=error msg="ContainerStatus for \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\": not found" May 17 00:36:41.179630 kubelet[2085]: E0517 00:36:41.179601 2085 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\": not found" containerID="d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739" May 17 00:36:41.179754 kubelet[2085]: I0517 00:36:41.179650 2085 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739"} err="failed to get container status \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\": rpc error: code = NotFound desc = an error occurred when try to find container \"d989c346dd560ceb1e8b3c950bf7a1bb90a8e15e415a21ca8260b6aed633d739\": not found" May 17 00:36:41.179828 kubelet[2085]: I0517 00:36:41.179756 2085 scope.go:117] "RemoveContainer" containerID="098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05" May 17 00:36:41.180029 env[1734]: time="2025-05-17T00:36:41.179965453Z" level=error msg="ContainerStatus for \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\": not found" May 17 00:36:41.180153 kubelet[2085]: E0517 00:36:41.180128 2085 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\": not found" containerID="098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05" May 17 00:36:41.180227 kubelet[2085]: I0517 00:36:41.180158 2085 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05"} err="failed to get container status \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\": rpc error: code = NotFound desc = an error occurred when try to find container \"098d56ddf5a64ac24b1dc127afba9b995455f3f4c35d5b8ad51126a298c68e05\": not found" May 17 00:36:41.180227 kubelet[2085]: I0517 00:36:41.180179 2085 scope.go:117] "RemoveContainer" containerID="68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85" May 17 00:36:41.180459 env[1734]: time="2025-05-17T00:36:41.180400048Z" level=error msg="ContainerStatus for \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\": not found" May 17 00:36:41.180577 kubelet[2085]: E0517 00:36:41.180549 2085 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\": not found" containerID="68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85" May 17 00:36:41.180650 kubelet[2085]: I0517 00:36:41.180582 2085 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85"} err="failed to get container status \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\": rpc error: code = NotFound desc = an error occurred when try to find container \"68c3d66a0e97952978009b596a2ca9989e1d6796cb0bce4420aa641ac3ea9c85\": not found" May 17 00:36:41.180650 kubelet[2085]: I0517 00:36:41.180605 2085 scope.go:117] "RemoveContainer" containerID="c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd" May 17 00:36:41.180832 env[1734]: time="2025-05-17T00:36:41.180782984Z" level=error msg="ContainerStatus for \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\": not found" May 17 00:36:41.180948 kubelet[2085]: E0517 00:36:41.180923 2085 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\": not found" containerID="c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd" May 17 00:36:41.181110 kubelet[2085]: I0517 00:36:41.180949 2085 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd"} err="failed to get container status \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\": rpc error: code = NotFound desc = an error occurred when try to find container \"c372c8e865536964aed828803c102dced855a7bab088f2c604c8208cd45798fd\": not found" May 17 00:36:41.181110 kubelet[2085]: I0517 00:36:41.180968 2085 scope.go:117] "RemoveContainer" containerID="abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c" May 17 00:36:41.181390 env[1734]: time="2025-05-17T00:36:41.181322152Z" level=error msg="ContainerStatus for \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\": not found" May 17 00:36:41.181491 kubelet[2085]: E0517 00:36:41.181468 2085 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\": not found" containerID="abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c" May 17 00:36:41.181570 kubelet[2085]: I0517 00:36:41.181496 2085 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c"} err="failed to get container status \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\": rpc error: code = NotFound desc = an error occurred when try to find container \"abc187caa16ba2df990beadf967c7c28eb6c8d8e45e78ca781212b0edb42cc5c\": not found" May 17 00:36:41.258409 systemd[1]: var-lib-kubelet-pods-2a104fbd\x2d8f68\x2d4e88\x2d91cf\x2d90e3dd90a9f4-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4lzj6.mount: Deactivated successfully. May 17 00:36:41.258549 systemd[1]: var-lib-kubelet-pods-2a104fbd\x2d8f68\x2d4e88\x2d91cf\x2d90e3dd90a9f4-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 17 00:36:41.258641 systemd[1]: var-lib-kubelet-pods-2a104fbd\x2d8f68\x2d4e88\x2d91cf\x2d90e3dd90a9f4-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 17 00:36:41.721357 kubelet[2085]: E0517 00:36:41.721270 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:42.721643 kubelet[2085]: E0517 00:36:42.721520 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:42.914778 kubelet[2085]: I0517 00:36:42.914719 2085 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" path="/var/lib/kubelet/pods/2a104fbd-8f68-4e88-91cf-90e3dd90a9f4/volumes" May 17 00:36:43.211270 kubelet[2085]: I0517 00:36:43.211217 2085 memory_manager.go:355] "RemoveStaleState removing state" podUID="2a104fbd-8f68-4e88-91cf-90e3dd90a9f4" containerName="cilium-agent" May 17 00:36:43.218208 systemd[1]: Created slice kubepods-besteffort-pod205b87b5_4b0b_4d10_b5ff_bb9ffbd12998.slice. May 17 00:36:43.249272 systemd[1]: Created slice kubepods-burstable-pod949e3c4e_3b61_4f3c_82e3_a0d23d7722fd.slice. May 17 00:36:43.335508 kubelet[2085]: I0517 00:36:43.335444 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/205b87b5-4b0b-4d10-b5ff-bb9ffbd12998-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-llz5d\" (UID: \"205b87b5-4b0b-4d10-b5ff-bb9ffbd12998\") " pod="kube-system/cilium-operator-6c4d7847fc-llz5d" May 17 00:36:43.335508 kubelet[2085]: I0517 00:36:43.335495 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws2nc\" (UniqueName: \"kubernetes.io/projected/205b87b5-4b0b-4d10-b5ff-bb9ffbd12998-kube-api-access-ws2nc\") pod \"cilium-operator-6c4d7847fc-llz5d\" (UID: \"205b87b5-4b0b-4d10-b5ff-bb9ffbd12998\") " pod="kube-system/cilium-operator-6c4d7847fc-llz5d" May 17 00:36:43.436495 kubelet[2085]: I0517 00:36:43.436441 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hostproc\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436495 kubelet[2085]: I0517 00:36:43.436493 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-cgroup\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436495 kubelet[2085]: I0517 00:36:43.436510 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-etc-cni-netd\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436733 kubelet[2085]: I0517 00:36:43.436528 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-xtables-lock\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436733 kubelet[2085]: I0517 00:36:43.436543 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-config-path\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436733 kubelet[2085]: I0517 00:36:43.436561 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-ipsec-secrets\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436733 kubelet[2085]: I0517 00:36:43.436582 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-run\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436733 kubelet[2085]: I0517 00:36:43.436596 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-bpf-maps\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436733 kubelet[2085]: I0517 00:36:43.436609 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-lib-modules\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436902 kubelet[2085]: I0517 00:36:43.436626 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-net\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436902 kubelet[2085]: I0517 00:36:43.436640 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hubble-tls\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436902 kubelet[2085]: I0517 00:36:43.436665 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k75jx\" (UniqueName: \"kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-kube-api-access-k75jx\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436902 kubelet[2085]: I0517 00:36:43.436696 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cni-path\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.436902 kubelet[2085]: I0517 00:36:43.436714 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-clustermesh-secrets\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.437212 kubelet[2085]: I0517 00:36:43.436732 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-kernel\") pod \"cilium-k2qp8\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " pod="kube-system/cilium-k2qp8" May 17 00:36:43.523600 env[1734]: time="2025-05-17T00:36:43.523472971Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-llz5d,Uid:205b87b5-4b0b-4d10-b5ff-bb9ffbd12998,Namespace:kube-system,Attempt:0,}" May 17 00:36:43.564012 env[1734]: time="2025-05-17T00:36:43.563900104Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:36:43.564194 env[1734]: time="2025-05-17T00:36:43.564023926Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:36:43.564194 env[1734]: time="2025-05-17T00:36:43.564056632Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:36:43.564348 env[1734]: time="2025-05-17T00:36:43.564242488Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/939439b197db49b03fed25c174023c334a8a8b2b948a730032e48b597bcf5ce4 pid=3871 runtime=io.containerd.runc.v2 May 17 00:36:43.594398 systemd[1]: Started cri-containerd-939439b197db49b03fed25c174023c334a8a8b2b948a730032e48b597bcf5ce4.scope. May 17 00:36:43.651251 env[1734]: time="2025-05-17T00:36:43.651201673Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-llz5d,Uid:205b87b5-4b0b-4d10-b5ff-bb9ffbd12998,Namespace:kube-system,Attempt:0,} returns sandbox id \"939439b197db49b03fed25c174023c334a8a8b2b948a730032e48b597bcf5ce4\"" May 17 00:36:43.653559 env[1734]: time="2025-05-17T00:36:43.653513404Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 17 00:36:43.722638 kubelet[2085]: E0517 00:36:43.722573 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:43.767050 env[1734]: time="2025-05-17T00:36:43.767000488Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-k2qp8,Uid:949e3c4e-3b61-4f3c-82e3-a0d23d7722fd,Namespace:kube-system,Attempt:0,}" May 17 00:36:43.789618 env[1734]: time="2025-05-17T00:36:43.788897952Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:36:43.789618 env[1734]: time="2025-05-17T00:36:43.788959204Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:36:43.789618 env[1734]: time="2025-05-17T00:36:43.788977704Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:36:43.789844 env[1734]: time="2025-05-17T00:36:43.789424286Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15 pid=3916 runtime=io.containerd.runc.v2 May 17 00:36:43.803313 systemd[1]: Started cri-containerd-093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15.scope. May 17 00:36:43.833456 env[1734]: time="2025-05-17T00:36:43.833386587Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-k2qp8,Uid:949e3c4e-3b61-4f3c-82e3-a0d23d7722fd,Namespace:kube-system,Attempt:0,} returns sandbox id \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\"" May 17 00:36:43.836626 env[1734]: time="2025-05-17T00:36:43.836577833Z" level=info msg="CreateContainer within sandbox \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:36:43.860777 env[1734]: time="2025-05-17T00:36:43.860703390Z" level=info msg="CreateContainer within sandbox \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\"" May 17 00:36:43.861667 env[1734]: time="2025-05-17T00:36:43.861620117Z" level=info msg="StartContainer for \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\"" May 17 00:36:43.879896 systemd[1]: Started cri-containerd-356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50.scope. May 17 00:36:43.894019 systemd[1]: cri-containerd-356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50.scope: Deactivated successfully. May 17 00:36:43.922801 env[1734]: time="2025-05-17T00:36:43.922727238Z" level=info msg="shim disconnected" id=356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50 May 17 00:36:43.922801 env[1734]: time="2025-05-17T00:36:43.922804290Z" level=warning msg="cleaning up after shim disconnected" id=356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50 namespace=k8s.io May 17 00:36:43.922801 env[1734]: time="2025-05-17T00:36:43.922818406Z" level=info msg="cleaning up dead shim" May 17 00:36:43.933231 env[1734]: time="2025-05-17T00:36:43.932989315Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:43Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3976 runtime=io.containerd.runc.v2\ntime=\"2025-05-17T00:36:43Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" May 17 00:36:43.933675 env[1734]: time="2025-05-17T00:36:43.933536646Z" level=error msg="copy shim log" error="read /proc/self/fd/58: file already closed" May 17 00:36:43.933888 env[1734]: time="2025-05-17T00:36:43.933842796Z" level=error msg="Failed to pipe stderr of container \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\"" error="reading from a closed fifo" May 17 00:36:43.934415 env[1734]: time="2025-05-17T00:36:43.934369380Z" level=error msg="Failed to pipe stdout of container \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\"" error="reading from a closed fifo" May 17 00:36:43.937816 env[1734]: time="2025-05-17T00:36:43.937736253Z" level=error msg="StartContainer for \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" May 17 00:36:43.938177 kubelet[2085]: E0517 00:36:43.938109 2085 log.go:32] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50" May 17 00:36:43.938425 kubelet[2085]: E0517 00:36:43.938389 2085 kuberuntime_manager.go:1341] "Unhandled Error" err=< May 17 00:36:43.938425 kubelet[2085]: init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; May 17 00:36:43.938425 kubelet[2085]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; May 17 00:36:43.938425 kubelet[2085]: rm /hostbin/cilium-mount May 17 00:36:43.938615 kubelet[2085]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k75jx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:&AppArmorProfile{Type:Unconfined,LocalhostProfile:nil,},},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cilium-k2qp8_kube-system(949e3c4e-3b61-4f3c-82e3-a0d23d7722fd): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown May 17 00:36:43.938615 kubelet[2085]: > logger="UnhandledError" May 17 00:36:43.940394 kubelet[2085]: E0517 00:36:43.940248 2085 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-k2qp8" podUID="949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" May 17 00:36:44.144561 env[1734]: time="2025-05-17T00:36:44.144422525Z" level=info msg="StopPodSandbox for \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\"" May 17 00:36:44.144561 env[1734]: time="2025-05-17T00:36:44.144502617Z" level=info msg="Container to stop \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:36:44.153880 systemd[1]: cri-containerd-093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15.scope: Deactivated successfully. May 17 00:36:44.188401 env[1734]: time="2025-05-17T00:36:44.188276311Z" level=info msg="shim disconnected" id=093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15 May 17 00:36:44.189376 env[1734]: time="2025-05-17T00:36:44.189337321Z" level=warning msg="cleaning up after shim disconnected" id=093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15 namespace=k8s.io May 17 00:36:44.189583 env[1734]: time="2025-05-17T00:36:44.189370683Z" level=info msg="cleaning up dead shim" May 17 00:36:44.199427 env[1734]: time="2025-05-17T00:36:44.199373516Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:44Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4007 runtime=io.containerd.runc.v2\n" May 17 00:36:44.199767 env[1734]: time="2025-05-17T00:36:44.199728878Z" level=info msg="TearDown network for sandbox \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" successfully" May 17 00:36:44.199767 env[1734]: time="2025-05-17T00:36:44.199763080Z" level=info msg="StopPodSandbox for \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" returns successfully" May 17 00:36:44.343060 kubelet[2085]: I0517 00:36:44.342997 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-bpf-maps\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343060 kubelet[2085]: I0517 00:36:44.343044 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-etc-cni-netd\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343079 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-config-path\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343117 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-run\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343174 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k75jx\" (UniqueName: \"kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-kube-api-access-k75jx\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343216 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cni-path\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343242 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-kernel\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343264 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hostproc\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343331 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-xtables-lock\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343364 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hubble-tls\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343413 kubelet[2085]: I0517 00:36:44.343395 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-clustermesh-secrets\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343914 kubelet[2085]: I0517 00:36:44.343418 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-cgroup\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343914 kubelet[2085]: I0517 00:36:44.343452 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-ipsec-secrets\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343914 kubelet[2085]: I0517 00:36:44.343476 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-lib-modules\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343914 kubelet[2085]: I0517 00:36:44.343507 2085 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-net\") pod \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\" (UID: \"949e3c4e-3b61-4f3c-82e3-a0d23d7722fd\") " May 17 00:36:44.343914 kubelet[2085]: I0517 00:36:44.343611 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.343914 kubelet[2085]: I0517 00:36:44.343652 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.343914 kubelet[2085]: I0517 00:36:44.343738 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.347489 kubelet[2085]: I0517 00:36:44.344280 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.347489 kubelet[2085]: I0517 00:36:44.344375 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.347489 kubelet[2085]: I0517 00:36:44.345924 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cni-path" (OuterVolumeSpecName: "cni-path") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.347489 kubelet[2085]: I0517 00:36:44.345991 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.347489 kubelet[2085]: I0517 00:36:44.346040 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hostproc" (OuterVolumeSpecName: "hostproc") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.347489 kubelet[2085]: I0517 00:36:44.346072 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.347489 kubelet[2085]: I0517 00:36:44.347057 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 17 00:36:44.348518 kubelet[2085]: I0517 00:36:44.348123 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 17 00:36:44.351517 kubelet[2085]: I0517 00:36:44.351476 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 17 00:36:44.353945 kubelet[2085]: I0517 00:36:44.353895 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-kube-api-access-k75jx" (OuterVolumeSpecName: "kube-api-access-k75jx") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "kube-api-access-k75jx". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 17 00:36:44.357274 kubelet[2085]: I0517 00:36:44.357223 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 17 00:36:44.357274 kubelet[2085]: I0517 00:36:44.357331 2085 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" (UID: "949e3c4e-3b61-4f3c-82e3-a0d23d7722fd"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 17 00:36:44.444581 kubelet[2085]: I0517 00:36:44.444526 2085 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-k75jx\" (UniqueName: \"kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-kube-api-access-k75jx\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.444581 kubelet[2085]: I0517 00:36:44.444573 2085 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-bpf-maps\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.444581 kubelet[2085]: I0517 00:36:44.444593 2085 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-etc-cni-netd\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444605 2085 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-config-path\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444617 2085 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-run\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444627 2085 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cni-path\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444651 2085 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hubble-tls\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444661 2085 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-kernel\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444673 2085 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-hostproc\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444684 2085 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-xtables-lock\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444695 2085 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-clustermesh-secrets\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444706 2085 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-cgroup\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444717 2085 reconciler_common.go:299] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-cilium-ipsec-secrets\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444726 2085 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-lib-modules\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.445205 kubelet[2085]: I0517 00:36:44.444739 2085 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd-host-proc-sys-net\") on node \"172.31.23.93\" DevicePath \"\"" May 17 00:36:44.458490 systemd[1]: var-lib-kubelet-pods-949e3c4e\x2d3b61\x2d4f3c\x2d82e3\x2da0d23d7722fd-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 17 00:36:44.458610 systemd[1]: var-lib-kubelet-pods-949e3c4e\x2d3b61\x2d4f3c\x2d82e3\x2da0d23d7722fd-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 17 00:36:44.723184 kubelet[2085]: E0517 00:36:44.723005 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:44.825662 kubelet[2085]: E0517 00:36:44.825618 2085 kubelet.go:3002] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 17 00:36:44.918023 systemd[1]: Removed slice kubepods-burstable-pod949e3c4e_3b61_4f3c_82e3_a0d23d7722fd.slice. May 17 00:36:45.149928 kubelet[2085]: I0517 00:36:45.149696 2085 scope.go:117] "RemoveContainer" containerID="356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50" May 17 00:36:45.156385 env[1734]: time="2025-05-17T00:36:45.156336915Z" level=info msg="RemoveContainer for \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\"" May 17 00:36:45.162655 env[1734]: time="2025-05-17T00:36:45.162599194Z" level=info msg="RemoveContainer for \"356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50\" returns successfully" May 17 00:36:45.208782 kubelet[2085]: I0517 00:36:45.208731 2085 memory_manager.go:355] "RemoveStaleState removing state" podUID="949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" containerName="mount-cgroup" May 17 00:36:45.225973 systemd[1]: Created slice kubepods-burstable-podcd36eac3_c7d5_4f6f_92c1_10fc90e48649.slice. May 17 00:36:45.297849 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2763955085.mount: Deactivated successfully. May 17 00:36:45.349509 kubelet[2085]: I0517 00:36:45.349459 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-cilium-run\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349709 kubelet[2085]: I0517 00:36:45.349536 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-cilium-cgroup\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349709 kubelet[2085]: I0517 00:36:45.349571 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-cni-path\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349709 kubelet[2085]: I0517 00:36:45.349618 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-etc-cni-netd\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349709 kubelet[2085]: I0517 00:36:45.349649 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-cilium-config-path\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349709 kubelet[2085]: I0517 00:36:45.349673 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-host-proc-sys-net\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349711 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-host-proc-sys-kernel\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349734 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gtdl\" (UniqueName: \"kubernetes.io/projected/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-kube-api-access-7gtdl\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349773 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-clustermesh-secrets\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349797 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-cilium-ipsec-secrets\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349822 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-hubble-tls\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349864 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-hostproc\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349888 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-lib-modules\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349912 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-xtables-lock\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.349967 kubelet[2085]: I0517 00:36:45.349957 2085 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/cd36eac3-c7d5-4f6f-92c1-10fc90e48649-bpf-maps\") pod \"cilium-qksv4\" (UID: \"cd36eac3-c7d5-4f6f-92c1-10fc90e48649\") " pod="kube-system/cilium-qksv4" May 17 00:36:45.538107 env[1734]: time="2025-05-17T00:36:45.538062046Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qksv4,Uid:cd36eac3-c7d5-4f6f-92c1-10fc90e48649,Namespace:kube-system,Attempt:0,}" May 17 00:36:45.573587 env[1734]: time="2025-05-17T00:36:45.573493884Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:36:45.573788 env[1734]: time="2025-05-17T00:36:45.573548994Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:36:45.573788 env[1734]: time="2025-05-17T00:36:45.573565655Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:36:45.573788 env[1734]: time="2025-05-17T00:36:45.573748187Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b pid=4035 runtime=io.containerd.runc.v2 May 17 00:36:45.606045 systemd[1]: Started cri-containerd-bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b.scope. May 17 00:36:45.654199 env[1734]: time="2025-05-17T00:36:45.654146811Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qksv4,Uid:cd36eac3-c7d5-4f6f-92c1-10fc90e48649,Namespace:kube-system,Attempt:0,} returns sandbox id \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\"" May 17 00:36:45.657632 env[1734]: time="2025-05-17T00:36:45.657582455Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:36:45.688392 env[1734]: time="2025-05-17T00:36:45.688323547Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b\"" May 17 00:36:45.690311 env[1734]: time="2025-05-17T00:36:45.690256455Z" level=info msg="StartContainer for \"e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b\"" May 17 00:36:45.715723 systemd[1]: Started cri-containerd-e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b.scope. May 17 00:36:45.723830 kubelet[2085]: E0517 00:36:45.723774 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:45.779956 env[1734]: time="2025-05-17T00:36:45.779907726Z" level=info msg="StartContainer for \"e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b\" returns successfully" May 17 00:36:45.802724 systemd[1]: cri-containerd-e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b.scope: Deactivated successfully. May 17 00:36:45.897806 env[1734]: time="2025-05-17T00:36:45.897747490Z" level=info msg="shim disconnected" id=e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b May 17 00:36:45.898308 env[1734]: time="2025-05-17T00:36:45.898266240Z" level=warning msg="cleaning up after shim disconnected" id=e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b namespace=k8s.io May 17 00:36:45.898414 env[1734]: time="2025-05-17T00:36:45.898398879Z" level=info msg="cleaning up dead shim" May 17 00:36:45.930976 env[1734]: time="2025-05-17T00:36:45.930924009Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:45Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4122 runtime=io.containerd.runc.v2\n" May 17 00:36:46.035979 kubelet[2085]: I0517 00:36:46.035919 2085 setters.go:602] "Node became not ready" node="172.31.23.93" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-17T00:36:46Z","lastTransitionTime":"2025-05-17T00:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 17 00:36:46.157900 env[1734]: time="2025-05-17T00:36:46.157500979Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 17 00:36:46.214066 env[1734]: time="2025-05-17T00:36:46.214000519Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f\"" May 17 00:36:46.215170 env[1734]: time="2025-05-17T00:36:46.215133534Z" level=info msg="StartContainer for \"973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f\"" May 17 00:36:46.267169 systemd[1]: Started cri-containerd-973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f.scope. May 17 00:36:46.313876 env[1734]: time="2025-05-17T00:36:46.313812882Z" level=info msg="StartContainer for \"973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f\" returns successfully" May 17 00:36:46.332563 systemd[1]: cri-containerd-973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f.scope: Deactivated successfully. May 17 00:36:46.339557 env[1734]: time="2025-05-17T00:36:46.339500444Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:46.344383 env[1734]: time="2025-05-17T00:36:46.344259524Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:46.347911 env[1734]: time="2025-05-17T00:36:46.347852312Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:36:46.348209 env[1734]: time="2025-05-17T00:36:46.348165394Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" May 17 00:36:46.351798 env[1734]: time="2025-05-17T00:36:46.351753724Z" level=info msg="CreateContainer within sandbox \"939439b197db49b03fed25c174023c334a8a8b2b948a730032e48b597bcf5ce4\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 17 00:36:46.413848 env[1734]: time="2025-05-17T00:36:46.413179407Z" level=info msg="CreateContainer within sandbox \"939439b197db49b03fed25c174023c334a8a8b2b948a730032e48b597bcf5ce4\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4\"" May 17 00:36:46.414379 env[1734]: time="2025-05-17T00:36:46.414348962Z" level=info msg="StartContainer for \"0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4\"" May 17 00:36:46.436075 systemd[1]: Started cri-containerd-0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4.scope. May 17 00:36:46.450366 env[1734]: time="2025-05-17T00:36:46.450304909Z" level=info msg="shim disconnected" id=973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f May 17 00:36:46.450808 env[1734]: time="2025-05-17T00:36:46.450768260Z" level=warning msg="cleaning up after shim disconnected" id=973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f namespace=k8s.io May 17 00:36:46.450958 env[1734]: time="2025-05-17T00:36:46.450939185Z" level=info msg="cleaning up dead shim" May 17 00:36:46.474205 env[1734]: time="2025-05-17T00:36:46.474137622Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:46Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4210 runtime=io.containerd.runc.v2\n" May 17 00:36:46.485372 env[1734]: time="2025-05-17T00:36:46.485232273Z" level=info msg="StartContainer for \"0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4\" returns successfully" May 17 00:36:46.724219 kubelet[2085]: E0517 00:36:46.724152 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:46.915157 kubelet[2085]: I0517 00:36:46.915121 2085 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="949e3c4e-3b61-4f3c-82e3-a0d23d7722fd" path="/var/lib/kubelet/pods/949e3c4e-3b61-4f3c-82e3-a0d23d7722fd/volumes" May 17 00:36:47.029388 kubelet[2085]: W0517 00:36:47.029231 2085 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod949e3c4e_3b61_4f3c_82e3_a0d23d7722fd.slice/cri-containerd-356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50.scope WatchSource:0}: container "356900a271a0dc63f8fb7258e32cb284bfc141ae259599ffa3daebd4b9216a50" in namespace "k8s.io": not found May 17 00:36:47.164310 env[1734]: time="2025-05-17T00:36:47.164238200Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 17 00:36:47.188398 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3022550645.mount: Deactivated successfully. May 17 00:36:47.195441 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2977996176.mount: Deactivated successfully. May 17 00:36:47.197729 kubelet[2085]: I0517 00:36:47.197654 2085 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-llz5d" podStartSLOduration=1.500536387 podStartE2EDuration="4.197631835s" podCreationTimestamp="2025-05-17 00:36:43 +0000 UTC" firstStartedPulling="2025-05-17 00:36:43.652776752 +0000 UTC m=+69.685704971" lastFinishedPulling="2025-05-17 00:36:46.3498722 +0000 UTC m=+72.382800419" observedRunningTime="2025-05-17 00:36:47.171975766 +0000 UTC m=+73.204903987" watchObservedRunningTime="2025-05-17 00:36:47.197631835 +0000 UTC m=+73.230560061" May 17 00:36:47.202021 env[1734]: time="2025-05-17T00:36:47.201958402Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58\"" May 17 00:36:47.202765 env[1734]: time="2025-05-17T00:36:47.202731160Z" level=info msg="StartContainer for \"83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58\"" May 17 00:36:47.251733 systemd[1]: Started cri-containerd-83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58.scope. May 17 00:36:47.320599 env[1734]: time="2025-05-17T00:36:47.320480262Z" level=info msg="StartContainer for \"83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58\" returns successfully" May 17 00:36:47.330032 systemd[1]: cri-containerd-83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58.scope: Deactivated successfully. May 17 00:36:47.366281 env[1734]: time="2025-05-17T00:36:47.366226239Z" level=info msg="shim disconnected" id=83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58 May 17 00:36:47.366281 env[1734]: time="2025-05-17T00:36:47.366301117Z" level=warning msg="cleaning up after shim disconnected" id=83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58 namespace=k8s.io May 17 00:36:47.366753 env[1734]: time="2025-05-17T00:36:47.366316178Z" level=info msg="cleaning up dead shim" May 17 00:36:47.376765 env[1734]: time="2025-05-17T00:36:47.376703675Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:47Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4280 runtime=io.containerd.runc.v2\n" May 17 00:36:47.724811 kubelet[2085]: E0517 00:36:47.724754 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:48.168561 env[1734]: time="2025-05-17T00:36:48.168431408Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 17 00:36:48.198886 env[1734]: time="2025-05-17T00:36:48.198815165Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a\"" May 17 00:36:48.199518 env[1734]: time="2025-05-17T00:36:48.199477088Z" level=info msg="StartContainer for \"6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a\"" May 17 00:36:48.230961 systemd[1]: Started cri-containerd-6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a.scope. May 17 00:36:48.265325 systemd[1]: cri-containerd-6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a.scope: Deactivated successfully. May 17 00:36:48.269401 env[1734]: time="2025-05-17T00:36:48.269324953Z" level=info msg="StartContainer for \"6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a\" returns successfully" May 17 00:36:48.308398 env[1734]: time="2025-05-17T00:36:48.308338186Z" level=info msg="shim disconnected" id=6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a May 17 00:36:48.308398 env[1734]: time="2025-05-17T00:36:48.308399425Z" level=warning msg="cleaning up after shim disconnected" id=6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a namespace=k8s.io May 17 00:36:48.308660 env[1734]: time="2025-05-17T00:36:48.308413140Z" level=info msg="cleaning up dead shim" May 17 00:36:48.318223 env[1734]: time="2025-05-17T00:36:48.318171138Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:36:48Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4336 runtime=io.containerd.runc.v2\n" May 17 00:36:48.462343 systemd[1]: run-containerd-runc-k8s.io-6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a-runc.7vxVe1.mount: Deactivated successfully. May 17 00:36:48.462469 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a-rootfs.mount: Deactivated successfully. May 17 00:36:48.725585 kubelet[2085]: E0517 00:36:48.725436 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:49.174563 env[1734]: time="2025-05-17T00:36:49.174512303Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 17 00:36:49.206754 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount320196619.mount: Deactivated successfully. May 17 00:36:49.219038 env[1734]: time="2025-05-17T00:36:49.218979037Z" level=info msg="CreateContainer within sandbox \"bdc209f36f7900667e74091317ebebf395e84448f6bbbb5101a138daec782f4b\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7\"" May 17 00:36:49.219975 env[1734]: time="2025-05-17T00:36:49.219881853Z" level=info msg="StartContainer for \"24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7\"" May 17 00:36:49.245086 systemd[1]: Started cri-containerd-24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7.scope. May 17 00:36:49.287187 env[1734]: time="2025-05-17T00:36:49.287098029Z" level=info msg="StartContainer for \"24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7\" returns successfully" May 17 00:36:49.726483 kubelet[2085]: E0517 00:36:49.726431 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:49.859417 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) May 17 00:36:50.143704 kubelet[2085]: W0517 00:36:50.143569 2085 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd36eac3_c7d5_4f6f_92c1_10fc90e48649.slice/cri-containerd-e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b.scope WatchSource:0}: task e194a1c47c759e638ba2308122183e75c8adc8f87684bc2a68de95819b86fe8b not found: not found May 17 00:36:50.400412 systemd[1]: run-containerd-runc-k8s.io-24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7-runc.z7nujI.mount: Deactivated successfully. May 17 00:36:50.727315 kubelet[2085]: E0517 00:36:50.727129 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:51.727963 kubelet[2085]: E0517 00:36:51.727908 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:52.575441 systemd[1]: run-containerd-runc-k8s.io-24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7-runc.IvKkX5.mount: Deactivated successfully. May 17 00:36:52.728843 kubelet[2085]: E0517 00:36:52.728753 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:52.767042 (udev-worker)[4921]: Network interface NamePolicy= disabled on kernel command line. May 17 00:36:52.767740 (udev-worker)[4429]: Network interface NamePolicy= disabled on kernel command line. May 17 00:36:52.769970 systemd-networkd[1459]: lxc_health: Link UP May 17 00:36:52.788408 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready May 17 00:36:52.787634 systemd-networkd[1459]: lxc_health: Gained carrier May 17 00:36:53.259271 kubelet[2085]: W0517 00:36:53.258433 2085 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd36eac3_c7d5_4f6f_92c1_10fc90e48649.slice/cri-containerd-973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f.scope WatchSource:0}: task 973954b722f0497115bfc641c8937dc0c3ec97d2e9cada35ad41ac64d5d92b4f not found: not found May 17 00:36:53.561358 kubelet[2085]: I0517 00:36:53.561191 2085 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-qksv4" podStartSLOduration=8.561167877 podStartE2EDuration="8.561167877s" podCreationTimestamp="2025-05-17 00:36:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:36:50.198965015 +0000 UTC m=+76.231893241" watchObservedRunningTime="2025-05-17 00:36:53.561167877 +0000 UTC m=+79.594096100" May 17 00:36:53.729953 kubelet[2085]: E0517 00:36:53.729902 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:54.043450 systemd-networkd[1459]: lxc_health: Gained IPv6LL May 17 00:36:54.669485 kubelet[2085]: E0517 00:36:54.669417 2085 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:54.730663 kubelet[2085]: E0517 00:36:54.730614 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:54.837167 systemd[1]: run-containerd-runc-k8s.io-24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7-runc.37lThF.mount: Deactivated successfully. May 17 00:36:55.732154 kubelet[2085]: E0517 00:36:55.732106 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:56.392800 kubelet[2085]: W0517 00:36:56.392743 2085 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd36eac3_c7d5_4f6f_92c1_10fc90e48649.slice/cri-containerd-83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58.scope WatchSource:0}: task 83c776d08428add02d7a8706cd1c32ade104d089b32c6204c5eb2b529a3c8e58 not found: not found May 17 00:36:56.733672 kubelet[2085]: E0517 00:36:56.733619 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:57.058561 systemd[1]: run-containerd-runc-k8s.io-24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7-runc.1nSs99.mount: Deactivated successfully. May 17 00:36:57.734886 kubelet[2085]: E0517 00:36:57.734830 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:58.735820 kubelet[2085]: E0517 00:36:58.735740 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:36:59.337517 systemd[1]: run-containerd-runc-k8s.io-24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7-runc.90F3W1.mount: Deactivated successfully. May 17 00:36:59.501645 kubelet[2085]: W0517 00:36:59.501598 2085 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd36eac3_c7d5_4f6f_92c1_10fc90e48649.slice/cri-containerd-6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a.scope WatchSource:0}: task 6885f4b7999ddeb908af7dbe1a5191c82f1df4076c0f4158d5311ae36b967f0a not found: not found May 17 00:36:59.736231 kubelet[2085]: E0517 00:36:59.736167 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:00.737311 kubelet[2085]: E0517 00:37:00.737145 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:01.612846 systemd[1]: run-containerd-runc-k8s.io-24c6eba30c22092814ed7b4516fc9574297e9ed02b2df982621758602b039aa7-runc.KENfi2.mount: Deactivated successfully. May 17 00:37:01.740049 kubelet[2085]: E0517 00:37:01.739984 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:02.740338 kubelet[2085]: E0517 00:37:02.740215 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:03.741451 kubelet[2085]: E0517 00:37:03.741384 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:04.742576 kubelet[2085]: E0517 00:37:04.742500 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:05.743009 kubelet[2085]: E0517 00:37:05.742946 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:06.743899 kubelet[2085]: E0517 00:37:06.743829 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:07.744753 kubelet[2085]: E0517 00:37:07.744700 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:08.744975 kubelet[2085]: E0517 00:37:08.744916 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:09.745790 kubelet[2085]: E0517 00:37:09.745719 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:10.746345 kubelet[2085]: E0517 00:37:10.746258 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:11.747437 kubelet[2085]: E0517 00:37:11.747381 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:12.748697 kubelet[2085]: E0517 00:37:12.748482 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:13.749443 kubelet[2085]: E0517 00:37:13.749368 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:14.668816 kubelet[2085]: E0517 00:37:14.668735 2085 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:14.750120 kubelet[2085]: E0517 00:37:14.750057 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:15.750958 kubelet[2085]: E0517 00:37:15.750884 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:16.751944 kubelet[2085]: E0517 00:37:16.751871 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:17.752764 kubelet[2085]: E0517 00:37:17.752704 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:18.603123 systemd[1]: cri-containerd-0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4.scope: Deactivated successfully. May 17 00:37:18.630527 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4-rootfs.mount: Deactivated successfully. May 17 00:37:18.657930 env[1734]: time="2025-05-17T00:37:18.657867201Z" level=info msg="shim disconnected" id=0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4 May 17 00:37:18.657930 env[1734]: time="2025-05-17T00:37:18.657926196Z" level=warning msg="cleaning up after shim disconnected" id=0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4 namespace=k8s.io May 17 00:37:18.657930 env[1734]: time="2025-05-17T00:37:18.657938007Z" level=info msg="cleaning up dead shim" May 17 00:37:18.667985 env[1734]: time="2025-05-17T00:37:18.667928421Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:37:18Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5061 runtime=io.containerd.runc.v2\n" May 17 00:37:18.753604 kubelet[2085]: E0517 00:37:18.753554 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:19.243591 kubelet[2085]: I0517 00:37:19.243541 2085 scope.go:117] "RemoveContainer" containerID="0e147711865c5e13034aae6d39681083cd3c3337c3ed3eb9dbe798d4e458bbb4" May 17 00:37:19.246048 env[1734]: time="2025-05-17T00:37:19.245996566Z" level=info msg="CreateContainer within sandbox \"939439b197db49b03fed25c174023c334a8a8b2b948a730032e48b597bcf5ce4\" for container &ContainerMetadata{Name:cilium-operator,Attempt:1,}" May 17 00:37:19.268105 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4006106455.mount: Deactivated successfully. May 17 00:37:19.280778 env[1734]: time="2025-05-17T00:37:19.280718371Z" level=info msg="CreateContainer within sandbox \"939439b197db49b03fed25c174023c334a8a8b2b948a730032e48b597bcf5ce4\" for &ContainerMetadata{Name:cilium-operator,Attempt:1,} returns container id \"1551793a2bfba0859d8ef8d31cf12f55de847bdaeaa73e3f06e5ee28b9ce29cd\"" May 17 00:37:19.281458 env[1734]: time="2025-05-17T00:37:19.281424230Z" level=info msg="StartContainer for \"1551793a2bfba0859d8ef8d31cf12f55de847bdaeaa73e3f06e5ee28b9ce29cd\"" May 17 00:37:19.304080 systemd[1]: Started cri-containerd-1551793a2bfba0859d8ef8d31cf12f55de847bdaeaa73e3f06e5ee28b9ce29cd.scope. May 17 00:37:19.352833 env[1734]: time="2025-05-17T00:37:19.352769054Z" level=info msg="StartContainer for \"1551793a2bfba0859d8ef8d31cf12f55de847bdaeaa73e3f06e5ee28b9ce29cd\" returns successfully" May 17 00:37:19.754459 kubelet[2085]: E0517 00:37:19.754392 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:20.755639 kubelet[2085]: E0517 00:37:20.755582 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:21.756512 kubelet[2085]: E0517 00:37:21.756254 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:22.757045 kubelet[2085]: E0517 00:37:22.756992 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:23.757317 kubelet[2085]: E0517 00:37:23.757256 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:24.758405 kubelet[2085]: E0517 00:37:24.758346 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:25.759080 kubelet[2085]: E0517 00:37:25.759019 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:26.760386 kubelet[2085]: E0517 00:37:26.760163 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:27.056721 kubelet[2085]: E0517 00:37:27.056431 2085 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" May 17 00:37:27.761157 kubelet[2085]: E0517 00:37:27.761099 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:28.761967 kubelet[2085]: E0517 00:37:28.761906 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:29.762534 kubelet[2085]: E0517 00:37:29.762469 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:30.763495 kubelet[2085]: E0517 00:37:30.763434 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:31.764636 kubelet[2085]: E0517 00:37:31.764594 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:32.765828 kubelet[2085]: E0517 00:37:32.765768 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:33.766918 kubelet[2085]: E0517 00:37:33.766858 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:34.668754 kubelet[2085]: E0517 00:37:34.668687 2085 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:34.721880 env[1734]: time="2025-05-17T00:37:34.721834209Z" level=info msg="StopPodSandbox for \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\"" May 17 00:37:34.722476 env[1734]: time="2025-05-17T00:37:34.721957484Z" level=info msg="TearDown network for sandbox \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" successfully" May 17 00:37:34.722476 env[1734]: time="2025-05-17T00:37:34.722007917Z" level=info msg="StopPodSandbox for \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" returns successfully" May 17 00:37:34.722607 env[1734]: time="2025-05-17T00:37:34.722571473Z" level=info msg="RemovePodSandbox for \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\"" May 17 00:37:34.722675 env[1734]: time="2025-05-17T00:37:34.722614616Z" level=info msg="Forcibly stopping sandbox \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\"" May 17 00:37:34.722740 env[1734]: time="2025-05-17T00:37:34.722719769Z" level=info msg="TearDown network for sandbox \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" successfully" May 17 00:37:34.730687 env[1734]: time="2025-05-17T00:37:34.730636684Z" level=info msg="RemovePodSandbox \"093501816d03975a119b290edc48b2b21e54a3509702ce6b1ada586b8705ee15\" returns successfully" May 17 00:37:34.731322 env[1734]: time="2025-05-17T00:37:34.731254331Z" level=info msg="StopPodSandbox for \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\"" May 17 00:37:34.731461 env[1734]: time="2025-05-17T00:37:34.731352521Z" level=info msg="TearDown network for sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" successfully" May 17 00:37:34.731461 env[1734]: time="2025-05-17T00:37:34.731386477Z" level=info msg="StopPodSandbox for \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" returns successfully" May 17 00:37:34.731667 env[1734]: time="2025-05-17T00:37:34.731643670Z" level=info msg="RemovePodSandbox for \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\"" May 17 00:37:34.731743 env[1734]: time="2025-05-17T00:37:34.731666750Z" level=info msg="Forcibly stopping sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\"" May 17 00:37:34.731743 env[1734]: time="2025-05-17T00:37:34.731723377Z" level=info msg="TearDown network for sandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" successfully" May 17 00:37:34.737643 env[1734]: time="2025-05-17T00:37:34.737580848Z" level=info msg="RemovePodSandbox \"22e29f65cbe3ecf7516d2147b3edb435f54c0299eb79297f6d439443273859c7\" returns successfully" May 17 00:37:34.767935 kubelet[2085]: E0517 00:37:34.767801 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:35.768773 kubelet[2085]: E0517 00:37:35.768690 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:36.768898 kubelet[2085]: E0517 00:37:36.768850 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:37.057084 kubelet[2085]: E0517 00:37:37.056722 2085 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" May 17 00:37:37.769992 kubelet[2085]: E0517 00:37:37.769943 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:38.771095 kubelet[2085]: E0517 00:37:38.771034 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:39.772349 kubelet[2085]: E0517 00:37:39.772270 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:40.773163 kubelet[2085]: E0517 00:37:40.773100 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:41.774071 kubelet[2085]: E0517 00:37:41.774009 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:42.627005 kubelet[2085]: E0517 00:37:42.626880 2085 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.29.206:6443/api/v1/namespaces/kube-system/events\": unexpected EOF" event="&Event{ObjectMeta:{cilium-operator-6c4d7847fc-llz5d.184029754fabb2f1 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:cilium-operator-6c4d7847fc-llz5d,UID:205b87b5-4b0b-4d10-b5ff-bb9ffbd12998,APIVersion:v1,ResourceVersion:866,FieldPath:spec.containers{cilium-operator},},Reason:Pulled,Message:Container image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" already present on machine,Source:EventSource{Component:kubelet,Host:172.31.23.93,},FirstTimestamp:2025-05-17 00:37:19.244317425 +0000 UTC m=+105.277245644,LastTimestamp:2025-05-17 00:37:19.244317425 +0000 UTC m=+105.277245644,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:172.31.23.93,}" May 17 00:37:42.629082 kubelet[2085]: E0517 00:37:42.627342 2085 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": unexpected EOF" May 17 00:37:42.640872 kubelet[2085]: E0517 00:37:42.640813 2085 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": read tcp 172.31.23.93:48358->172.31.29.206:6443: read: connection reset by peer" May 17 00:37:42.641836 kubelet[2085]: E0517 00:37:42.641788 2085 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": dial tcp 172.31.29.206:6443: connect: connection refused" May 17 00:37:42.641969 kubelet[2085]: I0517 00:37:42.641843 2085 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" May 17 00:37:42.642646 kubelet[2085]: E0517 00:37:42.642610 2085 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": dial tcp 172.31.29.206:6443: connect: connection refused" interval="200ms" May 17 00:37:42.775264 kubelet[2085]: E0517 00:37:42.775206 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:42.844815 kubelet[2085]: E0517 00:37:42.844751 2085 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": dial tcp 172.31.29.206:6443: connect: connection refused" interval="400ms" May 17 00:37:43.246780 kubelet[2085]: E0517 00:37:43.246714 2085 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": dial tcp 172.31.29.206:6443: connect: connection refused" interval="800ms" May 17 00:37:43.628964 kubelet[2085]: I0517 00:37:43.628561 2085 status_manager.go:890] "Failed to get status for pod" podUID="205b87b5-4b0b-4d10-b5ff-bb9ffbd12998" pod="kube-system/cilium-operator-6c4d7847fc-llz5d" err="Get \"https://172.31.29.206:6443/api/v1/namespaces/kube-system/pods/cilium-operator-6c4d7847fc-llz5d\": dial tcp 172.31.29.206:6443: connect: connection refused - error from a previous attempt: unexpected EOF" May 17 00:37:43.629349 kubelet[2085]: I0517 00:37:43.629317 2085 status_manager.go:890] "Failed to get status for pod" podUID="205b87b5-4b0b-4d10-b5ff-bb9ffbd12998" pod="kube-system/cilium-operator-6c4d7847fc-llz5d" err="Get \"https://172.31.29.206:6443/api/v1/namespaces/kube-system/pods/cilium-operator-6c4d7847fc-llz5d\": dial tcp 172.31.29.206:6443: connect: connection refused" May 17 00:37:43.630208 kubelet[2085]: I0517 00:37:43.630159 2085 status_manager.go:890] "Failed to get status for pod" podUID="205b87b5-4b0b-4d10-b5ff-bb9ffbd12998" pod="kube-system/cilium-operator-6c4d7847fc-llz5d" err="Get \"https://172.31.29.206:6443/api/v1/namespaces/kube-system/pods/cilium-operator-6c4d7847fc-llz5d\": dial tcp 172.31.29.206:6443: connect: connection refused" May 17 00:37:43.776376 kubelet[2085]: E0517 00:37:43.776311 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:44.776666 kubelet[2085]: E0517 00:37:44.776586 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:45.776847 kubelet[2085]: E0517 00:37:45.776772 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:46.777774 kubelet[2085]: E0517 00:37:46.777699 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:47.778921 kubelet[2085]: E0517 00:37:47.778865 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:48.780086 kubelet[2085]: E0517 00:37:48.779982 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:49.781263 kubelet[2085]: E0517 00:37:49.781198 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:50.781413 kubelet[2085]: E0517 00:37:50.781366 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:51.782135 kubelet[2085]: E0517 00:37:51.782069 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:52.782803 kubelet[2085]: E0517 00:37:52.782741 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:53.782962 kubelet[2085]: E0517 00:37:53.782900 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:54.047812 kubelet[2085]: E0517 00:37:54.047654 2085 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": context deadline exceeded" interval="1.6s" May 17 00:37:54.668578 kubelet[2085]: E0517 00:37:54.668508 2085 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:54.784164 kubelet[2085]: E0517 00:37:54.784097 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:55.785241 kubelet[2085]: E0517 00:37:55.785163 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:56.786082 kubelet[2085]: E0517 00:37:56.786022 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:57.786425 kubelet[2085]: E0517 00:37:57.786366 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:58.787305 kubelet[2085]: E0517 00:37:58.787221 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:37:59.788063 kubelet[2085]: E0517 00:37:59.788005 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:00.788478 kubelet[2085]: E0517 00:38:00.788413 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:01.789769 kubelet[2085]: E0517 00:38:01.789644 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:02.790879 kubelet[2085]: E0517 00:38:02.790816 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:03.791081 kubelet[2085]: E0517 00:38:03.790993 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:04.792146 kubelet[2085]: E0517 00:38:04.792085 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:05.649479 kubelet[2085]: E0517 00:38:05.649414 2085 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.206:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/172.31.23.93?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" interval="3.2s" May 17 00:38:05.793467 kubelet[2085]: E0517 00:38:05.793396 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:06.793636 kubelet[2085]: E0517 00:38:06.793572 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:07.793992 kubelet[2085]: E0517 00:38:07.793932 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" May 17 00:38:08.794887 kubelet[2085]: E0517 00:38:08.794633 2085 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"