Sep 13 00:47:43.018443 kernel: Linux version 5.15.192-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Fri Sep 12 23:13:49 -00 2025 Sep 13 00:47:43.018476 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=65d14b740db9e581daa1d0206188b16d2f1a39e5c5e0878b6855323cd7c584ec Sep 13 00:47:43.018494 kernel: BIOS-provided physical RAM map: Sep 13 00:47:43.018506 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 13 00:47:43.018517 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable Sep 13 00:47:43.018527 kernel: BIOS-e820: [mem 0x00000000786ce000-0x000000007894dfff] reserved Sep 13 00:47:43.018539 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Sep 13 00:47:43.018549 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Sep 13 00:47:43.018561 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable Sep 13 00:47:43.018570 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Sep 13 00:47:43.018580 kernel: NX (Execute Disable) protection: active Sep 13 00:47:43.018590 kernel: e820: update [mem 0x76813018-0x7681be57] usable ==> usable Sep 13 00:47:43.018600 kernel: e820: update [mem 0x76813018-0x7681be57] usable ==> usable Sep 13 00:47:43.018610 kernel: extended physical RAM map: Sep 13 00:47:43.018624 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 13 00:47:43.018635 kernel: reserve setup_data: [mem 0x0000000000100000-0x0000000076813017] usable Sep 13 00:47:43.018645 kernel: reserve setup_data: [mem 0x0000000076813018-0x000000007681be57] usable Sep 13 00:47:43.018655 kernel: reserve setup_data: [mem 0x000000007681be58-0x00000000786cdfff] usable Sep 13 00:47:43.018666 kernel: reserve setup_data: [mem 0x00000000786ce000-0x000000007894dfff] reserved Sep 13 00:47:43.018677 kernel: reserve setup_data: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Sep 13 00:47:43.018687 kernel: reserve setup_data: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Sep 13 00:47:43.018698 kernel: reserve setup_data: [mem 0x00000000789de000-0x000000007c97bfff] usable Sep 13 00:47:43.018709 kernel: reserve setup_data: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Sep 13 00:47:43.018719 kernel: efi: EFI v2.70 by EDK II Sep 13 00:47:43.018732 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77004a98 Sep 13 00:47:43.018743 kernel: SMBIOS 2.7 present. Sep 13 00:47:43.018754 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Sep 13 00:47:43.018764 kernel: Hypervisor detected: KVM Sep 13 00:47:43.018775 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 13 00:47:43.018785 kernel: kvm-clock: cpu 0, msr 2019f001, primary cpu clock Sep 13 00:47:43.018796 kernel: kvm-clock: using sched offset of 4454402064 cycles Sep 13 00:47:43.018808 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 13 00:47:43.018819 kernel: tsc: Detected 2499.996 MHz processor Sep 13 00:47:43.018830 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 13 00:47:43.019196 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 13 00:47:43.019217 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 Sep 13 00:47:43.019230 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 13 00:47:43.019243 kernel: Using GB pages for direct mapping Sep 13 00:47:43.019256 kernel: Secure boot disabled Sep 13 00:47:43.019270 kernel: ACPI: Early table checksum verification disabled Sep 13 00:47:43.019289 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) Sep 13 00:47:43.019303 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) Sep 13 00:47:43.019319 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Sep 13 00:47:43.019333 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Sep 13 00:47:43.019347 kernel: ACPI: FACS 0x00000000789D0000 000040 Sep 13 00:47:43.019361 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Sep 13 00:47:43.019376 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Sep 13 00:47:43.019389 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Sep 13 00:47:43.019416 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Sep 13 00:47:43.019432 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Sep 13 00:47:43.019447 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Sep 13 00:47:43.019461 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Sep 13 00:47:43.019475 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) Sep 13 00:47:43.019489 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] Sep 13 00:47:43.019503 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] Sep 13 00:47:43.019517 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] Sep 13 00:47:43.019532 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] Sep 13 00:47:43.019546 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] Sep 13 00:47:43.019562 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] Sep 13 00:47:43.019576 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] Sep 13 00:47:43.019590 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] Sep 13 00:47:43.019604 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] Sep 13 00:47:43.019618 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] Sep 13 00:47:43.019632 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] Sep 13 00:47:43.019646 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 13 00:47:43.019660 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 13 00:47:43.019674 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Sep 13 00:47:43.019691 kernel: NUMA: Initialized distance table, cnt=1 Sep 13 00:47:43.019705 kernel: NODE_DATA(0) allocated [mem 0x7a8ef000-0x7a8f4fff] Sep 13 00:47:43.019719 kernel: Zone ranges: Sep 13 00:47:43.019734 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 13 00:47:43.019748 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] Sep 13 00:47:43.019762 kernel: Normal empty Sep 13 00:47:43.019776 kernel: Movable zone start for each node Sep 13 00:47:43.019790 kernel: Early memory node ranges Sep 13 00:47:43.019804 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 13 00:47:43.019821 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] Sep 13 00:47:43.019835 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] Sep 13 00:47:43.019849 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] Sep 13 00:47:43.019863 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 13 00:47:43.019878 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 13 00:47:43.019892 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Sep 13 00:47:43.019906 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges Sep 13 00:47:43.019920 kernel: ACPI: PM-Timer IO Port: 0xb008 Sep 13 00:47:43.019935 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 13 00:47:43.019951 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Sep 13 00:47:43.019965 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 13 00:47:43.019980 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 13 00:47:43.019991 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 13 00:47:43.020001 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 13 00:47:43.020012 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 13 00:47:43.020023 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 13 00:47:43.020036 kernel: TSC deadline timer available Sep 13 00:47:43.020048 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Sep 13 00:47:43.020062 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices Sep 13 00:47:43.020074 kernel: Booting paravirtualized kernel on KVM Sep 13 00:47:43.020087 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 13 00:47:43.020099 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 Sep 13 00:47:43.020112 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u1048576 Sep 13 00:47:43.020124 kernel: pcpu-alloc: s188696 r8192 d32488 u1048576 alloc=1*2097152 Sep 13 00:47:43.020137 kernel: pcpu-alloc: [0] 0 1 Sep 13 00:47:43.020149 kernel: kvm-guest: stealtime: cpu 0, msr 7a41c0c0 Sep 13 00:47:43.020161 kernel: kvm-guest: PV spinlocks enabled Sep 13 00:47:43.020176 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 13 00:47:43.020187 kernel: Built 1 zonelists, mobility grouping on. Total pages: 501318 Sep 13 00:47:43.020200 kernel: Policy zone: DMA32 Sep 13 00:47:43.020214 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=65d14b740db9e581daa1d0206188b16d2f1a39e5c5e0878b6855323cd7c584ec Sep 13 00:47:43.020227 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 13 00:47:43.020239 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 13 00:47:43.020251 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 13 00:47:43.020263 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 13 00:47:43.020280 kernel: Memory: 1876640K/2037804K available (12295K kernel code, 2276K rwdata, 13732K rodata, 47492K init, 4088K bss, 160904K reserved, 0K cma-reserved) Sep 13 00:47:43.020292 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 13 00:47:43.020304 kernel: Kernel/User page tables isolation: enabled Sep 13 00:47:43.020316 kernel: ftrace: allocating 34614 entries in 136 pages Sep 13 00:47:43.020328 kernel: ftrace: allocated 136 pages with 2 groups Sep 13 00:47:43.020340 kernel: rcu: Hierarchical RCU implementation. Sep 13 00:47:43.020353 kernel: rcu: RCU event tracing is enabled. Sep 13 00:47:43.020378 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 13 00:47:43.020402 kernel: Rude variant of Tasks RCU enabled. Sep 13 00:47:43.020415 kernel: Tracing variant of Tasks RCU enabled. Sep 13 00:47:43.020428 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 13 00:47:43.020441 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 13 00:47:43.020456 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Sep 13 00:47:43.020468 kernel: random: crng init done Sep 13 00:47:43.020480 kernel: Console: colour dummy device 80x25 Sep 13 00:47:43.020493 kernel: printk: console [tty0] enabled Sep 13 00:47:43.020506 kernel: printk: console [ttyS0] enabled Sep 13 00:47:43.020518 kernel: ACPI: Core revision 20210730 Sep 13 00:47:43.020532 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Sep 13 00:47:43.020547 kernel: APIC: Switch to symmetric I/O mode setup Sep 13 00:47:43.020560 kernel: x2apic enabled Sep 13 00:47:43.020573 kernel: Switched APIC routing to physical x2apic. Sep 13 00:47:43.020586 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Sep 13 00:47:43.020599 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499996) Sep 13 00:47:43.020612 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Sep 13 00:47:43.020625 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Sep 13 00:47:43.020641 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 13 00:47:43.020653 kernel: Spectre V2 : Mitigation: Retpolines Sep 13 00:47:43.020666 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 13 00:47:43.020680 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Sep 13 00:47:43.020693 kernel: RETBleed: Vulnerable Sep 13 00:47:43.020706 kernel: Speculative Store Bypass: Vulnerable Sep 13 00:47:43.020719 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Sep 13 00:47:43.020731 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 13 00:47:43.020744 kernel: GDS: Unknown: Dependent on hypervisor status Sep 13 00:47:43.020757 kernel: active return thunk: its_return_thunk Sep 13 00:47:43.020769 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 13 00:47:43.020785 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 13 00:47:43.020799 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 13 00:47:43.020811 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 13 00:47:43.020824 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Sep 13 00:47:43.020837 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Sep 13 00:47:43.020849 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Sep 13 00:47:43.020862 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Sep 13 00:47:43.020875 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Sep 13 00:47:43.020887 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Sep 13 00:47:43.020900 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 13 00:47:43.020913 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Sep 13 00:47:43.020927 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Sep 13 00:47:43.020940 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Sep 13 00:47:43.020953 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Sep 13 00:47:43.020966 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Sep 13 00:47:43.020979 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Sep 13 00:47:43.020992 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Sep 13 00:47:43.021005 kernel: Freeing SMP alternatives memory: 32K Sep 13 00:47:43.021017 kernel: pid_max: default: 32768 minimum: 301 Sep 13 00:47:43.021029 kernel: LSM: Security Framework initializing Sep 13 00:47:43.021042 kernel: SELinux: Initializing. Sep 13 00:47:43.021055 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 13 00:47:43.021071 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 13 00:47:43.021084 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) Sep 13 00:47:43.021097 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Sep 13 00:47:43.021110 kernel: signal: max sigframe size: 3632 Sep 13 00:47:43.021122 kernel: rcu: Hierarchical SRCU implementation. Sep 13 00:47:43.021135 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 13 00:47:43.021147 kernel: smp: Bringing up secondary CPUs ... Sep 13 00:47:43.021223 kernel: x86: Booting SMP configuration: Sep 13 00:47:43.021238 kernel: .... node #0, CPUs: #1 Sep 13 00:47:43.021251 kernel: kvm-clock: cpu 1, msr 2019f041, secondary cpu clock Sep 13 00:47:43.021266 kernel: kvm-guest: stealtime: cpu 1, msr 7a51c0c0 Sep 13 00:47:43.021278 kernel: Transient Scheduler Attacks: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Sep 13 00:47:43.021294 kernel: Transient Scheduler Attacks: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 13 00:47:43.021307 kernel: smp: Brought up 1 node, 2 CPUs Sep 13 00:47:43.021319 kernel: smpboot: Max logical packages: 1 Sep 13 00:47:43.021334 kernel: smpboot: Total of 2 processors activated (9999.98 BogoMIPS) Sep 13 00:47:43.021348 kernel: devtmpfs: initialized Sep 13 00:47:43.021362 kernel: x86/mm: Memory block size: 128MB Sep 13 00:47:43.021379 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) Sep 13 00:47:43.021431 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 13 00:47:43.021444 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 13 00:47:43.021457 kernel: pinctrl core: initialized pinctrl subsystem Sep 13 00:47:43.021469 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 13 00:47:43.021481 kernel: audit: initializing netlink subsys (disabled) Sep 13 00:47:43.021495 kernel: audit: type=2000 audit(1757724463.941:1): state=initialized audit_enabled=0 res=1 Sep 13 00:47:43.021508 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 13 00:47:43.021521 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 13 00:47:43.021538 kernel: cpuidle: using governor menu Sep 13 00:47:43.021553 kernel: ACPI: bus type PCI registered Sep 13 00:47:43.021568 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 13 00:47:43.021582 kernel: dca service started, version 1.12.1 Sep 13 00:47:43.021597 kernel: PCI: Using configuration type 1 for base access Sep 13 00:47:43.021612 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 13 00:47:43.021627 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Sep 13 00:47:43.021642 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Sep 13 00:47:43.021657 kernel: ACPI: Added _OSI(Module Device) Sep 13 00:47:43.021675 kernel: ACPI: Added _OSI(Processor Device) Sep 13 00:47:43.021690 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 13 00:47:43.021705 kernel: ACPI: Added _OSI(Linux-Dell-Video) Sep 13 00:47:43.021720 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Sep 13 00:47:43.021735 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Sep 13 00:47:43.021749 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Sep 13 00:47:43.021764 kernel: ACPI: Interpreter enabled Sep 13 00:47:43.021779 kernel: ACPI: PM: (supports S0 S5) Sep 13 00:47:43.021795 kernel: ACPI: Using IOAPIC for interrupt routing Sep 13 00:47:43.021812 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 13 00:47:43.021827 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Sep 13 00:47:43.021842 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 13 00:47:43.022049 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Sep 13 00:47:43.022186 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. Sep 13 00:47:43.022206 kernel: acpiphp: Slot [3] registered Sep 13 00:47:43.022221 kernel: acpiphp: Slot [4] registered Sep 13 00:47:43.022236 kernel: acpiphp: Slot [5] registered Sep 13 00:47:43.022254 kernel: acpiphp: Slot [6] registered Sep 13 00:47:43.022269 kernel: acpiphp: Slot [7] registered Sep 13 00:47:43.022284 kernel: acpiphp: Slot [8] registered Sep 13 00:47:43.022298 kernel: acpiphp: Slot [9] registered Sep 13 00:47:43.022313 kernel: acpiphp: Slot [10] registered Sep 13 00:47:43.022328 kernel: acpiphp: Slot [11] registered Sep 13 00:47:43.022343 kernel: acpiphp: Slot [12] registered Sep 13 00:47:43.022358 kernel: acpiphp: Slot [13] registered Sep 13 00:47:43.022372 kernel: acpiphp: Slot [14] registered Sep 13 00:47:43.022389 kernel: acpiphp: Slot [15] registered Sep 13 00:47:43.022413 kernel: acpiphp: Slot [16] registered Sep 13 00:47:43.022425 kernel: acpiphp: Slot [17] registered Sep 13 00:47:43.028962 kernel: acpiphp: Slot [18] registered Sep 13 00:47:43.029001 kernel: acpiphp: Slot [19] registered Sep 13 00:47:43.029014 kernel: acpiphp: Slot [20] registered Sep 13 00:47:43.029029 kernel: acpiphp: Slot [21] registered Sep 13 00:47:43.029042 kernel: acpiphp: Slot [22] registered Sep 13 00:47:43.029055 kernel: acpiphp: Slot [23] registered Sep 13 00:47:43.029068 kernel: acpiphp: Slot [24] registered Sep 13 00:47:43.029086 kernel: acpiphp: Slot [25] registered Sep 13 00:47:43.029099 kernel: acpiphp: Slot [26] registered Sep 13 00:47:43.029112 kernel: acpiphp: Slot [27] registered Sep 13 00:47:43.029125 kernel: acpiphp: Slot [28] registered Sep 13 00:47:43.029138 kernel: acpiphp: Slot [29] registered Sep 13 00:47:43.029151 kernel: acpiphp: Slot [30] registered Sep 13 00:47:43.029164 kernel: acpiphp: Slot [31] registered Sep 13 00:47:43.029176 kernel: PCI host bridge to bus 0000:00 Sep 13 00:47:43.029347 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 13 00:47:43.029487 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 13 00:47:43.029599 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 13 00:47:43.029708 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Sep 13 00:47:43.029815 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] Sep 13 00:47:43.029930 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 13 00:47:43.030079 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Sep 13 00:47:43.030226 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Sep 13 00:47:43.030370 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 Sep 13 00:47:43.030531 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Sep 13 00:47:43.030661 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Sep 13 00:47:43.030796 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Sep 13 00:47:43.030936 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Sep 13 00:47:43.031069 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Sep 13 00:47:43.031216 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Sep 13 00:47:43.031350 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Sep 13 00:47:43.031519 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 Sep 13 00:47:43.031657 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x80000000-0x803fffff pref] Sep 13 00:47:43.031788 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Sep 13 00:47:43.031923 kernel: pci 0000:00:03.0: BAR 0: assigned to efifb Sep 13 00:47:43.032053 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 13 00:47:43.032197 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Sep 13 00:47:43.032329 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80404000-0x80407fff] Sep 13 00:47:43.040607 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Sep 13 00:47:43.040779 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80400000-0x80403fff] Sep 13 00:47:43.040801 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 13 00:47:43.040816 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 13 00:47:43.040831 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 13 00:47:43.040850 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 13 00:47:43.040864 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Sep 13 00:47:43.040878 kernel: iommu: Default domain type: Translated Sep 13 00:47:43.040892 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 13 00:47:43.041024 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Sep 13 00:47:43.041156 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 13 00:47:43.041288 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Sep 13 00:47:43.041308 kernel: vgaarb: loaded Sep 13 00:47:43.041325 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 13 00:47:43.041339 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 13 00:47:43.041353 kernel: PTP clock support registered Sep 13 00:47:43.041367 kernel: Registered efivars operations Sep 13 00:47:43.041383 kernel: PCI: Using ACPI for IRQ routing Sep 13 00:47:43.041424 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 13 00:47:43.041441 kernel: e820: reserve RAM buffer [mem 0x76813018-0x77ffffff] Sep 13 00:47:43.041456 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] Sep 13 00:47:43.041472 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] Sep 13 00:47:43.041491 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Sep 13 00:47:43.041508 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Sep 13 00:47:43.041525 kernel: clocksource: Switched to clocksource kvm-clock Sep 13 00:47:43.041539 kernel: VFS: Disk quotas dquot_6.6.0 Sep 13 00:47:43.041554 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 13 00:47:43.041570 kernel: pnp: PnP ACPI init Sep 13 00:47:43.041584 kernel: pnp: PnP ACPI: found 5 devices Sep 13 00:47:43.041600 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 13 00:47:43.041616 kernel: NET: Registered PF_INET protocol family Sep 13 00:47:43.041635 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 13 00:47:43.041650 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Sep 13 00:47:43.041666 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 13 00:47:43.041682 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 13 00:47:43.041698 kernel: TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear) Sep 13 00:47:43.041713 kernel: TCP: Hash tables configured (established 16384 bind 16384) Sep 13 00:47:43.041729 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 13 00:47:43.041744 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 13 00:47:43.041759 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 13 00:47:43.041778 kernel: NET: Registered PF_XDP protocol family Sep 13 00:47:43.041915 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 13 00:47:43.042053 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 13 00:47:43.042187 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 13 00:47:43.042297 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Sep 13 00:47:43.042423 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] Sep 13 00:47:43.042560 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Sep 13 00:47:43.042688 kernel: pci 0000:00:01.0: Activating ISA DMA hang workarounds Sep 13 00:47:43.042708 kernel: PCI: CLS 0 bytes, default 64 Sep 13 00:47:43.042723 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 13 00:47:43.042738 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Sep 13 00:47:43.042752 kernel: clocksource: Switched to clocksource tsc Sep 13 00:47:43.042767 kernel: Initialise system trusted keyrings Sep 13 00:47:43.042781 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Sep 13 00:47:43.042795 kernel: Key type asymmetric registered Sep 13 00:47:43.042808 kernel: Asymmetric key parser 'x509' registered Sep 13 00:47:43.042822 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 13 00:47:43.042837 kernel: io scheduler mq-deadline registered Sep 13 00:47:43.042852 kernel: io scheduler kyber registered Sep 13 00:47:43.042865 kernel: io scheduler bfq registered Sep 13 00:47:43.042879 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 13 00:47:43.042891 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 13 00:47:43.042905 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 13 00:47:43.042918 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 13 00:47:43.042932 kernel: i8042: Warning: Keylock active Sep 13 00:47:43.042950 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 13 00:47:43.042966 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 13 00:47:43.043122 kernel: rtc_cmos 00:00: RTC can wake from S4 Sep 13 00:47:43.043244 kernel: rtc_cmos 00:00: registered as rtc0 Sep 13 00:47:43.043363 kernel: rtc_cmos 00:00: setting system clock to 2025-09-13T00:47:42 UTC (1757724462) Sep 13 00:47:43.043499 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Sep 13 00:47:43.043516 kernel: intel_pstate: CPU model not supported Sep 13 00:47:43.043529 kernel: efifb: probing for efifb Sep 13 00:47:43.043547 kernel: efifb: framebuffer at 0x80000000, using 1876k, total 1875k Sep 13 00:47:43.043561 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Sep 13 00:47:43.043572 kernel: efifb: scrolling: redraw Sep 13 00:47:43.043584 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 13 00:47:43.043596 kernel: Console: switching to colour frame buffer device 100x37 Sep 13 00:47:43.043610 kernel: fb0: EFI VGA frame buffer device Sep 13 00:47:43.043651 kernel: pstore: Registered efi as persistent store backend Sep 13 00:47:43.043668 kernel: NET: Registered PF_INET6 protocol family Sep 13 00:47:43.043683 kernel: Segment Routing with IPv6 Sep 13 00:47:43.043702 kernel: In-situ OAM (IOAM) with IPv6 Sep 13 00:47:43.043717 kernel: NET: Registered PF_PACKET protocol family Sep 13 00:47:43.043732 kernel: Key type dns_resolver registered Sep 13 00:47:43.043745 kernel: IPI shorthand broadcast: enabled Sep 13 00:47:43.043760 kernel: sched_clock: Marking stable (377091091, 142298042)->(587555575, -68166442) Sep 13 00:47:43.043774 kernel: registered taskstats version 1 Sep 13 00:47:43.043788 kernel: Loading compiled-in X.509 certificates Sep 13 00:47:43.043802 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.192-flatcar: d4931373bb0d9b9f95da11f02ae07d3649cc6c37' Sep 13 00:47:43.045857 kernel: Key type .fscrypt registered Sep 13 00:47:43.045897 kernel: Key type fscrypt-provisioning registered Sep 13 00:47:43.045913 kernel: pstore: Using crash dump compression: deflate Sep 13 00:47:43.045930 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 13 00:47:43.045946 kernel: ima: Allocated hash algorithm: sha1 Sep 13 00:47:43.045962 kernel: ima: No architecture policies found Sep 13 00:47:43.045977 kernel: clk: Disabling unused clocks Sep 13 00:47:43.045992 kernel: Freeing unused kernel image (initmem) memory: 47492K Sep 13 00:47:43.046006 kernel: Write protecting the kernel read-only data: 28672k Sep 13 00:47:43.046020 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Sep 13 00:47:43.046037 kernel: Freeing unused kernel image (rodata/data gap) memory: 604K Sep 13 00:47:43.046054 kernel: Run /init as init process Sep 13 00:47:43.046069 kernel: with arguments: Sep 13 00:47:43.046084 kernel: /init Sep 13 00:47:43.046098 kernel: with environment: Sep 13 00:47:43.046113 kernel: HOME=/ Sep 13 00:47:43.046127 kernel: TERM=linux Sep 13 00:47:43.046141 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 13 00:47:43.046160 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 13 00:47:43.046181 systemd[1]: Detected virtualization amazon. Sep 13 00:47:43.046196 systemd[1]: Detected architecture x86-64. Sep 13 00:47:43.046212 systemd[1]: Running in initrd. Sep 13 00:47:43.046227 systemd[1]: No hostname configured, using default hostname. Sep 13 00:47:43.046240 systemd[1]: Hostname set to . Sep 13 00:47:43.046257 systemd[1]: Initializing machine ID from VM UUID. Sep 13 00:47:43.046272 systemd[1]: Queued start job for default target initrd.target. Sep 13 00:47:43.046288 systemd[1]: Started systemd-ask-password-console.path. Sep 13 00:47:43.046301 systemd[1]: Reached target cryptsetup.target. Sep 13 00:47:43.046314 systemd[1]: Reached target paths.target. Sep 13 00:47:43.046328 systemd[1]: Reached target slices.target. Sep 13 00:47:43.046345 systemd[1]: Reached target swap.target. Sep 13 00:47:43.046362 systemd[1]: Reached target timers.target. Sep 13 00:47:43.046379 systemd[1]: Listening on iscsid.socket. Sep 13 00:47:43.047517 systemd[1]: Listening on iscsiuio.socket. Sep 13 00:47:43.047544 systemd[1]: Listening on systemd-journald-audit.socket. Sep 13 00:47:43.047562 systemd[1]: Listening on systemd-journald-dev-log.socket. Sep 13 00:47:43.047580 systemd[1]: Listening on systemd-journald.socket. Sep 13 00:47:43.047597 systemd[1]: Listening on systemd-networkd.socket. Sep 13 00:47:43.047613 systemd[1]: Listening on systemd-udevd-control.socket. Sep 13 00:47:43.047637 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 13 00:47:43.047654 systemd[1]: Reached target sockets.target. Sep 13 00:47:43.047670 systemd[1]: Starting kmod-static-nodes.service... Sep 13 00:47:43.047687 systemd[1]: Finished network-cleanup.service. Sep 13 00:47:43.047704 systemd[1]: Starting systemd-fsck-usr.service... Sep 13 00:47:43.047721 systemd[1]: Starting systemd-journald.service... Sep 13 00:47:43.047737 systemd[1]: Starting systemd-modules-load.service... Sep 13 00:47:43.047754 systemd[1]: Starting systemd-resolved.service... Sep 13 00:47:43.047771 systemd[1]: Starting systemd-vconsole-setup.service... Sep 13 00:47:43.047790 systemd[1]: Finished kmod-static-nodes.service. Sep 13 00:47:43.047807 systemd[1]: Finished systemd-fsck-usr.service. Sep 13 00:47:43.047823 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 13 00:47:43.047841 systemd[1]: Finished systemd-vconsole-setup.service. Sep 13 00:47:43.047859 kernel: audit: type=1130 audit(1757724463.010:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.047876 systemd[1]: Starting dracut-cmdline-ask.service... Sep 13 00:47:43.047893 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Sep 13 00:47:43.047910 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Sep 13 00:47:43.047936 systemd-journald[185]: Journal started Sep 13 00:47:43.048025 systemd-journald[185]: Runtime Journal (/run/log/journal/ec2ec889107eb22543ff8a279d4c7d6a) is 4.8M, max 38.3M, 33.5M free. Sep 13 00:47:43.010000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.032725 systemd-modules-load[186]: Inserted module 'overlay' Sep 13 00:47:43.074990 kernel: audit: type=1130 audit(1757724463.050:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.075040 systemd[1]: Started systemd-journald.service. Sep 13 00:47:43.050000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.074000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.076412 systemd[1]: Finished dracut-cmdline-ask.service. Sep 13 00:47:43.087662 kernel: audit: type=1130 audit(1757724463.074:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.085316 systemd-resolved[187]: Positive Trust Anchors: Sep 13 00:47:43.085328 systemd-resolved[187]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 00:47:43.085382 systemd-resolved[187]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 13 00:47:43.096000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.097000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.089515 systemd-resolved[187]: Defaulting to hostname 'linux'. Sep 13 00:47:43.134222 kernel: audit: type=1130 audit(1757724463.096:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.134259 kernel: audit: type=1130 audit(1757724463.097:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.134287 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 13 00:47:43.134306 kernel: Bridge firewalling registered Sep 13 00:47:43.098029 systemd[1]: Started systemd-resolved.service. Sep 13 00:47:43.099370 systemd[1]: Reached target nss-lookup.target. Sep 13 00:47:43.101617 systemd[1]: Starting dracut-cmdline.service... Sep 13 00:47:43.121499 systemd-modules-load[186]: Inserted module 'br_netfilter' Sep 13 00:47:43.141799 dracut-cmdline[202]: dracut-dracut-053 Sep 13 00:47:43.145915 dracut-cmdline[202]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=65d14b740db9e581daa1d0206188b16d2f1a39e5c5e0878b6855323cd7c584ec Sep 13 00:47:43.160424 kernel: SCSI subsystem initialized Sep 13 00:47:43.178837 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 13 00:47:43.178912 kernel: device-mapper: uevent: version 1.0.3 Sep 13 00:47:43.178932 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Sep 13 00:47:43.186078 systemd-modules-load[186]: Inserted module 'dm_multipath' Sep 13 00:47:43.187000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.187649 systemd[1]: Finished systemd-modules-load.service. Sep 13 00:47:43.196879 kernel: audit: type=1130 audit(1757724463.187:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.195820 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:47:43.207503 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:47:43.208000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.216415 kernel: audit: type=1130 audit(1757724463.208:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.245421 kernel: Loading iSCSI transport class v2.0-870. Sep 13 00:47:43.264422 kernel: iscsi: registered transport (tcp) Sep 13 00:47:43.288440 kernel: iscsi: registered transport (qla4xxx) Sep 13 00:47:43.288522 kernel: QLogic iSCSI HBA Driver Sep 13 00:47:43.322256 systemd[1]: Finished dracut-cmdline.service. Sep 13 00:47:43.321000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.324240 systemd[1]: Starting dracut-pre-udev.service... Sep 13 00:47:43.332371 kernel: audit: type=1130 audit(1757724463.321:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.378452 kernel: raid6: avx512x4 gen() 18016 MB/s Sep 13 00:47:43.396442 kernel: raid6: avx512x4 xor() 7643 MB/s Sep 13 00:47:43.414452 kernel: raid6: avx512x2 gen() 17617 MB/s Sep 13 00:47:43.432445 kernel: raid6: avx512x2 xor() 24368 MB/s Sep 13 00:47:43.450448 kernel: raid6: avx512x1 gen() 17926 MB/s Sep 13 00:47:43.468455 kernel: raid6: avx512x1 xor() 21823 MB/s Sep 13 00:47:43.486457 kernel: raid6: avx2x4 gen() 17701 MB/s Sep 13 00:47:43.504442 kernel: raid6: avx2x4 xor() 7339 MB/s Sep 13 00:47:43.522462 kernel: raid6: avx2x2 gen() 17660 MB/s Sep 13 00:47:43.540444 kernel: raid6: avx2x2 xor() 18177 MB/s Sep 13 00:47:43.558436 kernel: raid6: avx2x1 gen() 13902 MB/s Sep 13 00:47:43.576432 kernel: raid6: avx2x1 xor() 15881 MB/s Sep 13 00:47:43.594464 kernel: raid6: sse2x4 gen() 9102 MB/s Sep 13 00:47:43.612454 kernel: raid6: sse2x4 xor() 6066 MB/s Sep 13 00:47:43.630460 kernel: raid6: sse2x2 gen() 10353 MB/s Sep 13 00:47:43.648451 kernel: raid6: sse2x2 xor() 6106 MB/s Sep 13 00:47:43.666444 kernel: raid6: sse2x1 gen() 9524 MB/s Sep 13 00:47:43.684783 kernel: raid6: sse2x1 xor() 4839 MB/s Sep 13 00:47:43.684834 kernel: raid6: using algorithm avx512x4 gen() 18016 MB/s Sep 13 00:47:43.684864 kernel: raid6: .... xor() 7643 MB/s, rmw enabled Sep 13 00:47:43.685958 kernel: raid6: using avx512x2 recovery algorithm Sep 13 00:47:43.700420 kernel: xor: automatically using best checksumming function avx Sep 13 00:47:43.804428 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Sep 13 00:47:43.813524 systemd[1]: Finished dracut-pre-udev.service. Sep 13 00:47:43.812000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.815165 systemd[1]: Starting systemd-udevd.service... Sep 13 00:47:43.820974 kernel: audit: type=1130 audit(1757724463.812:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.813000 audit: BPF prog-id=7 op=LOAD Sep 13 00:47:43.813000 audit: BPF prog-id=8 op=LOAD Sep 13 00:47:43.832176 systemd-udevd[385]: Using default interface naming scheme 'v252'. Sep 13 00:47:43.837852 systemd[1]: Started systemd-udevd.service. Sep 13 00:47:43.839742 systemd[1]: Starting dracut-pre-trigger.service... Sep 13 00:47:43.837000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.859984 dracut-pre-trigger[391]: rd.md=0: removing MD RAID activation Sep 13 00:47:43.890000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:43.891947 systemd[1]: Finished dracut-pre-trigger.service. Sep 13 00:47:43.893203 systemd[1]: Starting systemd-udev-trigger.service... Sep 13 00:47:43.936920 systemd[1]: Finished systemd-udev-trigger.service. Sep 13 00:47:43.936000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:44.009417 kernel: cryptd: max_cpu_qlen set to 1000 Sep 13 00:47:44.028496 kernel: ena 0000:00:05.0: ENA device version: 0.10 Sep 13 00:47:44.053764 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Sep 13 00:47:44.053934 kernel: AVX2 version of gcm_enc/dec engaged. Sep 13 00:47:44.053954 kernel: AES CTR mode by8 optimization enabled Sep 13 00:47:44.053976 kernel: nvme nvme0: pci function 0000:00:04.0 Sep 13 00:47:44.054151 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Sep 13 00:47:44.054174 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Sep 13 00:47:44.054313 kernel: nvme nvme0: 2/0/0 default/read/poll queues Sep 13 00:47:44.054477 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:d9:bf:d7:f9:c5 Sep 13 00:47:44.061290 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 13 00:47:44.061369 kernel: GPT:9289727 != 16777215 Sep 13 00:47:44.061388 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 13 00:47:44.062724 kernel: GPT:9289727 != 16777215 Sep 13 00:47:44.063774 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 13 00:47:44.066475 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 13 00:47:44.069566 (udev-worker)[442]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:47:44.120425 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/nvme0n1p6 scanned by (udev-worker) (430) Sep 13 00:47:44.156953 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 13 00:47:44.173800 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Sep 13 00:47:44.180264 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Sep 13 00:47:44.182028 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Sep 13 00:47:44.185234 systemd[1]: Starting disk-uuid.service... Sep 13 00:47:44.192470 disk-uuid[593]: Primary Header is updated. Sep 13 00:47:44.192470 disk-uuid[593]: Secondary Entries is updated. Sep 13 00:47:44.192470 disk-uuid[593]: Secondary Header is updated. Sep 13 00:47:44.196660 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Sep 13 00:47:44.201453 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 13 00:47:44.207418 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 13 00:47:45.221013 disk-uuid[594]: The operation has completed successfully. Sep 13 00:47:45.221793 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 13 00:47:45.333519 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 13 00:47:45.332000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.332000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.333625 systemd[1]: Finished disk-uuid.service. Sep 13 00:47:45.335309 systemd[1]: Starting verity-setup.service... Sep 13 00:47:45.353437 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 13 00:47:45.441430 systemd[1]: Found device dev-mapper-usr.device. Sep 13 00:47:45.443966 systemd[1]: Mounting sysusr-usr.mount... Sep 13 00:47:45.447077 systemd[1]: Finished verity-setup.service. Sep 13 00:47:45.446000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.546427 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Sep 13 00:47:45.546985 systemd[1]: Mounted sysusr-usr.mount. Sep 13 00:47:45.547924 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Sep 13 00:47:45.548690 systemd[1]: Starting ignition-setup.service... Sep 13 00:47:45.552024 systemd[1]: Starting parse-ip-for-networkd.service... Sep 13 00:47:45.571998 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 13 00:47:45.572069 kernel: BTRFS info (device nvme0n1p6): using free space tree Sep 13 00:47:45.572089 kernel: BTRFS info (device nvme0n1p6): has skinny extents Sep 13 00:47:45.606438 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 13 00:47:45.620550 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 13 00:47:45.630000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.631309 systemd[1]: Finished ignition-setup.service. Sep 13 00:47:45.634784 systemd[1]: Starting ignition-fetch-offline.service... Sep 13 00:47:45.641665 systemd[1]: Finished parse-ip-for-networkd.service. Sep 13 00:47:45.640000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.642000 audit: BPF prog-id=9 op=LOAD Sep 13 00:47:45.644167 systemd[1]: Starting systemd-networkd.service... Sep 13 00:47:45.668567 systemd-networkd[1107]: lo: Link UP Sep 13 00:47:45.669851 systemd-networkd[1107]: lo: Gained carrier Sep 13 00:47:45.670587 systemd-networkd[1107]: Enumeration completed Sep 13 00:47:45.670874 systemd-networkd[1107]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 00:47:45.672774 systemd[1]: Started systemd-networkd.service. Sep 13 00:47:45.672000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.674192 systemd[1]: Reached target network.target. Sep 13 00:47:45.676676 systemd[1]: Starting iscsiuio.service... Sep 13 00:47:45.676924 systemd-networkd[1107]: eth0: Link UP Sep 13 00:47:45.676930 systemd-networkd[1107]: eth0: Gained carrier Sep 13 00:47:45.684383 systemd[1]: Started iscsiuio.service. Sep 13 00:47:45.684000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.686976 systemd[1]: Starting iscsid.service... Sep 13 00:47:45.691799 iscsid[1112]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Sep 13 00:47:45.691799 iscsid[1112]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Sep 13 00:47:45.691799 iscsid[1112]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Sep 13 00:47:45.691799 iscsid[1112]: If using hardware iscsi like qla4xxx this message can be ignored. Sep 13 00:47:45.697000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.700595 iscsid[1112]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Sep 13 00:47:45.700595 iscsid[1112]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Sep 13 00:47:45.692737 systemd-networkd[1107]: eth0: DHCPv4 address 172.31.19.167/20, gateway 172.31.16.1 acquired from 172.31.16.1 Sep 13 00:47:45.693874 systemd[1]: Started iscsid.service. Sep 13 00:47:45.699932 systemd[1]: Starting dracut-initqueue.service... Sep 13 00:47:45.721446 systemd[1]: Finished dracut-initqueue.service. Sep 13 00:47:45.720000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:45.722241 systemd[1]: Reached target remote-fs-pre.target. Sep 13 00:47:45.723525 systemd[1]: Reached target remote-cryptsetup.target. Sep 13 00:47:45.724693 systemd[1]: Reached target remote-fs.target. Sep 13 00:47:45.727048 systemd[1]: Starting dracut-pre-mount.service... Sep 13 00:47:45.737322 systemd[1]: Finished dracut-pre-mount.service. Sep 13 00:47:45.736000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.167627 ignition[1103]: Ignition 2.14.0 Sep 13 00:47:46.167638 ignition[1103]: Stage: fetch-offline Sep 13 00:47:46.167927 ignition[1103]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:47:46.168051 ignition[1103]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Sep 13 00:47:46.185489 ignition[1103]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 13 00:47:46.185821 ignition[1103]: Ignition finished successfully Sep 13 00:47:46.188161 systemd[1]: Finished ignition-fetch-offline.service. Sep 13 00:47:46.187000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.190108 systemd[1]: Starting ignition-fetch.service... Sep 13 00:47:46.199801 ignition[1131]: Ignition 2.14.0 Sep 13 00:47:46.199813 ignition[1131]: Stage: fetch Sep 13 00:47:46.200018 ignition[1131]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:47:46.200049 ignition[1131]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Sep 13 00:47:46.207280 ignition[1131]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 13 00:47:46.208079 ignition[1131]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 13 00:47:46.278833 ignition[1131]: INFO : PUT result: OK Sep 13 00:47:46.293570 ignition[1131]: DEBUG : parsed url from cmdline: "" Sep 13 00:47:46.293570 ignition[1131]: INFO : no config URL provided Sep 13 00:47:46.293570 ignition[1131]: INFO : reading system config file "/usr/lib/ignition/user.ign" Sep 13 00:47:46.293570 ignition[1131]: INFO : no config at "/usr/lib/ignition/user.ign" Sep 13 00:47:46.296347 ignition[1131]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 13 00:47:46.296347 ignition[1131]: INFO : PUT result: OK Sep 13 00:47:46.296347 ignition[1131]: INFO : GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Sep 13 00:47:46.296347 ignition[1131]: INFO : GET result: OK Sep 13 00:47:46.296347 ignition[1131]: DEBUG : parsing config with SHA512: 78cb74422ada47e7289dfc01cc7e7479c8e5af79043cc3b2592136819a1bebe0621c20005e0954ea8633fea68dbb5a605ba0f0e8d79f890abd36ec76028c17f6 Sep 13 00:47:46.301907 unknown[1131]: fetched base config from "system" Sep 13 00:47:46.303080 unknown[1131]: fetched base config from "system" Sep 13 00:47:46.303288 unknown[1131]: fetched user config from "aws" Sep 13 00:47:46.304526 ignition[1131]: fetch: fetch complete Sep 13 00:47:46.304535 ignition[1131]: fetch: fetch passed Sep 13 00:47:46.306619 systemd[1]: Finished ignition-fetch.service. Sep 13 00:47:46.306000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.304617 ignition[1131]: Ignition finished successfully Sep 13 00:47:46.308948 systemd[1]: Starting ignition-kargs.service... Sep 13 00:47:46.321074 ignition[1137]: Ignition 2.14.0 Sep 13 00:47:46.321087 ignition[1137]: Stage: kargs Sep 13 00:47:46.321291 ignition[1137]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:47:46.321323 ignition[1137]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Sep 13 00:47:46.328404 ignition[1137]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 13 00:47:46.329368 ignition[1137]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 13 00:47:46.330102 ignition[1137]: INFO : PUT result: OK Sep 13 00:47:46.332000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.333660 systemd[1]: Finished ignition-kargs.service. Sep 13 00:47:46.332707 ignition[1137]: kargs: kargs passed Sep 13 00:47:46.335186 systemd[1]: Starting ignition-disks.service... Sep 13 00:47:46.332770 ignition[1137]: Ignition finished successfully Sep 13 00:47:46.344717 ignition[1143]: Ignition 2.14.0 Sep 13 00:47:46.344732 ignition[1143]: Stage: disks Sep 13 00:47:46.344933 ignition[1143]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:47:46.344966 ignition[1143]: parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Sep 13 00:47:46.351927 ignition[1143]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 13 00:47:46.352765 ignition[1143]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 13 00:47:46.353463 ignition[1143]: INFO : PUT result: OK Sep 13 00:47:46.355597 ignition[1143]: disks: disks passed Sep 13 00:47:46.355648 ignition[1143]: Ignition finished successfully Sep 13 00:47:46.355000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.356668 systemd[1]: Finished ignition-disks.service. Sep 13 00:47:46.357328 systemd[1]: Reached target initrd-root-device.target. Sep 13 00:47:46.357823 systemd[1]: Reached target local-fs-pre.target. Sep 13 00:47:46.358270 systemd[1]: Reached target local-fs.target. Sep 13 00:47:46.358791 systemd[1]: Reached target sysinit.target. Sep 13 00:47:46.359898 systemd[1]: Reached target basic.target. Sep 13 00:47:46.362074 systemd[1]: Starting systemd-fsck-root.service... Sep 13 00:47:46.403044 systemd-fsck[1151]: ROOT: clean, 629/553520 files, 56028/553472 blocks Sep 13 00:47:46.405956 systemd[1]: Finished systemd-fsck-root.service. Sep 13 00:47:46.405000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.407479 systemd[1]: Mounting sysroot.mount... Sep 13 00:47:46.427418 kernel: EXT4-fs (nvme0n1p9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Sep 13 00:47:46.429072 systemd[1]: Mounted sysroot.mount. Sep 13 00:47:46.431055 systemd[1]: Reached target initrd-root-fs.target. Sep 13 00:47:46.435618 systemd[1]: Mounting sysroot-usr.mount... Sep 13 00:47:46.438382 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Sep 13 00:47:46.440210 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 13 00:47:46.441264 systemd[1]: Reached target ignition-diskful.target. Sep 13 00:47:46.443670 systemd[1]: Mounted sysroot-usr.mount. Sep 13 00:47:46.447440 systemd[1]: Starting initrd-setup-root.service... Sep 13 00:47:46.460100 initrd-setup-root[1172]: cut: /sysroot/etc/passwd: No such file or directory Sep 13 00:47:46.480637 systemd[1]: Mounting sysroot-usr-share-oem.mount... Sep 13 00:47:46.491336 initrd-setup-root[1181]: cut: /sysroot/etc/group: No such file or directory Sep 13 00:47:46.497250 initrd-setup-root[1189]: cut: /sysroot/etc/shadow: No such file or directory Sep 13 00:47:46.503420 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 scanned by mount (1179) Sep 13 00:47:46.506372 initrd-setup-root[1197]: cut: /sysroot/etc/gshadow: No such file or directory Sep 13 00:47:46.511247 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 13 00:47:46.511272 kernel: BTRFS info (device nvme0n1p6): using free space tree Sep 13 00:47:46.511284 kernel: BTRFS info (device nvme0n1p6): has skinny extents Sep 13 00:47:46.525435 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 13 00:47:46.538152 systemd[1]: Mounted sysroot-usr-share-oem.mount. Sep 13 00:47:46.684274 systemd[1]: Finished initrd-setup-root.service. Sep 13 00:47:46.683000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.686124 systemd[1]: Starting ignition-mount.service... Sep 13 00:47:46.690568 systemd[1]: Starting sysroot-boot.service... Sep 13 00:47:46.698091 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. Sep 13 00:47:46.698224 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. Sep 13 00:47:46.712375 ignition[1233]: INFO : Ignition 2.14.0 Sep 13 00:47:46.713804 ignition[1233]: INFO : Stage: mount Sep 13 00:47:46.715148 ignition[1233]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:47:46.719414 ignition[1233]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Sep 13 00:47:46.731092 ignition[1233]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 13 00:47:46.732472 ignition[1233]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 13 00:47:46.734709 ignition[1233]: INFO : PUT result: OK Sep 13 00:47:46.735018 systemd[1]: Finished sysroot-boot.service. Sep 13 00:47:46.735000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.738469 ignition[1233]: INFO : mount: mount passed Sep 13 00:47:46.739250 ignition[1233]: INFO : Ignition finished successfully Sep 13 00:47:46.740709 systemd[1]: Finished ignition-mount.service. Sep 13 00:47:46.739000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:46.742431 systemd[1]: Starting ignition-files.service... Sep 13 00:47:46.751026 systemd[1]: Mounting sysroot-usr-share-oem.mount... Sep 13 00:47:46.773429 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 scanned by mount (1244) Sep 13 00:47:46.773489 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 13 00:47:46.775934 kernel: BTRFS info (device nvme0n1p6): using free space tree Sep 13 00:47:46.775996 kernel: BTRFS info (device nvme0n1p6): has skinny extents Sep 13 00:47:46.827443 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 13 00:47:46.830995 systemd[1]: Mounted sysroot-usr-share-oem.mount. Sep 13 00:47:46.842078 ignition[1263]: INFO : Ignition 2.14.0 Sep 13 00:47:46.842078 ignition[1263]: INFO : Stage: files Sep 13 00:47:46.844481 ignition[1263]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:47:46.844481 ignition[1263]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Sep 13 00:47:46.851408 ignition[1263]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 13 00:47:46.852332 ignition[1263]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 13 00:47:46.853279 ignition[1263]: INFO : PUT result: OK Sep 13 00:47:46.856027 ignition[1263]: DEBUG : files: compiled without relabeling support, skipping Sep 13 00:47:46.862871 ignition[1263]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 13 00:47:46.862871 ignition[1263]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 13 00:47:46.875540 ignition[1263]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 13 00:47:46.877158 ignition[1263]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 13 00:47:46.877158 ignition[1263]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 13 00:47:46.877040 unknown[1263]: wrote ssh authorized keys file for user: core Sep 13 00:47:46.881275 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Sep 13 00:47:46.881275 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Sep 13 00:47:46.881275 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 13 00:47:46.881275 ignition[1263]: INFO : GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Sep 13 00:47:46.931473 ignition[1263]: INFO : GET result: OK Sep 13 00:47:47.017570 systemd-networkd[1107]: eth0: Gained IPv6LL Sep 13 00:47:47.240223 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 13 00:47:47.242673 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 00:47:47.242673 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 13 00:47:47.242673 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 00:47:47.242673 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 00:47:47.242673 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/etc/eks/bootstrap.sh" Sep 13 00:47:47.242673 ignition[1263]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Sep 13 00:47:47.279448 ignition[1263]: INFO : op(1): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem885723981" Sep 13 00:47:47.279448 ignition[1263]: CRITICAL : op(1): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem885723981": device or resource busy Sep 13 00:47:47.279448 ignition[1263]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem885723981", trying btrfs: device or resource busy Sep 13 00:47:47.279448 ignition[1263]: INFO : op(2): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem885723981" Sep 13 00:47:47.279448 ignition[1263]: INFO : op(2): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem885723981" Sep 13 00:47:47.287362 ignition[1263]: INFO : op(3): [started] unmounting "/mnt/oem885723981" Sep 13 00:47:47.287362 ignition[1263]: INFO : op(3): [finished] unmounting "/mnt/oem885723981" Sep 13 00:47:47.287362 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/etc/eks/bootstrap.sh" Sep 13 00:47:47.287362 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 00:47:47.287362 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 13 00:47:47.287362 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 00:47:47.287362 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 13 00:47:47.287362 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 00:47:47.287362 ignition[1263]: INFO : GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 13 00:47:47.474133 ignition[1263]: INFO : GET result: OK Sep 13 00:47:47.592405 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 13 00:47:47.592405 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/home/core/install.sh" Sep 13 00:47:47.595486 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/home/core/install.sh" Sep 13 00:47:47.595486 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 13 00:47:47.595486 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 13 00:47:47.595486 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(d): [started] writing file "/sysroot/etc/systemd/system/nvidia.service" Sep 13 00:47:47.595486 ignition[1263]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Sep 13 00:47:47.603365 ignition[1263]: INFO : op(4): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2815138201" Sep 13 00:47:47.603365 ignition[1263]: CRITICAL : op(4): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2815138201": device or resource busy Sep 13 00:47:47.603365 ignition[1263]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem2815138201", trying btrfs: device or resource busy Sep 13 00:47:47.603365 ignition[1263]: INFO : op(5): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2815138201" Sep 13 00:47:47.609106 ignition[1263]: INFO : op(5): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2815138201" Sep 13 00:47:47.610530 ignition[1263]: INFO : op(6): [started] unmounting "/mnt/oem2815138201" Sep 13 00:47:47.610452 systemd[1]: mnt-oem2815138201.mount: Deactivated successfully. Sep 13 00:47:47.612248 ignition[1263]: INFO : op(6): [finished] unmounting "/mnt/oem2815138201" Sep 13 00:47:47.612248 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(d): [finished] writing file "/sysroot/etc/systemd/system/nvidia.service" Sep 13 00:47:47.612248 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 00:47:47.612248 ignition[1263]: INFO : GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Sep 13 00:47:48.020580 ignition[1263]: INFO : GET result: OK Sep 13 00:47:48.529546 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 13 00:47:48.529546 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/etc/amazon/ssm/amazon-ssm-agent.json" Sep 13 00:47:48.532737 ignition[1263]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Sep 13 00:47:48.536794 ignition[1263]: INFO : op(7): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2518041946" Sep 13 00:47:48.538350 ignition[1263]: CRITICAL : op(7): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2518041946": device or resource busy Sep 13 00:47:48.538350 ignition[1263]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem2518041946", trying btrfs: device or resource busy Sep 13 00:47:48.538350 ignition[1263]: INFO : op(8): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2518041946" Sep 13 00:47:48.542790 ignition[1263]: INFO : op(8): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2518041946" Sep 13 00:47:48.542790 ignition[1263]: INFO : op(9): [started] unmounting "/mnt/oem2518041946" Sep 13 00:47:48.542790 ignition[1263]: INFO : op(9): [finished] unmounting "/mnt/oem2518041946" Sep 13 00:47:48.542790 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/etc/amazon/ssm/amazon-ssm-agent.json" Sep 13 00:47:48.542790 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(10): [started] writing file "/sysroot/etc/amazon/ssm/seelog.xml" Sep 13 00:47:48.542790 ignition[1263]: INFO : oem config not found in "/usr/share/oem", looking on oem partition Sep 13 00:47:48.549983 ignition[1263]: INFO : op(a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem315361739" Sep 13 00:47:48.551458 ignition[1263]: CRITICAL : op(a): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem315361739": device or resource busy Sep 13 00:47:48.551458 ignition[1263]: ERROR : failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem315361739", trying btrfs: device or resource busy Sep 13 00:47:48.551458 ignition[1263]: INFO : op(b): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem315361739" Sep 13 00:47:48.555551 ignition[1263]: INFO : op(b): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem315361739" Sep 13 00:47:48.555551 ignition[1263]: INFO : op(c): [started] unmounting "/mnt/oem315361739" Sep 13 00:47:48.555551 ignition[1263]: INFO : op(c): [finished] unmounting "/mnt/oem315361739" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: createFilesystemsFiles: createFiles: op(10): [finished] writing file "/sysroot/etc/amazon/ssm/seelog.xml" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(11): [started] processing unit "coreos-metadata-sshkeys@.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(11): [finished] processing unit "coreos-metadata-sshkeys@.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(12): [started] processing unit "amazon-ssm-agent.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(12): op(13): [started] writing unit "amazon-ssm-agent.service" at "/sysroot/etc/systemd/system/amazon-ssm-agent.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(12): op(13): [finished] writing unit "amazon-ssm-agent.service" at "/sysroot/etc/systemd/system/amazon-ssm-agent.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(12): [finished] processing unit "amazon-ssm-agent.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(14): [started] processing unit "nvidia.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(14): [finished] processing unit "nvidia.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(15): [started] processing unit "containerd.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(15): op(16): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(15): op(16): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(15): [finished] processing unit "containerd.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(17): [started] processing unit "prepare-helm.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(17): op(18): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(17): op(18): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 13 00:47:48.555551 ignition[1263]: INFO : files: op(17): [finished] processing unit "prepare-helm.service" Sep 13 00:47:48.589293 kernel: kauditd_printk_skb: 24 callbacks suppressed Sep 13 00:47:48.589319 kernel: audit: type=1130 audit(1757724468.576:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.576000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.557595 systemd[1]: mnt-oem315361739.mount: Deactivated successfully. Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(19): [started] setting preset to enabled for "nvidia.service" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(19): [finished] setting preset to enabled for "nvidia.service" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(1a): [started] setting preset to enabled for "prepare-helm.service" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(1a): [finished] setting preset to enabled for "prepare-helm.service" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(1b): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(1b): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(1c): [started] setting preset to enabled for "amazon-ssm-agent.service" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: op(1c): [finished] setting preset to enabled for "amazon-ssm-agent.service" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: createResultFile: createFiles: op(1d): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: createResultFile: createFiles: op(1d): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 13 00:47:48.591031 ignition[1263]: INFO : files: files passed Sep 13 00:47:48.591031 ignition[1263]: INFO : Ignition finished successfully Sep 13 00:47:48.629727 kernel: audit: type=1130 audit(1757724468.596:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.629755 kernel: audit: type=1130 audit(1757724468.604:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.629767 kernel: audit: type=1131 audit(1757724468.604:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.596000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.604000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.604000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.575968 systemd[1]: Finished ignition-files.service. Sep 13 00:47:48.579242 systemd[1]: Starting initrd-setup-root-after-ignition.service... Sep 13 00:47:48.631737 initrd-setup-root-after-ignition[1287]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 13 00:47:48.590378 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Sep 13 00:47:48.591462 systemd[1]: Starting ignition-quench.service... Sep 13 00:47:48.596510 systemd[1]: Finished initrd-setup-root-after-ignition.service. Sep 13 00:47:48.598507 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 13 00:47:48.598608 systemd[1]: Finished ignition-quench.service. Sep 13 00:47:48.605910 systemd[1]: Reached target ignition-complete.target. Sep 13 00:47:48.622090 systemd[1]: Starting initrd-parse-etc.service... Sep 13 00:47:48.640421 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 13 00:47:48.640513 systemd[1]: Finished initrd-parse-etc.service. Sep 13 00:47:48.652020 kernel: audit: type=1130 audit(1757724468.640:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.652058 kernel: audit: type=1131 audit(1757724468.640:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.640000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.640000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.641815 systemd[1]: Reached target initrd-fs.target. Sep 13 00:47:48.652710 systemd[1]: Reached target initrd.target. Sep 13 00:47:48.654011 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Sep 13 00:47:48.654998 systemd[1]: Starting dracut-pre-pivot.service... Sep 13 00:47:48.669339 systemd[1]: Finished dracut-pre-pivot.service. Sep 13 00:47:48.668000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.670949 systemd[1]: Starting initrd-cleanup.service... Sep 13 00:47:48.677444 kernel: audit: type=1130 audit(1757724468.668:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.681421 systemd[1]: Stopped target nss-lookup.target. Sep 13 00:47:48.682263 systemd[1]: Stopped target remote-cryptsetup.target. Sep 13 00:47:48.683916 systemd[1]: Stopped target timers.target. Sep 13 00:47:48.685012 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 13 00:47:48.691677 kernel: audit: type=1131 audit(1757724468.684:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.684000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.685279 systemd[1]: Stopped dracut-pre-pivot.service. Sep 13 00:47:48.686313 systemd[1]: Stopped target initrd.target. Sep 13 00:47:48.692449 systemd[1]: Stopped target basic.target. Sep 13 00:47:48.693639 systemd[1]: Stopped target ignition-complete.target. Sep 13 00:47:48.694797 systemd[1]: Stopped target ignition-diskful.target. Sep 13 00:47:48.696214 systemd[1]: Stopped target initrd-root-device.target. Sep 13 00:47:48.697889 systemd[1]: Stopped target remote-fs.target. Sep 13 00:47:48.699374 systemd[1]: Stopped target remote-fs-pre.target. Sep 13 00:47:48.700532 systemd[1]: Stopped target sysinit.target. Sep 13 00:47:48.701899 systemd[1]: Stopped target local-fs.target. Sep 13 00:47:48.703175 systemd[1]: Stopped target local-fs-pre.target. Sep 13 00:47:48.704930 systemd[1]: Stopped target swap.target. Sep 13 00:47:48.712964 kernel: audit: type=1131 audit(1757724468.706:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.706000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.706060 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 13 00:47:48.706286 systemd[1]: Stopped dracut-pre-mount.service. Sep 13 00:47:48.720593 kernel: audit: type=1131 audit(1757724468.713:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.713000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.707704 systemd[1]: Stopped target cryptsetup.target. Sep 13 00:47:48.720000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.713752 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 13 00:47:48.721000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.713970 systemd[1]: Stopped dracut-initqueue.service. Sep 13 00:47:48.715294 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 13 00:47:48.715549 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Sep 13 00:47:48.721621 systemd[1]: ignition-files.service: Deactivated successfully. Sep 13 00:47:48.721831 systemd[1]: Stopped ignition-files.service. Sep 13 00:47:48.724268 systemd[1]: Stopping ignition-mount.service... Sep 13 00:47:48.731385 iscsid[1112]: iscsid shutting down. Sep 13 00:47:48.732733 systemd[1]: Stopping iscsid.service... Sep 13 00:47:48.738252 systemd[1]: Stopping sysroot-boot.service... Sep 13 00:47:48.740751 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 13 00:47:48.742459 systemd[1]: Stopped systemd-udev-trigger.service. Sep 13 00:47:48.743000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.745176 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 13 00:47:48.746525 systemd[1]: Stopped dracut-pre-trigger.service. Sep 13 00:47:48.748697 ignition[1301]: INFO : Ignition 2.14.0 Sep 13 00:47:48.748697 ignition[1301]: INFO : Stage: umount Sep 13 00:47:48.748000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.752950 ignition[1301]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 13 00:47:48.752950 ignition[1301]: DEBUG : parsing config with SHA512: 6629d8e825d60c9c9d4629d8547ef9a0b839d6b01b7f61a481a1f23308c924b8b0bbf10cae7f7fe3bcaf88b23d1a81baa7771c3670728d4d2a1e665216a1de7b Sep 13 00:47:48.756000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.755627 systemd[1]: iscsid.service: Deactivated successfully. Sep 13 00:47:48.755781 systemd[1]: Stopped iscsid.service. Sep 13 00:47:48.762000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.762000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.758932 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 13 00:47:48.760541 systemd[1]: Finished initrd-cleanup.service. Sep 13 00:47:48.769275 systemd[1]: Stopping iscsiuio.service... Sep 13 00:47:48.773102 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 13 00:47:48.773743 systemd[1]: iscsiuio.service: Deactivated successfully. Sep 13 00:47:48.776000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.778000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.775472 systemd[1]: Stopped iscsiuio.service. Sep 13 00:47:48.781563 ignition[1301]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 13 00:47:48.781563 ignition[1301]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 13 00:47:48.781563 ignition[1301]: INFO : PUT result: OK Sep 13 00:47:48.778563 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 13 00:47:48.778697 systemd[1]: Stopped sysroot-boot.service. Sep 13 00:47:48.785931 ignition[1301]: INFO : umount: umount passed Sep 13 00:47:48.785931 ignition[1301]: INFO : Ignition finished successfully Sep 13 00:47:48.785000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.786000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.785675 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 13 00:47:48.787000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.785801 systemd[1]: Stopped ignition-mount.service. Sep 13 00:47:48.788000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.786608 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 13 00:47:48.786672 systemd[1]: Stopped ignition-disks.service. Sep 13 00:47:48.790000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.787832 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 13 00:47:48.787893 systemd[1]: Stopped ignition-kargs.service. Sep 13 00:47:48.788955 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 13 00:47:48.789014 systemd[1]: Stopped ignition-fetch.service. Sep 13 00:47:48.790107 systemd[1]: Stopped target network.target. Sep 13 00:47:48.791290 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 13 00:47:48.791357 systemd[1]: Stopped ignition-fetch-offline.service. Sep 13 00:47:48.792465 systemd[1]: Stopped target paths.target. Sep 13 00:47:48.793517 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 13 00:47:48.796452 systemd[1]: Stopped systemd-ask-password-console.path. Sep 13 00:47:48.797447 systemd[1]: Stopped target slices.target. Sep 13 00:47:48.798452 systemd[1]: Stopped target sockets.target. Sep 13 00:47:48.801000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.799611 systemd[1]: iscsid.socket: Deactivated successfully. Sep 13 00:47:48.802000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.799672 systemd[1]: Closed iscsid.socket. Sep 13 00:47:48.800690 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 13 00:47:48.800739 systemd[1]: Closed iscsiuio.socket. Sep 13 00:47:48.801762 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 13 00:47:48.801826 systemd[1]: Stopped ignition-setup.service. Sep 13 00:47:48.802905 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 13 00:47:48.802961 systemd[1]: Stopped initrd-setup-root.service. Sep 13 00:47:48.804286 systemd[1]: Stopping systemd-networkd.service... Sep 13 00:47:48.805469 systemd[1]: Stopping systemd-resolved.service... Sep 13 00:47:48.810000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.808458 systemd-networkd[1107]: eth0: DHCPv6 lease lost Sep 13 00:47:48.812000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.811434 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 13 00:47:48.811575 systemd[1]: Stopped systemd-resolved.service. Sep 13 00:47:48.815000 audit: BPF prog-id=6 op=UNLOAD Sep 13 00:47:48.813109 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 13 00:47:48.817000 audit: BPF prog-id=9 op=UNLOAD Sep 13 00:47:48.813243 systemd[1]: Stopped systemd-networkd.service. Sep 13 00:47:48.818000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.814379 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 13 00:47:48.819000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.814437 systemd[1]: Closed systemd-networkd.socket. Sep 13 00:47:48.821000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.816609 systemd[1]: Stopping network-cleanup.service... Sep 13 00:47:48.818881 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 13 00:47:48.818963 systemd[1]: Stopped parse-ip-for-networkd.service. Sep 13 00:47:48.820256 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 00:47:48.820321 systemd[1]: Stopped systemd-sysctl.service. Sep 13 00:47:48.821485 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 13 00:47:48.821546 systemd[1]: Stopped systemd-modules-load.service. Sep 13 00:47:48.832000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.822756 systemd[1]: Stopping systemd-udevd.service... Sep 13 00:47:48.832029 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 13 00:47:48.832921 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 13 00:47:48.833103 systemd[1]: Stopped systemd-udevd.service. Sep 13 00:47:48.839000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.835578 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 13 00:47:48.840000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.835637 systemd[1]: Closed systemd-udevd-control.socket. Sep 13 00:47:48.841000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.836558 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 13 00:47:48.846000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.847000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.848000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.850000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.836605 systemd[1]: Closed systemd-udevd-kernel.socket. Sep 13 00:47:48.839483 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 13 00:47:48.839553 systemd[1]: Stopped dracut-pre-udev.service. Sep 13 00:47:48.853000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.853000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:48.840701 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 13 00:47:48.840761 systemd[1]: Stopped dracut-cmdline.service. Sep 13 00:47:48.841875 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 13 00:47:48.841937 systemd[1]: Stopped dracut-cmdline-ask.service. Sep 13 00:47:48.844253 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Sep 13 00:47:48.846452 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 13 00:47:48.846525 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Sep 13 00:47:48.847738 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 13 00:47:48.847794 systemd[1]: Stopped kmod-static-nodes.service. Sep 13 00:47:48.848817 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 13 00:47:48.848871 systemd[1]: Stopped systemd-vconsole-setup.service. Sep 13 00:47:48.850417 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 13 00:47:48.850539 systemd[1]: Stopped network-cleanup.service. Sep 13 00:47:48.853692 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 13 00:47:48.853816 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Sep 13 00:47:48.855049 systemd[1]: Reached target initrd-switch-root.target. Sep 13 00:47:48.857221 systemd[1]: Starting initrd-switch-root.service... Sep 13 00:47:48.879248 systemd[1]: Switching root. Sep 13 00:47:48.880000 audit: BPF prog-id=5 op=UNLOAD Sep 13 00:47:48.880000 audit: BPF prog-id=4 op=UNLOAD Sep 13 00:47:48.880000 audit: BPF prog-id=3 op=UNLOAD Sep 13 00:47:48.883000 audit: BPF prog-id=8 op=UNLOAD Sep 13 00:47:48.883000 audit: BPF prog-id=7 op=UNLOAD Sep 13 00:47:48.902269 systemd-journald[185]: Journal stopped Sep 13 00:47:54.291313 systemd-journald[185]: Received SIGTERM from PID 1 (systemd). Sep 13 00:47:54.291409 kernel: SELinux: Class mctp_socket not defined in policy. Sep 13 00:47:54.291430 kernel: SELinux: Class anon_inode not defined in policy. Sep 13 00:47:54.291453 kernel: SELinux: the above unknown classes and permissions will be allowed Sep 13 00:47:54.291470 kernel: SELinux: policy capability network_peer_controls=1 Sep 13 00:47:54.291491 kernel: SELinux: policy capability open_perms=1 Sep 13 00:47:54.291508 kernel: SELinux: policy capability extended_socket_class=1 Sep 13 00:47:54.291526 kernel: SELinux: policy capability always_check_network=0 Sep 13 00:47:54.291543 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 13 00:47:54.291559 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 13 00:47:54.291579 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 13 00:47:54.291597 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 13 00:47:54.291615 systemd[1]: Successfully loaded SELinux policy in 80.455ms. Sep 13 00:47:54.291649 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.014ms. Sep 13 00:47:54.291670 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 13 00:47:54.291694 systemd[1]: Detected virtualization amazon. Sep 13 00:47:54.291713 systemd[1]: Detected architecture x86-64. Sep 13 00:47:54.291735 systemd[1]: Detected first boot. Sep 13 00:47:54.291753 systemd[1]: Initializing machine ID from VM UUID. Sep 13 00:47:54.291772 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Sep 13 00:47:54.291790 systemd[1]: Populated /etc with preset unit settings. Sep 13 00:47:54.291809 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:47:54.291837 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:47:54.291858 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:47:54.291876 systemd[1]: Queued start job for default target multi-user.target. Sep 13 00:47:54.291896 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device. Sep 13 00:47:54.291916 systemd[1]: Created slice system-addon\x2dconfig.slice. Sep 13 00:47:54.291934 systemd[1]: Created slice system-addon\x2drun.slice. Sep 13 00:47:54.291953 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. Sep 13 00:47:54.291972 systemd[1]: Created slice system-getty.slice. Sep 13 00:47:54.291990 systemd[1]: Created slice system-modprobe.slice. Sep 13 00:47:54.292009 systemd[1]: Created slice system-serial\x2dgetty.slice. Sep 13 00:47:54.292027 systemd[1]: Created slice system-system\x2dcloudinit.slice. Sep 13 00:47:54.292045 systemd[1]: Created slice system-systemd\x2dfsck.slice. Sep 13 00:47:54.292065 systemd[1]: Created slice user.slice. Sep 13 00:47:54.292083 systemd[1]: Started systemd-ask-password-console.path. Sep 13 00:47:54.292104 systemd[1]: Started systemd-ask-password-wall.path. Sep 13 00:47:54.292124 systemd[1]: Set up automount boot.automount. Sep 13 00:47:54.292152 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Sep 13 00:47:54.292176 systemd[1]: Reached target integritysetup.target. Sep 13 00:47:54.299463 systemd[1]: Reached target remote-cryptsetup.target. Sep 13 00:47:54.299507 systemd[1]: Reached target remote-fs.target. Sep 13 00:47:54.299526 systemd[1]: Reached target slices.target. Sep 13 00:47:54.299545 systemd[1]: Reached target swap.target. Sep 13 00:47:54.299564 systemd[1]: Reached target torcx.target. Sep 13 00:47:54.299582 systemd[1]: Reached target veritysetup.target. Sep 13 00:47:54.299600 systemd[1]: Listening on systemd-coredump.socket. Sep 13 00:47:54.299626 systemd[1]: Listening on systemd-initctl.socket. Sep 13 00:47:54.299645 kernel: kauditd_printk_skb: 49 callbacks suppressed Sep 13 00:47:54.299669 kernel: audit: type=1400 audit(1757724474.040:87): avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 13 00:47:54.299688 systemd[1]: Listening on systemd-journald-audit.socket. Sep 13 00:47:54.299707 kernel: audit: type=1335 audit(1757724474.040:88): pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Sep 13 00:47:54.299722 systemd[1]: Listening on systemd-journald-dev-log.socket. Sep 13 00:47:54.299743 systemd[1]: Listening on systemd-journald.socket. Sep 13 00:47:54.299768 systemd[1]: Listening on systemd-networkd.socket. Sep 13 00:47:54.299789 systemd[1]: Listening on systemd-udevd-control.socket. Sep 13 00:47:54.299810 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 13 00:47:54.299832 systemd[1]: Listening on systemd-userdbd.socket. Sep 13 00:47:54.299855 systemd[1]: Mounting dev-hugepages.mount... Sep 13 00:47:54.299877 systemd[1]: Mounting dev-mqueue.mount... Sep 13 00:47:54.299898 systemd[1]: Mounting media.mount... Sep 13 00:47:54.299920 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:47:54.299942 systemd[1]: Mounting sys-kernel-debug.mount... Sep 13 00:47:54.299965 systemd[1]: Mounting sys-kernel-tracing.mount... Sep 13 00:47:54.299987 systemd[1]: Mounting tmp.mount... Sep 13 00:47:54.300008 systemd[1]: Starting flatcar-tmpfiles.service... Sep 13 00:47:54.300029 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:47:54.300050 systemd[1]: Starting kmod-static-nodes.service... Sep 13 00:47:54.300071 systemd[1]: Starting modprobe@configfs.service... Sep 13 00:47:54.300093 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:47:54.300114 systemd[1]: Starting modprobe@drm.service... Sep 13 00:47:54.300135 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:47:54.300159 systemd[1]: Starting modprobe@fuse.service... Sep 13 00:47:54.300180 systemd[1]: Starting modprobe@loop.service... Sep 13 00:47:54.300203 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 13 00:47:54.300224 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Sep 13 00:47:54.300246 systemd[1]: (This warning is only shown for the first unit using IP firewalling.) Sep 13 00:47:54.300266 systemd[1]: Starting systemd-journald.service... Sep 13 00:47:54.300287 systemd[1]: Starting systemd-modules-load.service... Sep 13 00:47:54.300308 systemd[1]: Starting systemd-network-generator.service... Sep 13 00:47:54.300330 systemd[1]: Starting systemd-remount-fs.service... Sep 13 00:47:54.300354 systemd[1]: Starting systemd-udev-trigger.service... Sep 13 00:47:54.300376 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:47:54.300430 systemd[1]: Mounted dev-hugepages.mount. Sep 13 00:47:54.300452 systemd[1]: Mounted dev-mqueue.mount. Sep 13 00:47:54.300474 systemd[1]: Mounted media.mount. Sep 13 00:47:54.300495 kernel: fuse: init (API version 7.34) Sep 13 00:47:54.300516 systemd[1]: Mounted sys-kernel-debug.mount. Sep 13 00:47:54.300536 systemd[1]: Mounted sys-kernel-tracing.mount. Sep 13 00:47:54.300558 systemd[1]: Mounted tmp.mount. Sep 13 00:47:54.300581 systemd[1]: Finished kmod-static-nodes.service. Sep 13 00:47:54.300603 kernel: audit: type=1130 audit(1757724474.281:89): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.300624 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 13 00:47:54.300646 kernel: audit: type=1305 audit(1757724474.282:90): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Sep 13 00:47:54.300666 systemd[1]: Finished modprobe@configfs.service. Sep 13 00:47:54.300693 systemd-journald[1452]: Journal started Sep 13 00:47:54.300767 systemd-journald[1452]: Runtime Journal (/run/log/journal/ec2ec889107eb22543ff8a279d4c7d6a) is 4.8M, max 38.3M, 33.5M free. Sep 13 00:47:54.040000 audit[1]: AVC avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 13 00:47:54.040000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Sep 13 00:47:54.281000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.282000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Sep 13 00:47:54.302448 kernel: audit: type=1300 audit(1757724474.282:90): arch=c000003e syscall=46 success=yes exit=60 a0=4 a1=7ffe709b6590 a2=4000 a3=7ffe709b662c items=0 ppid=1 pid=1452 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:47:54.302495 kernel: audit: type=1327 audit(1757724474.282:90): proctitle="/usr/lib/systemd/systemd-journald" Sep 13 00:47:54.282000 audit[1452]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=4 a1=7ffe709b6590 a2=4000 a3=7ffe709b662c items=0 ppid=1 pid=1452 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:47:54.282000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Sep 13 00:47:54.322413 kernel: loop: module loaded Sep 13 00:47:54.324000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.335597 kernel: audit: type=1130 audit(1757724474.324:91): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.335681 systemd[1]: Started systemd-journald.service. Sep 13 00:47:54.331000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.345426 kernel: audit: type=1131 audit(1757724474.331:92): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.347929 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:47:54.348192 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:47:54.349917 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 00:47:54.343000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.348000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.358084 systemd[1]: Finished modprobe@drm.service. Sep 13 00:47:54.365724 kernel: audit: type=1130 audit(1757724474.343:93): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.373756 kernel: audit: type=1130 audit(1757724474.348:94): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.348000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.365000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.365000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.367000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.370000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.370000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.367611 systemd[1]: Finished flatcar-tmpfiles.service. Sep 13 00:47:54.369267 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:47:54.370342 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:47:54.372231 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 13 00:47:54.372486 systemd[1]: Finished modprobe@fuse.service. Sep 13 00:47:54.374823 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:47:54.373000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.373000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.375201 systemd[1]: Finished modprobe@loop.service. Sep 13 00:47:54.375000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.375000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.378832 systemd[1]: Finished systemd-modules-load.service. Sep 13 00:47:54.380000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.382743 systemd[1]: Finished systemd-network-generator.service. Sep 13 00:47:54.382000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.384987 systemd[1]: Finished systemd-remount-fs.service. Sep 13 00:47:54.384000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.386953 systemd[1]: Reached target network-pre.target. Sep 13 00:47:54.390210 systemd[1]: Mounting sys-fs-fuse-connections.mount... Sep 13 00:47:54.393390 systemd[1]: Mounting sys-kernel-config.mount... Sep 13 00:47:54.394884 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 13 00:47:54.405176 systemd[1]: Starting systemd-hwdb-update.service... Sep 13 00:47:54.407947 systemd[1]: Starting systemd-journal-flush.service... Sep 13 00:47:54.409269 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:47:54.411240 systemd[1]: Starting systemd-random-seed.service... Sep 13 00:47:54.412387 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:47:54.414175 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:47:54.417172 systemd[1]: Starting systemd-sysusers.service... Sep 13 00:47:54.423280 systemd[1]: Finished systemd-udev-trigger.service. Sep 13 00:47:54.424000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.426512 systemd[1]: Mounted sys-fs-fuse-connections.mount. Sep 13 00:47:54.434812 systemd-journald[1452]: Time spent on flushing to /var/log/journal/ec2ec889107eb22543ff8a279d4c7d6a is 43.572ms for 1172 entries. Sep 13 00:47:54.434812 systemd-journald[1452]: System Journal (/var/log/journal/ec2ec889107eb22543ff8a279d4c7d6a) is 8.0M, max 195.6M, 187.6M free. Sep 13 00:47:54.507370 systemd-journald[1452]: Received client request to flush runtime journal. Sep 13 00:47:54.460000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.471000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.508000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.428601 systemd[1]: Mounted sys-kernel-config.mount. Sep 13 00:47:54.433812 systemd[1]: Starting systemd-udev-settle.service... Sep 13 00:47:54.510164 udevadm[1496]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Sep 13 00:47:54.461012 systemd[1]: Finished systemd-random-seed.service. Sep 13 00:47:54.462148 systemd[1]: Reached target first-boot-complete.target. Sep 13 00:47:54.471836 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:47:54.508672 systemd[1]: Finished systemd-journal-flush.service. Sep 13 00:47:54.602000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:54.602366 systemd[1]: Finished systemd-sysusers.service. Sep 13 00:47:54.605324 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Sep 13 00:47:54.704845 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Sep 13 00:47:54.704000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.097369 systemd[1]: Finished systemd-hwdb-update.service. Sep 13 00:47:55.096000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.099521 systemd[1]: Starting systemd-udevd.service... Sep 13 00:47:55.121686 systemd-udevd[1506]: Using default interface naming scheme 'v252'. Sep 13 00:47:55.188000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.189113 systemd[1]: Started systemd-udevd.service. Sep 13 00:47:55.192120 systemd[1]: Starting systemd-networkd.service... Sep 13 00:47:55.214372 systemd[1]: Starting systemd-userdbd.service... Sep 13 00:47:55.250873 systemd[1]: Found device dev-ttyS0.device. Sep 13 00:47:55.285732 systemd[1]: Started systemd-userdbd.service. Sep 13 00:47:55.285000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.329386 (udev-worker)[1522]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:47:55.339421 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Sep 13 00:47:55.358587 kernel: ACPI: button: Power Button [PWRF] Sep 13 00:47:55.358696 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input3 Sep 13 00:47:55.378000 audit[1514]: AVC avc: denied { confidentiality } for pid=1514 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Sep 13 00:47:55.378000 audit[1514]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=561f38bf3140 a1=338ec a2=7f632909dbc5 a3=5 items=110 ppid=1506 pid=1514 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:47:55.378000 audit: CWD cwd="/" Sep 13 00:47:55.378000 audit: PATH item=0 name=(null) inode=43 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=1 name=(null) inode=14210 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=2 name=(null) inode=14210 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=3 name=(null) inode=14211 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=4 name=(null) inode=14210 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=5 name=(null) inode=14212 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=6 name=(null) inode=14210 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=7 name=(null) inode=14213 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=8 name=(null) inode=14213 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=9 name=(null) inode=14214 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=10 name=(null) inode=14213 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=11 name=(null) inode=14215 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=12 name=(null) inode=14213 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=13 name=(null) inode=14216 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=14 name=(null) inode=14213 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=15 name=(null) inode=14217 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=16 name=(null) inode=14213 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=17 name=(null) inode=14218 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=18 name=(null) inode=14210 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=19 name=(null) inode=14219 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=20 name=(null) inode=14219 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=21 name=(null) inode=14220 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=22 name=(null) inode=14219 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=23 name=(null) inode=14221 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=24 name=(null) inode=14219 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=25 name=(null) inode=14222 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=26 name=(null) inode=14219 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=27 name=(null) inode=14223 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=28 name=(null) inode=14219 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=29 name=(null) inode=14224 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=30 name=(null) inode=14210 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=31 name=(null) inode=14225 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=32 name=(null) inode=14225 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=33 name=(null) inode=14226 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=34 name=(null) inode=14225 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=35 name=(null) inode=14227 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=36 name=(null) inode=14225 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=37 name=(null) inode=14228 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=38 name=(null) inode=14225 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=39 name=(null) inode=14229 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=40 name=(null) inode=14225 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=41 name=(null) inode=14230 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=42 name=(null) inode=14210 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=43 name=(null) inode=14231 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=44 name=(null) inode=14231 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=45 name=(null) inode=14232 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=46 name=(null) inode=14231 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=47 name=(null) inode=14233 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=48 name=(null) inode=14231 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=49 name=(null) inode=14234 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.400453 kernel: ACPI: button: Sleep Button [SLPF] Sep 13 00:47:55.378000 audit: PATH item=50 name=(null) inode=14231 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=51 name=(null) inode=14235 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=52 name=(null) inode=14231 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=53 name=(null) inode=14236 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=54 name=(null) inode=43 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=55 name=(null) inode=14237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=56 name=(null) inode=14237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=57 name=(null) inode=14238 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=58 name=(null) inode=14237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=59 name=(null) inode=14239 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=60 name=(null) inode=14237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=61 name=(null) inode=14240 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=62 name=(null) inode=14240 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=63 name=(null) inode=14241 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=64 name=(null) inode=14240 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=65 name=(null) inode=14242 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=66 name=(null) inode=14240 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=67 name=(null) inode=14243 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=68 name=(null) inode=14240 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=69 name=(null) inode=14244 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=70 name=(null) inode=14240 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=71 name=(null) inode=14245 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=72 name=(null) inode=14237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=73 name=(null) inode=14246 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.404000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.378000 audit: PATH item=74 name=(null) inode=14246 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=75 name=(null) inode=14247 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=76 name=(null) inode=14246 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=77 name=(null) inode=14248 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=78 name=(null) inode=14246 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=79 name=(null) inode=14249 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=80 name=(null) inode=14246 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=81 name=(null) inode=14250 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=82 name=(null) inode=14246 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=83 name=(null) inode=14251 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=84 name=(null) inode=14237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=85 name=(null) inode=14252 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=86 name=(null) inode=14252 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=87 name=(null) inode=14253 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=88 name=(null) inode=14252 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=89 name=(null) inode=14254 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=90 name=(null) inode=14252 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=91 name=(null) inode=14255 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=92 name=(null) inode=14252 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=93 name=(null) inode=14256 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=94 name=(null) inode=14252 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=95 name=(null) inode=14257 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=96 name=(null) inode=14237 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=97 name=(null) inode=14258 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=98 name=(null) inode=14258 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=99 name=(null) inode=14259 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=100 name=(null) inode=14258 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=101 name=(null) inode=14260 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=102 name=(null) inode=14258 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=103 name=(null) inode=14261 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=104 name=(null) inode=14258 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=105 name=(null) inode=14262 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=106 name=(null) inode=14258 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=107 name=(null) inode=14263 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PATH item=109 name=(null) inode=14830 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 13 00:47:55.378000 audit: PROCTITLE proctitle="(udev-worker)" Sep 13 00:47:55.404787 systemd-networkd[1513]: lo: Link UP Sep 13 00:47:55.404797 systemd-networkd[1513]: lo: Gained carrier Sep 13 00:47:55.405413 systemd-networkd[1513]: Enumeration completed Sep 13 00:47:55.405585 systemd[1]: Started systemd-networkd.service. Sep 13 00:47:55.408164 systemd[1]: Starting systemd-networkd-wait-online.service... Sep 13 00:47:55.411666 systemd-networkd[1513]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 13 00:47:55.418237 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Sep 13 00:47:55.417773 systemd-networkd[1513]: eth0: Link UP Sep 13 00:47:55.417955 systemd-networkd[1513]: eth0: Gained carrier Sep 13 00:47:55.428609 systemd-networkd[1513]: eth0: DHCPv4 address 172.31.19.167/20, gateway 172.31.16.1 acquired from 172.31.16.1 Sep 13 00:47:55.434538 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Sep 13 00:47:55.461414 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input4 Sep 13 00:47:55.477502 kernel: mousedev: PS/2 mouse device common for all mice Sep 13 00:47:55.620167 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 13 00:47:55.621588 systemd[1]: Finished systemd-udev-settle.service. Sep 13 00:47:55.621000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.628844 systemd[1]: Starting lvm2-activation-early.service... Sep 13 00:47:55.681231 lvm[1622]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 00:47:55.711000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.711862 systemd[1]: Finished lvm2-activation-early.service. Sep 13 00:47:55.712543 systemd[1]: Reached target cryptsetup.target. Sep 13 00:47:55.714503 systemd[1]: Starting lvm2-activation.service... Sep 13 00:47:55.721030 lvm[1624]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 13 00:47:55.742905 systemd[1]: Finished lvm2-activation.service. Sep 13 00:47:55.742000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.744165 systemd[1]: Reached target local-fs-pre.target. Sep 13 00:47:55.745097 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 13 00:47:55.745142 systemd[1]: Reached target local-fs.target. Sep 13 00:47:55.745978 systemd[1]: Reached target machines.target. Sep 13 00:47:55.748794 systemd[1]: Starting ldconfig.service... Sep 13 00:47:55.750519 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:47:55.750615 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:47:55.752143 systemd[1]: Starting systemd-boot-update.service... Sep 13 00:47:55.754193 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Sep 13 00:47:55.760131 systemd[1]: Starting systemd-machine-id-commit.service... Sep 13 00:47:55.763988 systemd[1]: Starting systemd-sysext.service... Sep 13 00:47:55.765885 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1627 (bootctl) Sep 13 00:47:55.769738 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Sep 13 00:47:55.790609 systemd[1]: Unmounting usr-share-oem.mount... Sep 13 00:47:55.797364 systemd[1]: usr-share-oem.mount: Deactivated successfully. Sep 13 00:47:55.797712 systemd[1]: Unmounted usr-share-oem.mount. Sep 13 00:47:55.818431 kernel: loop0: detected capacity change from 0 to 221472 Sep 13 00:47:55.831803 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Sep 13 00:47:55.831000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.960427 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 13 00:47:55.981418 kernel: loop1: detected capacity change from 0 to 221472 Sep 13 00:47:55.988726 systemd-fsck[1639]: fsck.fat 4.2 (2021-01-31) Sep 13 00:47:55.988726 systemd-fsck[1639]: /dev/nvme0n1p1: 790 files, 120761/258078 clusters Sep 13 00:47:55.991252 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Sep 13 00:47:55.990000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:55.994283 systemd[1]: Mounting boot.mount... Sep 13 00:47:56.020913 (sd-sysext)[1644]: Using extensions 'kubernetes'. Sep 13 00:47:56.024248 systemd[1]: Mounted boot.mount. Sep 13 00:47:56.024311 (sd-sysext)[1644]: Merged extensions into '/usr'. Sep 13 00:47:56.071524 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 13 00:47:56.073329 systemd[1]: Finished systemd-boot-update.service. Sep 13 00:47:56.073000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.076211 systemd[1]: Finished systemd-machine-id-commit.service. Sep 13 00:47:56.075000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.078337 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:47:56.080888 systemd[1]: Mounting usr-share-oem.mount... Sep 13 00:47:56.082087 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.084345 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:47:56.088992 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:47:56.092711 systemd[1]: Starting modprobe@loop.service... Sep 13 00:47:56.096464 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.096693 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:47:56.096899 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:47:56.109913 systemd[1]: Mounted usr-share-oem.mount. Sep 13 00:47:56.111253 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:47:56.112292 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:47:56.111000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.111000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.114225 systemd[1]: Finished systemd-sysext.service. Sep 13 00:47:56.114000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.116213 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:47:56.116583 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:47:56.116000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.116000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.118744 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:47:56.118987 systemd[1]: Finished modprobe@loop.service. Sep 13 00:47:56.118000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.118000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.123071 systemd[1]: Starting ensure-sysext.service... Sep 13 00:47:56.123818 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:47:56.123922 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.126149 systemd[1]: Starting systemd-tmpfiles-setup.service... Sep 13 00:47:56.139880 systemd[1]: Reloading. Sep 13 00:47:56.154031 systemd-tmpfiles[1676]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Sep 13 00:47:56.156874 systemd-tmpfiles[1676]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 13 00:47:56.161377 systemd-tmpfiles[1676]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 13 00:47:56.220112 /usr/lib/systemd/system-generators/torcx-generator[1696]: time="2025-09-13T00:47:56Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:47:56.220149 /usr/lib/systemd/system-generators/torcx-generator[1696]: time="2025-09-13T00:47:56Z" level=info msg="torcx already run" Sep 13 00:47:56.414581 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:47:56.414916 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:47:56.439211 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:47:56.514180 systemd[1]: Finished systemd-tmpfiles-setup.service. Sep 13 00:47:56.513000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.519900 systemd[1]: Starting audit-rules.service... Sep 13 00:47:56.522461 systemd[1]: Starting clean-ca-certificates.service... Sep 13 00:47:56.525260 systemd[1]: Starting systemd-journal-catalog-update.service... Sep 13 00:47:56.534815 systemd[1]: Starting systemd-resolved.service... Sep 13 00:47:56.538064 systemd[1]: Starting systemd-timesyncd.service... Sep 13 00:47:56.544459 systemd[1]: Starting systemd-update-utmp.service... Sep 13 00:47:56.546430 systemd[1]: Finished clean-ca-certificates.service. Sep 13 00:47:56.545000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.559000 audit[1766]: SYSTEM_BOOT pid=1766 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.568891 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.571907 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:47:56.577312 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:47:56.581243 systemd[1]: Starting modprobe@loop.service... Sep 13 00:47:56.585647 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.586222 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:47:56.586501 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:47:56.589943 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:47:56.590193 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:47:56.590000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.590000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.593030 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:47:56.593256 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:47:56.594000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.594000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.597297 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:47:56.597000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.597000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.603000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.597578 systemd[1]: Finished modprobe@loop.service. Sep 13 00:47:56.599339 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:47:56.599583 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.603134 systemd[1]: Finished systemd-update-utmp.service. Sep 13 00:47:56.611666 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.614726 systemd[1]: Starting modprobe@dm_mod.service... Sep 13 00:47:56.617795 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:47:56.620599 systemd[1]: Starting modprobe@loop.service... Sep 13 00:47:56.621466 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.622188 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:47:56.622618 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:47:56.627721 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:47:56.627958 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:47:56.629000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.629000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.630887 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:47:56.636287 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.638262 systemd[1]: Starting modprobe@drm.service... Sep 13 00:47:56.644325 systemd[1]: Starting modprobe@efi_pstore.service... Sep 13 00:47:56.645214 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.645486 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:47:56.645711 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 13 00:47:56.650000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.650000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.651066 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 13 00:47:56.654000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.651295 systemd[1]: Finished modprobe@dm_mod.service. Sep 13 00:47:56.655757 systemd[1]: Finished ensure-sysext.service. Sep 13 00:47:56.676000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.676000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.677442 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 13 00:47:56.677699 systemd[1]: Finished modprobe@drm.service. Sep 13 00:47:56.681000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.682000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.684000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.684000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.682160 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 13 00:47:56.682481 systemd[1]: Finished modprobe@loop.service. Sep 13 00:47:56.683725 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 13 00:47:56.684800 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 13 00:47:56.685027 systemd[1]: Finished modprobe@efi_pstore.service. Sep 13 00:47:56.685920 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 13 00:47:56.709954 systemd[1]: Finished systemd-journal-catalog-update.service. Sep 13 00:47:56.709000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 13 00:47:56.751000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Sep 13 00:47:56.751000 audit[1804]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffd03aa0f10 a2=420 a3=0 items=0 ppid=1760 pid=1804 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 13 00:47:56.751000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Sep 13 00:47:56.753270 augenrules[1804]: No rules Sep 13 00:47:56.754150 systemd[1]: Finished audit-rules.service. Sep 13 00:47:56.792418 systemd-resolved[1764]: Positive Trust Anchors: Sep 13 00:47:56.792432 systemd-resolved[1764]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 13 00:47:56.792463 systemd-resolved[1764]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 13 00:47:56.793255 systemd[1]: Started systemd-timesyncd.service. Sep 13 00:47:56.793787 systemd[1]: Reached target time-set.target. Sep 13 00:47:56.797921 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:47:56.797952 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 13 00:47:56.838210 systemd-resolved[1764]: Defaulting to hostname 'linux'. Sep 13 00:47:56.840009 systemd[1]: Started systemd-resolved.service. Sep 13 00:47:56.840483 systemd[1]: Reached target network.target. Sep 13 00:47:56.840802 systemd[1]: Reached target nss-lookup.target. Sep 13 00:47:58.829305 systemd-resolved[1764]: Clock change detected. Flushing caches. Sep 13 00:47:58.829531 systemd-timesyncd[1765]: Contacted time server 69.89.207.199:123 (0.flatcar.pool.ntp.org). Sep 13 00:47:58.829783 systemd-timesyncd[1765]: Initial clock synchronization to Sat 2025-09-13 00:47:58.829098 UTC. Sep 13 00:47:58.950325 systemd-networkd[1513]: eth0: Gained IPv6LL Sep 13 00:47:58.952351 systemd[1]: Finished systemd-networkd-wait-online.service. Sep 13 00:47:58.952807 systemd[1]: Reached target network-online.target. Sep 13 00:47:58.991977 ldconfig[1626]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 13 00:47:59.001517 systemd[1]: Finished ldconfig.service. Sep 13 00:47:59.003427 systemd[1]: Starting systemd-update-done.service... Sep 13 00:47:59.011945 systemd[1]: Finished systemd-update-done.service. Sep 13 00:47:59.012468 systemd[1]: Reached target sysinit.target. Sep 13 00:47:59.012924 systemd[1]: Started motdgen.path. Sep 13 00:47:59.013311 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Sep 13 00:47:59.013810 systemd[1]: Started logrotate.timer. Sep 13 00:47:59.014299 systemd[1]: Started mdadm.timer. Sep 13 00:47:59.014636 systemd[1]: Started systemd-tmpfiles-clean.timer. Sep 13 00:47:59.015055 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 13 00:47:59.015104 systemd[1]: Reached target paths.target. Sep 13 00:47:59.015801 systemd[1]: Reached target timers.target. Sep 13 00:47:59.016487 systemd[1]: Listening on dbus.socket. Sep 13 00:47:59.018320 systemd[1]: Starting docker.socket... Sep 13 00:47:59.021110 systemd[1]: Listening on sshd.socket. Sep 13 00:47:59.021742 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:47:59.022438 systemd[1]: Listening on docker.socket. Sep 13 00:47:59.022910 systemd[1]: Reached target sockets.target. Sep 13 00:47:59.023352 systemd[1]: Reached target basic.target. Sep 13 00:47:59.024036 systemd[1]: System is tainted: cgroupsv1 Sep 13 00:47:59.024161 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 13 00:47:59.024194 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 13 00:47:59.025425 systemd[1]: Started amazon-ssm-agent.service. Sep 13 00:47:59.027416 systemd[1]: Starting containerd.service... Sep 13 00:47:59.029651 systemd[1]: Starting coreos-metadata-sshkeys@core.service... Sep 13 00:47:59.032114 systemd[1]: Starting dbus.service... Sep 13 00:47:59.034381 systemd[1]: Starting enable-oem-cloudinit.service... Sep 13 00:47:59.036743 systemd[1]: Starting extend-filesystems.service... Sep 13 00:47:59.042486 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Sep 13 00:47:59.051624 systemd[1]: Starting kubelet.service... Sep 13 00:47:59.054786 systemd[1]: Starting motdgen.service... Sep 13 00:47:59.068360 systemd[1]: Started nvidia.service. Sep 13 00:47:59.088491 jq[1822]: false Sep 13 00:47:59.070989 systemd[1]: Starting prepare-helm.service... Sep 13 00:47:59.080557 systemd[1]: Starting ssh-key-proc-cmdline.service... Sep 13 00:47:59.085061 systemd[1]: Starting sshd-keygen.service... Sep 13 00:47:59.093381 systemd[1]: Starting systemd-logind.service... Sep 13 00:47:59.100293 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 13 00:47:59.100387 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 13 00:47:59.102381 systemd[1]: Starting update-engine.service... Sep 13 00:47:59.114032 systemd[1]: Starting update-ssh-keys-after-ignition.service... Sep 13 00:47:59.130950 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 13 00:47:59.131335 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Sep 13 00:47:59.143182 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 13 00:47:59.143537 systemd[1]: Finished ssh-key-proc-cmdline.service. Sep 13 00:47:59.202115 tar[1852]: linux-amd64/helm Sep 13 00:47:59.227323 dbus-daemon[1821]: [system] SELinux support is enabled Sep 13 00:47:59.228014 systemd[1]: Started dbus.service. Sep 13 00:47:59.231786 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 13 00:47:59.231824 systemd[1]: Reached target system-config.target. Sep 13 00:47:59.237932 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 13 00:47:59.237970 systemd[1]: Reached target user-config.target. Sep 13 00:47:59.251610 extend-filesystems[1823]: Found loop1 Sep 13 00:47:59.257565 jq[1839]: true Sep 13 00:47:59.284828 extend-filesystems[1823]: Found nvme0n1 Sep 13 00:47:59.290424 extend-filesystems[1823]: Found nvme0n1p1 Sep 13 00:47:59.292337 extend-filesystems[1823]: Found nvme0n1p2 Sep 13 00:47:59.293210 extend-filesystems[1823]: Found nvme0n1p3 Sep 13 00:47:59.294556 extend-filesystems[1823]: Found usr Sep 13 00:47:59.295699 extend-filesystems[1823]: Found nvme0n1p4 Sep 13 00:47:59.297534 extend-filesystems[1823]: Found nvme0n1p6 Sep 13 00:47:59.298606 extend-filesystems[1823]: Found nvme0n1p7 Sep 13 00:47:59.298606 extend-filesystems[1823]: Found nvme0n1p9 Sep 13 00:47:59.298606 extend-filesystems[1823]: Checking size of /dev/nvme0n1p9 Sep 13 00:47:59.317777 dbus-daemon[1821]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1513 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Sep 13 00:47:59.322619 systemd[1]: Starting systemd-hostnamed.service... Sep 13 00:47:59.358449 jq[1862]: true Sep 13 00:47:59.367868 systemd[1]: motdgen.service: Deactivated successfully. Sep 13 00:47:59.368309 systemd[1]: Finished motdgen.service. Sep 13 00:47:59.385351 extend-filesystems[1823]: Resized partition /dev/nvme0n1p9 Sep 13 00:47:59.393615 amazon-ssm-agent[1817]: 2025/09/13 00:47:59 Failed to load instance info from vault. RegistrationKey does not exist. Sep 13 00:47:59.398327 amazon-ssm-agent[1817]: Initializing new seelog logger Sep 13 00:47:59.403649 extend-filesystems[1882]: resize2fs 1.46.5 (30-Dec-2021) Sep 13 00:47:59.417183 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Sep 13 00:47:59.419207 amazon-ssm-agent[1817]: New Seelog Logger Creation Complete Sep 13 00:47:59.419336 amazon-ssm-agent[1817]: 2025/09/13 00:47:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 13 00:47:59.419336 amazon-ssm-agent[1817]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 13 00:47:59.420461 amazon-ssm-agent[1817]: 2025/09/13 00:47:59 processing appconfig overrides Sep 13 00:47:59.517210 env[1854]: time="2025-09-13T00:47:59.517129932Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Sep 13 00:47:59.528168 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Sep 13 00:47:59.548588 update_engine[1837]: I0913 00:47:59.533097 1837 main.cc:92] Flatcar Update Engine starting Sep 13 00:47:59.551273 extend-filesystems[1882]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Sep 13 00:47:59.551273 extend-filesystems[1882]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 13 00:47:59.551273 extend-filesystems[1882]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Sep 13 00:47:59.560453 extend-filesystems[1823]: Resized filesystem in /dev/nvme0n1p9 Sep 13 00:47:59.567447 bash[1907]: Updated "/home/core/.ssh/authorized_keys" Sep 13 00:47:59.552225 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 13 00:47:59.567675 update_engine[1837]: I0913 00:47:59.553679 1837 update_check_scheduler.cc:74] Next update check in 6m59s Sep 13 00:47:59.552546 systemd[1]: Finished extend-filesystems.service. Sep 13 00:47:59.556741 systemd[1]: Started update-engine.service. Sep 13 00:47:59.561277 systemd[1]: Started locksmithd.service. Sep 13 00:47:59.563694 systemd[1]: Finished update-ssh-keys-after-ignition.service. Sep 13 00:47:59.771621 systemd[1]: nvidia.service: Deactivated successfully. Sep 13 00:47:59.785020 systemd-logind[1836]: Watching system buttons on /dev/input/event1 (Power Button) Sep 13 00:47:59.785055 systemd-logind[1836]: Watching system buttons on /dev/input/event2 (Sleep Button) Sep 13 00:47:59.785077 systemd-logind[1836]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 13 00:47:59.785350 systemd-logind[1836]: New seat seat0. Sep 13 00:47:59.788516 systemd[1]: Started systemd-logind.service. Sep 13 00:47:59.825471 env[1854]: time="2025-09-13T00:47:59.825406326Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 13 00:47:59.825806 env[1854]: time="2025-09-13T00:47:59.825784162Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:47:59.834596 env[1854]: time="2025-09-13T00:47:59.834533379Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.192-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:47:59.834771 env[1854]: time="2025-09-13T00:47:59.834749776Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:47:59.835332 env[1854]: time="2025-09-13T00:47:59.835293565Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:47:59.836214 env[1854]: time="2025-09-13T00:47:59.836189481Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 13 00:47:59.836343 env[1854]: time="2025-09-13T00:47:59.836324436Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Sep 13 00:47:59.836434 env[1854]: time="2025-09-13T00:47:59.836418375Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 13 00:47:59.836669 env[1854]: time="2025-09-13T00:47:59.836635156Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:47:59.838644 env[1854]: time="2025-09-13T00:47:59.838609776Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 13 00:47:59.839119 env[1854]: time="2025-09-13T00:47:59.839077059Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 13 00:47:59.841224 env[1854]: time="2025-09-13T00:47:59.841194655Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 13 00:47:59.841457 env[1854]: time="2025-09-13T00:47:59.841436563Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Sep 13 00:47:59.843228 env[1854]: time="2025-09-13T00:47:59.843197221Z" level=info msg="metadata content store policy set" policy=shared Sep 13 00:47:59.852291 env[1854]: time="2025-09-13T00:47:59.852228342Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 13 00:47:59.852486 env[1854]: time="2025-09-13T00:47:59.852468134Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 13 00:47:59.852584 env[1854]: time="2025-09-13T00:47:59.852567892Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 13 00:47:59.852707 env[1854]: time="2025-09-13T00:47:59.852689718Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.852889 env[1854]: time="2025-09-13T00:47:59.852871885Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.852987 env[1854]: time="2025-09-13T00:47:59.852971582Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.853079 env[1854]: time="2025-09-13T00:47:59.853064219Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.853185 env[1854]: time="2025-09-13T00:47:59.853168240Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.853282 env[1854]: time="2025-09-13T00:47:59.853266785Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.853376 env[1854]: time="2025-09-13T00:47:59.853361081Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.853475 env[1854]: time="2025-09-13T00:47:59.853459167Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.853588 env[1854]: time="2025-09-13T00:47:59.853571294Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 13 00:47:59.853902 env[1854]: time="2025-09-13T00:47:59.853872085Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 13 00:47:59.858457 env[1854]: time="2025-09-13T00:47:59.858417588Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 13 00:47:59.859329 env[1854]: time="2025-09-13T00:47:59.859300637Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 13 00:47:59.860874 env[1854]: time="2025-09-13T00:47:59.860823496Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.861030 env[1854]: time="2025-09-13T00:47:59.861010897Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 13 00:47:59.866445 env[1854]: time="2025-09-13T00:47:59.866316586Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.874678 env[1854]: time="2025-09-13T00:47:59.874590410Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.875223 env[1854]: time="2025-09-13T00:47:59.875189870Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.875344 env[1854]: time="2025-09-13T00:47:59.875328507Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.875434 env[1854]: time="2025-09-13T00:47:59.875419642Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.875536 env[1854]: time="2025-09-13T00:47:59.875521708Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.877260 env[1854]: time="2025-09-13T00:47:59.877218445Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.877420 env[1854]: time="2025-09-13T00:47:59.877399322Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.877541 env[1854]: time="2025-09-13T00:47:59.877524892Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 13 00:47:59.877865 env[1854]: time="2025-09-13T00:47:59.877844776Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.877977 env[1854]: time="2025-09-13T00:47:59.877960783Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.878081 env[1854]: time="2025-09-13T00:47:59.878065406Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.878187 env[1854]: time="2025-09-13T00:47:59.878170399Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 13 00:47:59.878295 env[1854]: time="2025-09-13T00:47:59.878272650Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Sep 13 00:47:59.878383 env[1854]: time="2025-09-13T00:47:59.878368035Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 13 00:47:59.878483 env[1854]: time="2025-09-13T00:47:59.878466613Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Sep 13 00:47:59.878611 env[1854]: time="2025-09-13T00:47:59.878595190Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 13 00:47:59.879112 env[1854]: time="2025-09-13T00:47:59.879022238Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.880952277Z" level=info msg="Connect containerd service" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.881031795Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.882366340Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.882500070Z" level=info msg="Start subscribing containerd event" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.882565341Z" level=info msg="Start recovering state" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.882647287Z" level=info msg="Start event monitor" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.882675468Z" level=info msg="Start snapshots syncer" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.882687957Z" level=info msg="Start cni network conf syncer for default" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.882701559Z" level=info msg="Start streaming server" Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.883246608Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 13 00:47:59.883463 env[1854]: time="2025-09-13T00:47:59.883342575Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 13 00:47:59.890484 coreos-metadata[1819]: Sep 13 00:47:59.890 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Sep 13 00:47:59.895882 coreos-metadata[1819]: Sep 13 00:47:59.895 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-keys: Attempt #1 Sep 13 00:47:59.897464 coreos-metadata[1819]: Sep 13 00:47:59.897 INFO Fetch successful Sep 13 00:47:59.897464 coreos-metadata[1819]: Sep 13 00:47:59.897 INFO Fetching http://169.254.169.254/2019-10-01/meta-data/public-keys/0/openssh-key: Attempt #1 Sep 13 00:47:59.898214 coreos-metadata[1819]: Sep 13 00:47:59.898 INFO Fetch successful Sep 13 00:47:59.900368 unknown[1819]: wrote ssh authorized keys file for user: core Sep 13 00:47:59.912619 env[1854]: time="2025-09-13T00:47:59.912577425Z" level=info msg="containerd successfully booted in 0.406776s" Sep 13 00:47:59.912702 systemd[1]: Started containerd.service. Sep 13 00:47:59.916225 update-ssh-keys[1939]: Updated "/home/core/.ssh/authorized_keys" Sep 13 00:47:59.916585 systemd[1]: Finished coreos-metadata-sshkeys@core.service. Sep 13 00:47:59.963513 dbus-daemon[1821]: [system] Successfully activated service 'org.freedesktop.hostname1' Sep 13 00:47:59.963693 systemd[1]: Started systemd-hostnamed.service. Sep 13 00:47:59.971378 dbus-daemon[1821]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.6' (uid=0 pid=1870 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Sep 13 00:47:59.975454 systemd[1]: Starting polkit.service... Sep 13 00:47:59.999948 polkitd[1962]: Started polkitd version 121 Sep 13 00:48:00.027612 polkitd[1962]: Loading rules from directory /etc/polkit-1/rules.d Sep 13 00:48:00.032334 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Create new startup processor Sep 13 00:48:00.033747 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [LongRunningPluginsManager] registered plugins: {} Sep 13 00:48:00.037362 polkitd[1962]: Loading rules from directory /usr/share/polkit-1/rules.d Sep 13 00:48:00.041489 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing bookkeeping folders Sep 13 00:48:00.041660 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO removing the completed state files Sep 13 00:48:00.041888 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing bookkeeping folders for long running plugins Sep 13 00:48:00.041982 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing replies folder for MDS reply requests that couldn't reach the service Sep 13 00:48:00.042084 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing healthcheck folders for long running plugins Sep 13 00:48:00.042209 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing locations for inventory plugin Sep 13 00:48:00.042293 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing default location for custom inventory Sep 13 00:48:00.042398 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing default location for file inventory Sep 13 00:48:00.042471 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Initializing default location for role inventory Sep 13 00:48:00.042538 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Init the cloudwatchlogs publisher Sep 13 00:48:00.042611 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:configureDocker Sep 13 00:48:00.042675 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:runDockerAction Sep 13 00:48:00.042763 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:refreshAssociation Sep 13 00:48:00.042829 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:downloadContent Sep 13 00:48:00.043064 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:runDocument Sep 13 00:48:00.043167 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:softwareInventory Sep 13 00:48:00.043238 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:runPowerShellScript Sep 13 00:48:00.043299 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:updateSsmAgent Sep 13 00:48:00.043585 polkitd[1962]: Finished loading, compiling and executing 2 rules Sep 13 00:48:00.044320 dbus-daemon[1821]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Sep 13 00:48:00.044522 systemd[1]: Started polkit.service. Sep 13 00:48:00.044682 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform independent plugin aws:configurePackage Sep 13 00:48:00.044755 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Successfully loaded platform dependent plugin aws:runShellScript Sep 13 00:48:00.044818 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO Starting Agent: amazon-ssm-agent - v2.3.1319.0 Sep 13 00:48:00.044889 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO OS: linux, Arch: amd64 Sep 13 00:48:00.045363 polkitd[1962]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Sep 13 00:48:00.047057 amazon-ssm-agent[1817]: datastore file /var/lib/amazon/ssm/i-0543061661a412af4/longrunningplugins/datastore/store doesn't exist - no long running plugins to execute Sep 13 00:48:00.062942 systemd-hostnamed[1870]: Hostname set to (transient) Sep 13 00:48:00.063074 systemd-resolved[1764]: System hostname changed to 'ip-172-31-19-167'. Sep 13 00:48:00.134932 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] Starting session document processing engine... Sep 13 00:48:00.229676 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] [EngineProcessor] Starting Sep 13 00:48:00.324390 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] SSM Agent is trying to setup control channel for Session Manager module. Sep 13 00:48:00.418829 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] Setting up websocket for controlchannel for instance: i-0543061661a412af4, requestId: 83de83ac-995d-47be-9c68-8d2ab8c92135 Sep 13 00:48:00.513517 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] Starting document processing engine... Sep 13 00:48:00.575966 sshd_keygen[1863]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 13 00:48:00.608407 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] [EngineProcessor] Starting Sep 13 00:48:00.613187 systemd[1]: Finished sshd-keygen.service. Sep 13 00:48:00.616072 systemd[1]: Starting issuegen.service... Sep 13 00:48:00.633211 systemd[1]: issuegen.service: Deactivated successfully. Sep 13 00:48:00.633542 systemd[1]: Finished issuegen.service. Sep 13 00:48:00.636718 systemd[1]: Starting systemd-user-sessions.service... Sep 13 00:48:00.648230 systemd[1]: Finished systemd-user-sessions.service. Sep 13 00:48:00.651242 systemd[1]: Started getty@tty1.service. Sep 13 00:48:00.657481 systemd[1]: Started serial-getty@ttyS0.service. Sep 13 00:48:00.662561 systemd[1]: Reached target getty.target. Sep 13 00:48:00.703503 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] [EngineProcessor] Initial processing Sep 13 00:48:00.715685 tar[1852]: linux-amd64/LICENSE Sep 13 00:48:00.716107 tar[1852]: linux-amd64/README.md Sep 13 00:48:00.723666 systemd[1]: Finished prepare-helm.service. Sep 13 00:48:00.804290 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] Starting message polling Sep 13 00:48:00.863094 locksmithd[1915]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 13 00:48:00.899991 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] Starting send replies to MDS Sep 13 00:48:00.995645 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [instanceID=i-0543061661a412af4] Starting association polling Sep 13 00:48:01.091564 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] [Association] [EngineProcessor] Starting Sep 13 00:48:01.188871 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] [Association] Launching response handler Sep 13 00:48:01.284756 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] [Association] [EngineProcessor] Initial processing Sep 13 00:48:01.390879 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] [Association] Initializing association scheduling service Sep 13 00:48:01.487592 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessagingDeliveryService] [Association] Association scheduling service initialized Sep 13 00:48:01.584491 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] listening reply. Sep 13 00:48:01.681570 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [HealthCheck] HealthCheck reporting agent health. Sep 13 00:48:01.779134 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [OfflineService] Starting document processing engine... Sep 13 00:48:01.876527 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [OfflineService] [EngineProcessor] Starting Sep 13 00:48:01.974177 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [OfflineService] [EngineProcessor] Initial processing Sep 13 00:48:02.072098 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [OfflineService] Starting message polling Sep 13 00:48:02.170724 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [OfflineService] Starting send replies to MDS Sep 13 00:48:02.269004 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [LongRunningPluginsManager] starting long running plugin manager Sep 13 00:48:02.367645 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [LongRunningPluginsManager] there aren't any long running plugin to execute Sep 13 00:48:02.429393 systemd[1]: Started kubelet.service. Sep 13 00:48:02.430926 systemd[1]: Reached target multi-user.target. Sep 13 00:48:02.459824 systemd[1]: Starting systemd-update-utmp-runlevel.service... Sep 13 00:48:02.478444 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [LongRunningPluginsManager] There are no long running plugins currently getting executed - skipping their healthcheck Sep 13 00:48:02.524369 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Sep 13 00:48:02.524713 systemd[1]: Finished systemd-update-utmp-runlevel.service. Sep 13 00:48:02.555611 systemd[1]: Startup finished in 7.582s (kernel) + 10.878s (userspace) = 18.460s. Sep 13 00:48:02.575801 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [StartupProcessor] Executing startup processor tasks Sep 13 00:48:02.676109 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [StartupProcessor] Write to serial port: Amazon SSM Agent v2.3.1319.0 is running Sep 13 00:48:02.775299 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [StartupProcessor] Write to serial port: OsProductName: Flatcar Container Linux by Kinvolk Sep 13 00:48:02.874733 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [StartupProcessor] Write to serial port: OsVersion: 3510.3.8 Sep 13 00:48:02.975790 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] Opening websocket connection to: wss://ssmmessages.us-west-2.amazonaws.com/v1/control-channel/i-0543061661a412af4?role=subscribe&stream=input Sep 13 00:48:03.075539 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] Successfully opened websocket connection to: wss://ssmmessages.us-west-2.amazonaws.com/v1/control-channel/i-0543061661a412af4?role=subscribe&stream=input Sep 13 00:48:03.175653 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] Starting receiving message from control channel Sep 13 00:48:03.277816 amazon-ssm-agent[1817]: 2025-09-13 00:48:00 INFO [MessageGatewayService] [EngineProcessor] Initial processing Sep 13 00:48:04.494387 kubelet[2053]: E0913 00:48:04.493864 2053 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:48:04.496645 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:48:04.496883 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:48:08.735172 systemd[1]: Created slice system-sshd.slice. Sep 13 00:48:08.736577 systemd[1]: Started sshd@0-172.31.19.167:22-147.75.109.163:54738.service. Sep 13 00:48:08.950702 sshd[2062]: Accepted publickey for core from 147.75.109.163 port 54738 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:48:08.954199 sshd[2062]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:48:08.970052 systemd[1]: Created slice user-500.slice. Sep 13 00:48:08.972365 systemd[1]: Starting user-runtime-dir@500.service... Sep 13 00:48:08.978255 systemd-logind[1836]: New session 1 of user core. Sep 13 00:48:08.985815 systemd[1]: Finished user-runtime-dir@500.service. Sep 13 00:48:08.987590 systemd[1]: Starting user@500.service... Sep 13 00:48:08.993892 (systemd)[2067]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:48:09.100050 systemd[2067]: Queued start job for default target default.target. Sep 13 00:48:09.100901 systemd[2067]: Reached target paths.target. Sep 13 00:48:09.100933 systemd[2067]: Reached target sockets.target. Sep 13 00:48:09.100952 systemd[2067]: Reached target timers.target. Sep 13 00:48:09.100970 systemd[2067]: Reached target basic.target. Sep 13 00:48:09.101186 systemd[1]: Started user@500.service. Sep 13 00:48:09.102199 systemd[1]: Started session-1.scope. Sep 13 00:48:09.104064 systemd[2067]: Reached target default.target. Sep 13 00:48:09.108085 systemd[2067]: Startup finished in 101ms. Sep 13 00:48:09.243735 systemd[1]: Started sshd@1-172.31.19.167:22-147.75.109.163:54748.service. Sep 13 00:48:09.414937 sshd[2076]: Accepted publickey for core from 147.75.109.163 port 54748 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:48:09.416664 sshd[2076]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:48:09.421271 systemd-logind[1836]: New session 2 of user core. Sep 13 00:48:09.422746 systemd[1]: Started session-2.scope. Sep 13 00:48:09.551026 sshd[2076]: pam_unix(sshd:session): session closed for user core Sep 13 00:48:09.554058 systemd[1]: sshd@1-172.31.19.167:22-147.75.109.163:54748.service: Deactivated successfully. Sep 13 00:48:09.554947 systemd-logind[1836]: Session 2 logged out. Waiting for processes to exit. Sep 13 00:48:09.555015 systemd[1]: session-2.scope: Deactivated successfully. Sep 13 00:48:09.556188 systemd-logind[1836]: Removed session 2. Sep 13 00:48:09.574250 systemd[1]: Started sshd@2-172.31.19.167:22-147.75.109.163:54750.service. Sep 13 00:48:09.739324 sshd[2083]: Accepted publickey for core from 147.75.109.163 port 54750 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:48:09.740977 sshd[2083]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:48:09.746202 systemd-logind[1836]: New session 3 of user core. Sep 13 00:48:09.746857 systemd[1]: Started session-3.scope. Sep 13 00:48:09.867091 sshd[2083]: pam_unix(sshd:session): session closed for user core Sep 13 00:48:09.870818 systemd[1]: sshd@2-172.31.19.167:22-147.75.109.163:54750.service: Deactivated successfully. Sep 13 00:48:09.872387 systemd[1]: session-3.scope: Deactivated successfully. Sep 13 00:48:09.872419 systemd-logind[1836]: Session 3 logged out. Waiting for processes to exit. Sep 13 00:48:09.873906 systemd-logind[1836]: Removed session 3. Sep 13 00:48:09.892292 systemd[1]: Started sshd@3-172.31.19.167:22-147.75.109.163:54764.service. Sep 13 00:48:10.061204 sshd[2090]: Accepted publickey for core from 147.75.109.163 port 54764 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:48:10.062758 sshd[2090]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:48:10.067851 systemd-logind[1836]: New session 4 of user core. Sep 13 00:48:10.068418 systemd[1]: Started session-4.scope. Sep 13 00:48:10.198821 sshd[2090]: pam_unix(sshd:session): session closed for user core Sep 13 00:48:10.202522 systemd[1]: sshd@3-172.31.19.167:22-147.75.109.163:54764.service: Deactivated successfully. Sep 13 00:48:10.204016 systemd[1]: session-4.scope: Deactivated successfully. Sep 13 00:48:10.204108 systemd-logind[1836]: Session 4 logged out. Waiting for processes to exit. Sep 13 00:48:10.205584 systemd-logind[1836]: Removed session 4. Sep 13 00:48:10.223953 systemd[1]: Started sshd@4-172.31.19.167:22-147.75.109.163:53340.service. Sep 13 00:48:10.395027 sshd[2097]: Accepted publickey for core from 147.75.109.163 port 53340 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:48:10.396569 sshd[2097]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:48:10.402736 systemd[1]: Started session-5.scope. Sep 13 00:48:10.403075 systemd-logind[1836]: New session 5 of user core. Sep 13 00:48:10.536918 sudo[2101]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 13 00:48:10.537180 sudo[2101]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 13 00:48:10.562673 systemd[1]: Starting docker.service... Sep 13 00:48:10.602969 env[2111]: time="2025-09-13T00:48:10.602905674Z" level=info msg="Starting up" Sep 13 00:48:10.605438 env[2111]: time="2025-09-13T00:48:10.605192147Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 13 00:48:10.605438 env[2111]: time="2025-09-13T00:48:10.605220732Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 13 00:48:10.605438 env[2111]: time="2025-09-13T00:48:10.605273702Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 13 00:48:10.605438 env[2111]: time="2025-09-13T00:48:10.605289158Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 13 00:48:10.608283 env[2111]: time="2025-09-13T00:48:10.608245995Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 13 00:48:10.608283 env[2111]: time="2025-09-13T00:48:10.608269855Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 13 00:48:10.608468 env[2111]: time="2025-09-13T00:48:10.608292749Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 13 00:48:10.608468 env[2111]: time="2025-09-13T00:48:10.608306019Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 13 00:48:10.617001 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2111304591-merged.mount: Deactivated successfully. Sep 13 00:48:10.647202 env[2111]: time="2025-09-13T00:48:10.647138770Z" level=warning msg="Your kernel does not support cgroup blkio weight" Sep 13 00:48:10.647396 env[2111]: time="2025-09-13T00:48:10.647383925Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" Sep 13 00:48:10.647628 env[2111]: time="2025-09-13T00:48:10.647613620Z" level=info msg="Loading containers: start." Sep 13 00:48:10.856179 kernel: Initializing XFRM netlink socket Sep 13 00:48:10.922342 env[2111]: time="2025-09-13T00:48:10.922278752Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Sep 13 00:48:10.923540 (udev-worker)[2121]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:48:11.034470 systemd-networkd[1513]: docker0: Link UP Sep 13 00:48:11.052019 env[2111]: time="2025-09-13T00:48:11.051969377Z" level=info msg="Loading containers: done." Sep 13 00:48:11.067461 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1062177937-merged.mount: Deactivated successfully. Sep 13 00:48:11.074109 env[2111]: time="2025-09-13T00:48:11.074055121Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 13 00:48:11.074369 env[2111]: time="2025-09-13T00:48:11.074343460Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Sep 13 00:48:11.074507 env[2111]: time="2025-09-13T00:48:11.074484742Z" level=info msg="Daemon has completed initialization" Sep 13 00:48:11.094781 systemd[1]: Started docker.service. Sep 13 00:48:11.102095 env[2111]: time="2025-09-13T00:48:11.102036522Z" level=info msg="API listen on /run/docker.sock" Sep 13 00:48:12.401998 env[1854]: time="2025-09-13T00:48:12.401951737Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\"" Sep 13 00:48:12.942733 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2915644362.mount: Deactivated successfully. Sep 13 00:48:14.408836 env[1854]: time="2025-09-13T00:48:14.408767069Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:14.412107 env[1854]: time="2025-09-13T00:48:14.411588176Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:14.424341 env[1854]: time="2025-09-13T00:48:14.424288288Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:14.430983 env[1854]: time="2025-09-13T00:48:14.430932301Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:9abeb8a2d3e53e356d1f2e5d5dc2081cf28f23242651b0552c9e38f4a7ae960e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:14.432796 env[1854]: time="2025-09-13T00:48:14.432729619Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.13\" returns image reference \"sha256:368da3301bb03f4bef9f7dc2084f5fc5954b0ac1bf1e49ca502e3a7604011e54\"" Sep 13 00:48:14.433874 env[1854]: time="2025-09-13T00:48:14.433839288Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\"" Sep 13 00:48:14.724483 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 13 00:48:14.724740 systemd[1]: Stopped kubelet.service. Sep 13 00:48:14.727184 systemd[1]: Starting kubelet.service... Sep 13 00:48:14.953738 systemd[1]: Started kubelet.service. Sep 13 00:48:15.045129 kubelet[2240]: E0913 00:48:15.045004 2240 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:48:15.048490 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:48:15.048706 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:48:16.092113 env[1854]: time="2025-09-13T00:48:16.092060277Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:16.095182 env[1854]: time="2025-09-13T00:48:16.095121524Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:16.097820 env[1854]: time="2025-09-13T00:48:16.097772035Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:16.100502 env[1854]: time="2025-09-13T00:48:16.100458705Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:facc91288697a288a691520949fe4eec40059ef065c89da8e10481d14e131b09,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:16.101474 env[1854]: time="2025-09-13T00:48:16.101427608Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.13\" returns image reference \"sha256:cbd19105c6bcbedf394f51c8bb963def5195c300fc7d04bc39d48d14d23c0ff0\"" Sep 13 00:48:16.102201 env[1854]: time="2025-09-13T00:48:16.102180248Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\"" Sep 13 00:48:17.454638 env[1854]: time="2025-09-13T00:48:17.454580242Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:17.457220 env[1854]: time="2025-09-13T00:48:17.457174002Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:17.459294 env[1854]: time="2025-09-13T00:48:17.459228280Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:17.461759 env[1854]: time="2025-09-13T00:48:17.461711307Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:c5ce150dcce2419fdef9f9875fef43014355ccebf937846ed3a2971953f9b241,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:17.462791 env[1854]: time="2025-09-13T00:48:17.462730127Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.13\" returns image reference \"sha256:d019d989e2b1f0b08ea7eebd4dd7673bdd6ba2218a3c5a6bd53f6848d5fc1af6\"" Sep 13 00:48:17.463548 env[1854]: time="2025-09-13T00:48:17.463519200Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\"" Sep 13 00:48:18.515417 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1739608745.mount: Deactivated successfully. Sep 13 00:48:19.200118 env[1854]: time="2025-09-13T00:48:19.200072621Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:19.202117 env[1854]: time="2025-09-13T00:48:19.202061375Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:19.204250 env[1854]: time="2025-09-13T00:48:19.204209464Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.31.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:19.205557 env[1854]: time="2025-09-13T00:48:19.205525453Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:a39637326e88d128d38da6ff2b2ceb4e856475887bfcb5f7a55734d4f63d9fae,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:19.205996 env[1854]: time="2025-09-13T00:48:19.205959572Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.13\" returns image reference \"sha256:21d97a49eeb0b08ecaba421a84a79ca44cf2bc57773c085bbfda537488790ad7\"" Sep 13 00:48:19.206736 env[1854]: time="2025-09-13T00:48:19.206609617Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 13 00:48:19.751235 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4023638291.mount: Deactivated successfully. Sep 13 00:48:20.724738 env[1854]: time="2025-09-13T00:48:20.724680624Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:20.727357 env[1854]: time="2025-09-13T00:48:20.727304966Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:20.729435 env[1854]: time="2025-09-13T00:48:20.729402836Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:20.731266 env[1854]: time="2025-09-13T00:48:20.731223979Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:20.732320 env[1854]: time="2025-09-13T00:48:20.732275206Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 13 00:48:20.733673 env[1854]: time="2025-09-13T00:48:20.733637572Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 13 00:48:21.182758 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3311608087.mount: Deactivated successfully. Sep 13 00:48:21.190949 env[1854]: time="2025-09-13T00:48:21.190903056Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:21.193461 env[1854]: time="2025-09-13T00:48:21.193409389Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:21.195168 env[1854]: time="2025-09-13T00:48:21.195094897Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:21.197183 env[1854]: time="2025-09-13T00:48:21.197119057Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:21.197519 env[1854]: time="2025-09-13T00:48:21.197488236Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 13 00:48:21.197992 env[1854]: time="2025-09-13T00:48:21.197954775Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Sep 13 00:48:21.672875 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1536708362.mount: Deactivated successfully. Sep 13 00:48:23.992520 env[1854]: time="2025-09-13T00:48:23.992441285Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:23.995291 env[1854]: time="2025-09-13T00:48:23.995245739Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:23.997927 env[1854]: time="2025-09-13T00:48:23.997877966Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:23.999710 env[1854]: time="2025-09-13T00:48:23.999576954Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:24.000675 env[1854]: time="2025-09-13T00:48:24.000640270Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Sep 13 00:48:25.224139 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 13 00:48:25.224440 systemd[1]: Stopped kubelet.service. Sep 13 00:48:25.227739 systemd[1]: Starting kubelet.service... Sep 13 00:48:26.019893 systemd[1]: Started kubelet.service. Sep 13 00:48:26.104050 kubelet[2272]: E0913 00:48:26.103999 2272 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 13 00:48:26.106879 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 13 00:48:26.107102 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 13 00:48:26.782140 amazon-ssm-agent[1817]: 2025-09-13 00:48:26 INFO [MessagingDeliveryService] [Association] No associations on boot. Requerying for associations after 30 seconds. Sep 13 00:48:27.281164 systemd[1]: Stopped kubelet.service. Sep 13 00:48:27.284410 systemd[1]: Starting kubelet.service... Sep 13 00:48:27.324237 systemd[1]: Reloading. Sep 13 00:48:27.416544 /usr/lib/systemd/system-generators/torcx-generator[2306]: time="2025-09-13T00:48:27Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:48:27.418364 /usr/lib/systemd/system-generators/torcx-generator[2306]: time="2025-09-13T00:48:27Z" level=info msg="torcx already run" Sep 13 00:48:27.571841 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:48:27.572782 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:48:27.602811 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:48:27.709202 systemd[1]: Started kubelet.service. Sep 13 00:48:27.712045 systemd[1]: Stopping kubelet.service... Sep 13 00:48:27.712995 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 00:48:27.713366 systemd[1]: Stopped kubelet.service. Sep 13 00:48:27.716411 systemd[1]: Starting kubelet.service... Sep 13 00:48:27.931093 systemd[1]: Started kubelet.service. Sep 13 00:48:27.991775 kubelet[2383]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:48:27.992212 kubelet[2383]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 13 00:48:27.992273 kubelet[2383]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:48:27.992431 kubelet[2383]: I0913 00:48:27.992406 2383 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 00:48:28.589687 kubelet[2383]: I0913 00:48:28.589642 2383 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 13 00:48:28.589687 kubelet[2383]: I0913 00:48:28.589676 2383 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 00:48:28.590019 kubelet[2383]: I0913 00:48:28.589994 2383 server.go:934] "Client rotation is on, will bootstrap in background" Sep 13 00:48:28.655762 kubelet[2383]: I0913 00:48:28.655525 2383 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 00:48:28.656257 kubelet[2383]: E0913 00:48:28.656233 2383 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.19.167:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:28.667253 kubelet[2383]: E0913 00:48:28.667223 2383 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 00:48:28.667433 kubelet[2383]: I0913 00:48:28.667406 2383 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 00:48:28.671763 kubelet[2383]: I0913 00:48:28.671688 2383 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 00:48:28.672027 kubelet[2383]: I0913 00:48:28.672009 2383 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 13 00:48:28.672160 kubelet[2383]: I0913 00:48:28.672122 2383 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 00:48:28.672348 kubelet[2383]: I0913 00:48:28.672167 2383 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-19-167","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Sep 13 00:48:28.672457 kubelet[2383]: I0913 00:48:28.672350 2383 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 00:48:28.672457 kubelet[2383]: I0913 00:48:28.672359 2383 container_manager_linux.go:300] "Creating device plugin manager" Sep 13 00:48:28.672457 kubelet[2383]: I0913 00:48:28.672456 2383 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:48:28.685884 kubelet[2383]: I0913 00:48:28.685839 2383 kubelet.go:408] "Attempting to sync node with API server" Sep 13 00:48:28.685884 kubelet[2383]: I0913 00:48:28.685890 2383 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 00:48:28.686067 kubelet[2383]: I0913 00:48:28.685926 2383 kubelet.go:314] "Adding apiserver pod source" Sep 13 00:48:28.686067 kubelet[2383]: I0913 00:48:28.685947 2383 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 00:48:28.698166 kubelet[2383]: W0913 00:48:28.698092 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.19.167:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-19-167&limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:28.698547 kubelet[2383]: E0913 00:48:28.698338 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.19.167:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-19-167&limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:28.699862 kubelet[2383]: W0913 00:48:28.699807 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.19.167:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:28.699959 kubelet[2383]: E0913 00:48:28.699864 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.19.167:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:28.700138 kubelet[2383]: I0913 00:48:28.700121 2383 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 13 00:48:28.700984 kubelet[2383]: I0913 00:48:28.700952 2383 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 00:48:28.712845 kubelet[2383]: W0913 00:48:28.712804 2383 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 13 00:48:28.721909 kubelet[2383]: I0913 00:48:28.721870 2383 server.go:1274] "Started kubelet" Sep 13 00:48:28.728002 kubelet[2383]: I0913 00:48:28.727945 2383 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 00:48:28.728928 kubelet[2383]: I0913 00:48:28.728887 2383 server.go:449] "Adding debug handlers to kubelet server" Sep 13 00:48:28.737180 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Sep 13 00:48:28.737308 kubelet[2383]: I0913 00:48:28.737019 2383 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 00:48:28.740604 kubelet[2383]: I0913 00:48:28.740558 2383 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 00:48:28.740776 kubelet[2383]: I0913 00:48:28.740760 2383 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 00:48:28.743771 kubelet[2383]: E0913 00:48:28.740960 2383 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.19.167:6443/api/v1/namespaces/default/events\": dial tcp 172.31.19.167:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-19-167.1864b120bb5ceeb9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-19-167,UID:ip-172-31-19-167,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-19-167,},FirstTimestamp:2025-09-13 00:48:28.721835705 +0000 UTC m=+0.779177588,LastTimestamp:2025-09-13 00:48:28.721835705 +0000 UTC m=+0.779177588,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-19-167,}" Sep 13 00:48:28.745459 kubelet[2383]: I0913 00:48:28.745425 2383 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 00:48:28.747053 kubelet[2383]: I0913 00:48:28.746930 2383 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 13 00:48:28.747202 kubelet[2383]: E0913 00:48:28.747186 2383 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-19-167\" not found" Sep 13 00:48:28.747769 kubelet[2383]: I0913 00:48:28.747748 2383 factory.go:221] Registration of the systemd container factory successfully Sep 13 00:48:28.747953 kubelet[2383]: I0913 00:48:28.747938 2383 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 00:48:28.750332 kubelet[2383]: I0913 00:48:28.750310 2383 factory.go:221] Registration of the containerd container factory successfully Sep 13 00:48:28.753123 kubelet[2383]: I0913 00:48:28.753081 2383 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 13 00:48:28.753263 kubelet[2383]: I0913 00:48:28.753167 2383 reconciler.go:26] "Reconciler: start to sync state" Sep 13 00:48:28.761632 kubelet[2383]: E0913 00:48:28.761595 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.19.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-19-167?timeout=10s\": dial tcp 172.31.19.167:6443: connect: connection refused" interval="200ms" Sep 13 00:48:28.767134 kubelet[2383]: I0913 00:48:28.767091 2383 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 00:48:28.768534 kubelet[2383]: I0913 00:48:28.768498 2383 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 00:48:28.768534 kubelet[2383]: I0913 00:48:28.768529 2383 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 13 00:48:28.768662 kubelet[2383]: I0913 00:48:28.768548 2383 kubelet.go:2321] "Starting kubelet main sync loop" Sep 13 00:48:28.768662 kubelet[2383]: E0913 00:48:28.768593 2383 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 00:48:28.779751 kubelet[2383]: W0913 00:48:28.779697 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.19.167:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:28.779942 kubelet[2383]: E0913 00:48:28.779924 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.19.167:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:28.780079 kubelet[2383]: W0913 00:48:28.780054 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.19.167:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:28.780165 kubelet[2383]: E0913 00:48:28.780135 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.19.167:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:28.782690 kubelet[2383]: E0913 00:48:28.782664 2383 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 00:48:28.791065 kubelet[2383]: I0913 00:48:28.791010 2383 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 13 00:48:28.791065 kubelet[2383]: I0913 00:48:28.791046 2383 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 13 00:48:28.791231 kubelet[2383]: I0913 00:48:28.791081 2383 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:48:28.793703 kubelet[2383]: I0913 00:48:28.793666 2383 policy_none.go:49] "None policy: Start" Sep 13 00:48:28.794399 kubelet[2383]: I0913 00:48:28.794354 2383 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 13 00:48:28.794399 kubelet[2383]: I0913 00:48:28.794392 2383 state_mem.go:35] "Initializing new in-memory state store" Sep 13 00:48:28.800181 kubelet[2383]: I0913 00:48:28.799696 2383 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 00:48:28.800181 kubelet[2383]: I0913 00:48:28.799877 2383 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 00:48:28.800181 kubelet[2383]: I0913 00:48:28.799889 2383 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 00:48:28.801435 kubelet[2383]: I0913 00:48:28.801418 2383 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 00:48:28.803004 kubelet[2383]: E0913 00:48:28.802989 2383 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-19-167\" not found" Sep 13 00:48:28.915074 kubelet[2383]: I0913 00:48:28.912564 2383 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-19-167" Sep 13 00:48:28.915074 kubelet[2383]: E0913 00:48:28.913256 2383 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.19.167:6443/api/v1/nodes\": dial tcp 172.31.19.167:6443: connect: connection refused" node="ip-172-31-19-167" Sep 13 00:48:28.962970 kubelet[2383]: E0913 00:48:28.962926 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.19.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-19-167?timeout=10s\": dial tcp 172.31.19.167:6443: connect: connection refused" interval="400ms" Sep 13 00:48:29.054649 kubelet[2383]: I0913 00:48:29.054586 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1e30b5c1e43171410f5dad087060f636-k8s-certs\") pod \"kube-apiserver-ip-172-31-19-167\" (UID: \"1e30b5c1e43171410f5dad087060f636\") " pod="kube-system/kube-apiserver-ip-172-31-19-167" Sep 13 00:48:29.054649 kubelet[2383]: I0913 00:48:29.054650 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-k8s-certs\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:29.055213 kubelet[2383]: I0913 00:48:29.054683 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-kubeconfig\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:29.055213 kubelet[2383]: I0913 00:48:29.054721 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:29.055213 kubelet[2383]: I0913 00:48:29.054752 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d7c4cbbdb680656b071548db8d01d06d-kubeconfig\") pod \"kube-scheduler-ip-172-31-19-167\" (UID: \"d7c4cbbdb680656b071548db8d01d06d\") " pod="kube-system/kube-scheduler-ip-172-31-19-167" Sep 13 00:48:29.055213 kubelet[2383]: I0913 00:48:29.054778 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1e30b5c1e43171410f5dad087060f636-ca-certs\") pod \"kube-apiserver-ip-172-31-19-167\" (UID: \"1e30b5c1e43171410f5dad087060f636\") " pod="kube-system/kube-apiserver-ip-172-31-19-167" Sep 13 00:48:29.055213 kubelet[2383]: I0913 00:48:29.054806 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1e30b5c1e43171410f5dad087060f636-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-19-167\" (UID: \"1e30b5c1e43171410f5dad087060f636\") " pod="kube-system/kube-apiserver-ip-172-31-19-167" Sep 13 00:48:29.055364 kubelet[2383]: I0913 00:48:29.054849 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-ca-certs\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:29.055364 kubelet[2383]: I0913 00:48:29.054877 2383 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:29.115184 kubelet[2383]: I0913 00:48:29.115158 2383 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-19-167" Sep 13 00:48:29.115747 kubelet[2383]: E0913 00:48:29.115713 2383 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.19.167:6443/api/v1/nodes\": dial tcp 172.31.19.167:6443: connect: connection refused" node="ip-172-31-19-167" Sep 13 00:48:29.190317 env[1854]: time="2025-09-13T00:48:29.189411941Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-19-167,Uid:1e30b5c1e43171410f5dad087060f636,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:29.208496 env[1854]: time="2025-09-13T00:48:29.208446438Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-19-167,Uid:d7c4cbbdb680656b071548db8d01d06d,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:29.213200 env[1854]: time="2025-09-13T00:48:29.213135285Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-19-167,Uid:ed7fb32cfe077e607f31e48642251dfc,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:29.363475 kubelet[2383]: E0913 00:48:29.363385 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.19.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-19-167?timeout=10s\": dial tcp 172.31.19.167:6443: connect: connection refused" interval="800ms" Sep 13 00:48:29.517964 kubelet[2383]: I0913 00:48:29.517921 2383 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-19-167" Sep 13 00:48:29.518442 kubelet[2383]: E0913 00:48:29.518409 2383 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.19.167:6443/api/v1/nodes\": dial tcp 172.31.19.167:6443: connect: connection refused" node="ip-172-31-19-167" Sep 13 00:48:29.634696 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3032504678.mount: Deactivated successfully. Sep 13 00:48:29.643015 env[1854]: time="2025-09-13T00:48:29.642964517Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.648063 env[1854]: time="2025-09-13T00:48:29.647546918Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.648690 env[1854]: time="2025-09-13T00:48:29.648656287Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.650884 env[1854]: time="2025-09-13T00:48:29.650843299Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.652243 env[1854]: time="2025-09-13T00:48:29.652200721Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.653260 env[1854]: time="2025-09-13T00:48:29.653221283Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.654627 env[1854]: time="2025-09-13T00:48:29.654569581Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.658632 env[1854]: time="2025-09-13T00:48:29.658563749Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.660741 env[1854]: time="2025-09-13T00:48:29.660691099Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.662619 env[1854]: time="2025-09-13T00:48:29.662579265Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.668671 env[1854]: time="2025-09-13T00:48:29.668623520Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.670491 env[1854]: time="2025-09-13T00:48:29.670449877Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:29.687188 env[1854]: time="2025-09-13T00:48:29.687072356Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:48:29.688129 env[1854]: time="2025-09-13T00:48:29.687132741Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:48:29.688129 env[1854]: time="2025-09-13T00:48:29.687166864Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:48:29.688129 env[1854]: time="2025-09-13T00:48:29.687453408Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/7def50844e0e64aa3ba34fce3c5b6ee9e20193733487a0af58f0d072379b7f73 pid=2420 runtime=io.containerd.runc.v2 Sep 13 00:48:29.738950 env[1854]: time="2025-09-13T00:48:29.733181889Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:48:29.738950 env[1854]: time="2025-09-13T00:48:29.733344549Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:48:29.738950 env[1854]: time="2025-09-13T00:48:29.733429991Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:48:29.738950 env[1854]: time="2025-09-13T00:48:29.733668153Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/fec989cef7de0aefcf0cea7bbc9e4b70c1b79e7dc0968b572dc65abc939cb360 pid=2448 runtime=io.containerd.runc.v2 Sep 13 00:48:29.760300 kubelet[2383]: W0913 00:48:29.760123 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.19.167:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:29.760300 kubelet[2383]: E0913 00:48:29.760242 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.19.167:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:29.761826 env[1854]: time="2025-09-13T00:48:29.761498138Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:48:29.761826 env[1854]: time="2025-09-13T00:48:29.761562592Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:48:29.761826 env[1854]: time="2025-09-13T00:48:29.761581260Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:48:29.763371 env[1854]: time="2025-09-13T00:48:29.762203916Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/cb7a499ed5522a7253b700a2bcf6a6918813badde9aa26906a4028e1c1673a1c pid=2472 runtime=io.containerd.runc.v2 Sep 13 00:48:29.831114 env[1854]: time="2025-09-13T00:48:29.830989115Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-19-167,Uid:1e30b5c1e43171410f5dad087060f636,Namespace:kube-system,Attempt:0,} returns sandbox id \"7def50844e0e64aa3ba34fce3c5b6ee9e20193733487a0af58f0d072379b7f73\"" Sep 13 00:48:29.842114 env[1854]: time="2025-09-13T00:48:29.842068920Z" level=info msg="CreateContainer within sandbox \"7def50844e0e64aa3ba34fce3c5b6ee9e20193733487a0af58f0d072379b7f73\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 13 00:48:29.866142 env[1854]: time="2025-09-13T00:48:29.866077216Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-19-167,Uid:ed7fb32cfe077e607f31e48642251dfc,Namespace:kube-system,Attempt:0,} returns sandbox id \"fec989cef7de0aefcf0cea7bbc9e4b70c1b79e7dc0968b572dc65abc939cb360\"" Sep 13 00:48:29.870299 env[1854]: time="2025-09-13T00:48:29.870258331Z" level=info msg="CreateContainer within sandbox \"fec989cef7de0aefcf0cea7bbc9e4b70c1b79e7dc0968b572dc65abc939cb360\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 13 00:48:29.875299 env[1854]: time="2025-09-13T00:48:29.875251754Z" level=info msg="CreateContainer within sandbox \"7def50844e0e64aa3ba34fce3c5b6ee9e20193733487a0af58f0d072379b7f73\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"acfb4aa5c02333f60f2c7ec14ccb9b88778d4cc72c66b2d1441afee0d4c067ee\"" Sep 13 00:48:29.876297 env[1854]: time="2025-09-13T00:48:29.876259806Z" level=info msg="StartContainer for \"acfb4aa5c02333f60f2c7ec14ccb9b88778d4cc72c66b2d1441afee0d4c067ee\"" Sep 13 00:48:29.884238 env[1854]: time="2025-09-13T00:48:29.884137162Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-19-167,Uid:d7c4cbbdb680656b071548db8d01d06d,Namespace:kube-system,Attempt:0,} returns sandbox id \"cb7a499ed5522a7253b700a2bcf6a6918813badde9aa26906a4028e1c1673a1c\"" Sep 13 00:48:29.887877 env[1854]: time="2025-09-13T00:48:29.887833231Z" level=info msg="CreateContainer within sandbox \"cb7a499ed5522a7253b700a2bcf6a6918813badde9aa26906a4028e1c1673a1c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 13 00:48:29.898326 env[1854]: time="2025-09-13T00:48:29.898263415Z" level=info msg="CreateContainer within sandbox \"fec989cef7de0aefcf0cea7bbc9e4b70c1b79e7dc0968b572dc65abc939cb360\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"74b189b03804fcdfee3a762ed74baf732bd631cc28d7367d6dbb4bd21504f283\"" Sep 13 00:48:29.899772 env[1854]: time="2025-09-13T00:48:29.899724533Z" level=info msg="StartContainer for \"74b189b03804fcdfee3a762ed74baf732bd631cc28d7367d6dbb4bd21504f283\"" Sep 13 00:48:29.925841 env[1854]: time="2025-09-13T00:48:29.925785367Z" level=info msg="CreateContainer within sandbox \"cb7a499ed5522a7253b700a2bcf6a6918813badde9aa26906a4028e1c1673a1c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"dae12014b4cd4a8cdb9364f499f957163f235bd00597fda1eb6478b54f440430\"" Sep 13 00:48:29.927172 env[1854]: time="2025-09-13T00:48:29.927046739Z" level=info msg="StartContainer for \"dae12014b4cd4a8cdb9364f499f957163f235bd00597fda1eb6478b54f440430\"" Sep 13 00:48:30.020703 env[1854]: time="2025-09-13T00:48:30.020610950Z" level=info msg="StartContainer for \"acfb4aa5c02333f60f2c7ec14ccb9b88778d4cc72c66b2d1441afee0d4c067ee\" returns successfully" Sep 13 00:48:30.049270 kubelet[2383]: W0913 00:48:30.046849 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.19.167:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-19-167&limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:30.049270 kubelet[2383]: E0913 00:48:30.046926 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.19.167:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-19-167&limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:30.078844 env[1854]: time="2025-09-13T00:48:30.078784424Z" level=info msg="StartContainer for \"74b189b03804fcdfee3a762ed74baf732bd631cc28d7367d6dbb4bd21504f283\" returns successfully" Sep 13 00:48:30.088432 env[1854]: time="2025-09-13T00:48:30.088318453Z" level=info msg="StartContainer for \"dae12014b4cd4a8cdb9364f499f957163f235bd00597fda1eb6478b54f440430\" returns successfully" Sep 13 00:48:30.095166 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Sep 13 00:48:30.120628 kubelet[2383]: W0913 00:48:30.120453 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.19.167:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:30.120628 kubelet[2383]: E0913 00:48:30.120568 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.19.167:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:30.164619 kubelet[2383]: E0913 00:48:30.164546 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.19.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-19-167?timeout=10s\": dial tcp 172.31.19.167:6443: connect: connection refused" interval="1.6s" Sep 13 00:48:30.202584 kubelet[2383]: W0913 00:48:30.202503 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.19.167:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:30.202834 kubelet[2383]: E0913 00:48:30.202813 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.19.167:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:30.320180 kubelet[2383]: I0913 00:48:30.320145 2383 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-19-167" Sep 13 00:48:30.320762 kubelet[2383]: E0913 00:48:30.320736 2383 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.19.167:6443/api/v1/nodes\": dial tcp 172.31.19.167:6443: connect: connection refused" node="ip-172-31-19-167" Sep 13 00:48:30.827168 kubelet[2383]: E0913 00:48:30.827116 2383 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.19.167:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:31.478881 kubelet[2383]: E0913 00:48:31.478768 2383 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.19.167:6443/api/v1/namespaces/default/events\": dial tcp 172.31.19.167:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-19-167.1864b120bb5ceeb9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-19-167,UID:ip-172-31-19-167,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-19-167,},FirstTimestamp:2025-09-13 00:48:28.721835705 +0000 UTC m=+0.779177588,LastTimestamp:2025-09-13 00:48:28.721835705 +0000 UTC m=+0.779177588,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-19-167,}" Sep 13 00:48:31.768690 kubelet[2383]: E0913 00:48:31.768566 2383 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.19.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-19-167?timeout=10s\": dial tcp 172.31.19.167:6443: connect: connection refused" interval="3.2s" Sep 13 00:48:31.800899 kubelet[2383]: W0913 00:48:31.800860 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.19.167:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:31.801067 kubelet[2383]: E0913 00:48:31.800909 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.19.167:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:31.923518 kubelet[2383]: W0913 00:48:31.923482 2383 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.19.167:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.19.167:6443: connect: connection refused Sep 13 00:48:31.923798 kubelet[2383]: E0913 00:48:31.923776 2383 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.19.167:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.19.167:6443: connect: connection refused" logger="UnhandledError" Sep 13 00:48:31.924073 kubelet[2383]: I0913 00:48:31.924060 2383 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-19-167" Sep 13 00:48:31.924431 kubelet[2383]: E0913 00:48:31.924408 2383 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.19.167:6443/api/v1/nodes\": dial tcp 172.31.19.167:6443: connect: connection refused" node="ip-172-31-19-167" Sep 13 00:48:34.703537 kubelet[2383]: I0913 00:48:34.703492 2383 apiserver.go:52] "Watching apiserver" Sep 13 00:48:34.753738 kubelet[2383]: I0913 00:48:34.753704 2383 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 13 00:48:34.881498 kubelet[2383]: E0913 00:48:34.881461 2383 csi_plugin.go:305] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "ip-172-31-19-167" not found Sep 13 00:48:34.973529 kubelet[2383]: E0913 00:48:34.973502 2383 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-19-167\" not found" node="ip-172-31-19-167" Sep 13 00:48:35.126983 kubelet[2383]: I0913 00:48:35.126946 2383 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-19-167" Sep 13 00:48:35.140507 kubelet[2383]: I0913 00:48:35.140468 2383 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-19-167" Sep 13 00:48:36.912600 systemd[1]: Reloading. Sep 13 00:48:37.019199 /usr/lib/systemd/system-generators/torcx-generator[2670]: time="2025-09-13T00:48:37Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 13 00:48:37.019228 /usr/lib/systemd/system-generators/torcx-generator[2670]: time="2025-09-13T00:48:37Z" level=info msg="torcx already run" Sep 13 00:48:37.166884 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 13 00:48:37.166909 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 13 00:48:37.191310 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 13 00:48:37.299160 systemd[1]: Stopping kubelet.service... Sep 13 00:48:37.322877 systemd[1]: kubelet.service: Deactivated successfully. Sep 13 00:48:37.323343 systemd[1]: Stopped kubelet.service. Sep 13 00:48:37.326253 systemd[1]: Starting kubelet.service... Sep 13 00:48:39.037923 systemd[1]: Started kubelet.service. Sep 13 00:48:39.165469 kubelet[2740]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:48:39.166480 kubelet[2740]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 13 00:48:39.166619 kubelet[2740]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 13 00:48:39.166961 kubelet[2740]: I0913 00:48:39.166911 2740 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 13 00:48:39.177391 kubelet[2740]: I0913 00:48:39.177361 2740 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 13 00:48:39.177563 kubelet[2740]: I0913 00:48:39.177552 2740 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 13 00:48:39.177931 kubelet[2740]: I0913 00:48:39.177918 2740 server.go:934] "Client rotation is on, will bootstrap in background" Sep 13 00:48:39.180119 kubelet[2740]: I0913 00:48:39.180098 2740 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 13 00:48:39.189609 kubelet[2740]: I0913 00:48:39.189574 2740 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 13 00:48:39.190649 sudo[2754]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 13 00:48:39.191042 sudo[2754]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Sep 13 00:48:39.196182 kubelet[2740]: E0913 00:48:39.196132 2740 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 13 00:48:39.196571 kubelet[2740]: I0913 00:48:39.196554 2740 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 13 00:48:39.199901 kubelet[2740]: I0913 00:48:39.199878 2740 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 13 00:48:39.200580 kubelet[2740]: I0913 00:48:39.200560 2740 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 13 00:48:39.200879 kubelet[2740]: I0913 00:48:39.200819 2740 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 13 00:48:39.201205 kubelet[2740]: I0913 00:48:39.200953 2740 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-19-167","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Sep 13 00:48:39.201396 kubelet[2740]: I0913 00:48:39.201384 2740 topology_manager.go:138] "Creating topology manager with none policy" Sep 13 00:48:39.201460 kubelet[2740]: I0913 00:48:39.201451 2740 container_manager_linux.go:300] "Creating device plugin manager" Sep 13 00:48:39.201543 kubelet[2740]: I0913 00:48:39.201534 2740 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:48:39.201704 kubelet[2740]: I0913 00:48:39.201694 2740 kubelet.go:408] "Attempting to sync node with API server" Sep 13 00:48:39.202228 kubelet[2740]: I0913 00:48:39.202201 2740 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 13 00:48:39.202314 kubelet[2740]: I0913 00:48:39.202294 2740 kubelet.go:314] "Adding apiserver pod source" Sep 13 00:48:39.202314 kubelet[2740]: I0913 00:48:39.202310 2740 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 13 00:48:39.222892 kubelet[2740]: I0913 00:48:39.222856 2740 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 13 00:48:39.223523 kubelet[2740]: I0913 00:48:39.223501 2740 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 13 00:48:39.224381 kubelet[2740]: I0913 00:48:39.224360 2740 server.go:1274] "Started kubelet" Sep 13 00:48:39.233374 kubelet[2740]: I0913 00:48:39.230725 2740 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 13 00:48:39.241754 kubelet[2740]: I0913 00:48:39.241686 2740 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 13 00:48:39.243823 kubelet[2740]: I0913 00:48:39.243222 2740 server.go:449] "Adding debug handlers to kubelet server" Sep 13 00:48:39.245130 kubelet[2740]: I0913 00:48:39.244686 2740 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 13 00:48:39.245130 kubelet[2740]: I0913 00:48:39.244962 2740 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 13 00:48:39.245359 kubelet[2740]: I0913 00:48:39.245339 2740 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 13 00:48:39.255573 kubelet[2740]: I0913 00:48:39.248754 2740 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 13 00:48:39.257160 kubelet[2740]: I0913 00:48:39.257123 2740 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 13 00:48:39.257500 kubelet[2740]: I0913 00:48:39.257484 2740 reconciler.go:26] "Reconciler: start to sync state" Sep 13 00:48:39.264265 kubelet[2740]: I0913 00:48:39.264219 2740 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 13 00:48:39.265909 kubelet[2740]: I0913 00:48:39.265797 2740 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 13 00:48:39.265909 kubelet[2740]: I0913 00:48:39.265836 2740 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 13 00:48:39.265909 kubelet[2740]: I0913 00:48:39.265859 2740 kubelet.go:2321] "Starting kubelet main sync loop" Sep 13 00:48:39.265909 kubelet[2740]: E0913 00:48:39.265907 2740 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 13 00:48:39.272538 kubelet[2740]: I0913 00:48:39.272514 2740 factory.go:221] Registration of the containerd container factory successfully Sep 13 00:48:39.272753 kubelet[2740]: I0913 00:48:39.272739 2740 factory.go:221] Registration of the systemd container factory successfully Sep 13 00:48:39.277940 kubelet[2740]: I0913 00:48:39.277912 2740 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 13 00:48:39.295384 kubelet[2740]: E0913 00:48:39.293404 2740 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 13 00:48:39.366230 kubelet[2740]: E0913 00:48:39.366188 2740 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 13 00:48:39.377394 kubelet[2740]: I0913 00:48:39.377364 2740 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 13 00:48:39.377586 kubelet[2740]: I0913 00:48:39.377571 2740 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 13 00:48:39.377709 kubelet[2740]: I0913 00:48:39.377699 2740 state_mem.go:36] "Initialized new in-memory state store" Sep 13 00:48:39.378013 kubelet[2740]: I0913 00:48:39.377998 2740 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 13 00:48:39.378135 kubelet[2740]: I0913 00:48:39.378110 2740 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 13 00:48:39.378253 kubelet[2740]: I0913 00:48:39.378244 2740 policy_none.go:49] "None policy: Start" Sep 13 00:48:39.379365 kubelet[2740]: I0913 00:48:39.379347 2740 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 13 00:48:39.379504 kubelet[2740]: I0913 00:48:39.379494 2740 state_mem.go:35] "Initializing new in-memory state store" Sep 13 00:48:39.379896 kubelet[2740]: I0913 00:48:39.379882 2740 state_mem.go:75] "Updated machine memory state" Sep 13 00:48:39.382728 kubelet[2740]: I0913 00:48:39.382707 2740 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 13 00:48:39.383709 kubelet[2740]: I0913 00:48:39.383669 2740 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 13 00:48:39.383913 kubelet[2740]: I0913 00:48:39.383868 2740 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 13 00:48:39.388240 kubelet[2740]: I0913 00:48:39.388142 2740 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 13 00:48:39.497801 kubelet[2740]: I0913 00:48:39.497756 2740 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-19-167" Sep 13 00:48:39.526438 kubelet[2740]: I0913 00:48:39.526407 2740 kubelet_node_status.go:111] "Node was previously registered" node="ip-172-31-19-167" Sep 13 00:48:39.526706 kubelet[2740]: I0913 00:48:39.526693 2740 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-19-167" Sep 13 00:48:39.661168 kubelet[2740]: I0913 00:48:39.661058 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1e30b5c1e43171410f5dad087060f636-k8s-certs\") pod \"kube-apiserver-ip-172-31-19-167\" (UID: \"1e30b5c1e43171410f5dad087060f636\") " pod="kube-system/kube-apiserver-ip-172-31-19-167" Sep 13 00:48:39.661343 kubelet[2740]: I0913 00:48:39.661326 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-ca-certs\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:39.661473 kubelet[2740]: I0913 00:48:39.661461 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-kubeconfig\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:39.661562 kubelet[2740]: I0913 00:48:39.661530 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:39.661656 kubelet[2740]: I0913 00:48:39.661627 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d7c4cbbdb680656b071548db8d01d06d-kubeconfig\") pod \"kube-scheduler-ip-172-31-19-167\" (UID: \"d7c4cbbdb680656b071548db8d01d06d\") " pod="kube-system/kube-scheduler-ip-172-31-19-167" Sep 13 00:48:39.661656 kubelet[2740]: I0913 00:48:39.661653 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1e30b5c1e43171410f5dad087060f636-ca-certs\") pod \"kube-apiserver-ip-172-31-19-167\" (UID: \"1e30b5c1e43171410f5dad087060f636\") " pod="kube-system/kube-apiserver-ip-172-31-19-167" Sep 13 00:48:39.661784 kubelet[2740]: I0913 00:48:39.661680 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1e30b5c1e43171410f5dad087060f636-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-19-167\" (UID: \"1e30b5c1e43171410f5dad087060f636\") " pod="kube-system/kube-apiserver-ip-172-31-19-167" Sep 13 00:48:39.661784 kubelet[2740]: I0913 00:48:39.661703 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:39.661784 kubelet[2740]: I0913 00:48:39.661723 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ed7fb32cfe077e607f31e48642251dfc-k8s-certs\") pod \"kube-controller-manager-ip-172-31-19-167\" (UID: \"ed7fb32cfe077e607f31e48642251dfc\") " pod="kube-system/kube-controller-manager-ip-172-31-19-167" Sep 13 00:48:40.205000 kubelet[2740]: I0913 00:48:40.204956 2740 apiserver.go:52] "Watching apiserver" Sep 13 00:48:40.257914 kubelet[2740]: I0913 00:48:40.257875 2740 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 13 00:48:40.343251 kubelet[2740]: E0913 00:48:40.343213 2740 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ip-172-31-19-167\" already exists" pod="kube-system/kube-apiserver-ip-172-31-19-167" Sep 13 00:48:40.353362 kubelet[2740]: I0913 00:48:40.353285 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-19-167" podStartSLOduration=1.35326917 podStartE2EDuration="1.35326917s" podCreationTimestamp="2025-09-13 00:48:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:48:40.35145546 +0000 UTC m=+1.279542697" watchObservedRunningTime="2025-09-13 00:48:40.35326917 +0000 UTC m=+1.281356402" Sep 13 00:48:40.379375 kubelet[2740]: I0913 00:48:40.379309 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-19-167" podStartSLOduration=1.379291719 podStartE2EDuration="1.379291719s" podCreationTimestamp="2025-09-13 00:48:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:48:40.363224068 +0000 UTC m=+1.291311309" watchObservedRunningTime="2025-09-13 00:48:40.379291719 +0000 UTC m=+1.307378957" Sep 13 00:48:40.398330 kubelet[2740]: I0913 00:48:40.398266 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-19-167" podStartSLOduration=1.398220229 podStartE2EDuration="1.398220229s" podCreationTimestamp="2025-09-13 00:48:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:48:40.379636572 +0000 UTC m=+1.307723814" watchObservedRunningTime="2025-09-13 00:48:40.398220229 +0000 UTC m=+1.326307466" Sep 13 00:48:40.471868 sudo[2754]: pam_unix(sudo:session): session closed for user root Sep 13 00:48:41.609855 kubelet[2740]: I0913 00:48:41.609809 2740 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 13 00:48:41.610261 kubelet[2740]: I0913 00:48:41.610245 2740 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 13 00:48:41.610298 env[1854]: time="2025-09-13T00:48:41.610073697Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 13 00:48:42.580003 kubelet[2740]: I0913 00:48:42.579960 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-xtables-lock\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580225 kubelet[2740]: I0913 00:48:42.580013 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/27addee7-1eb3-4955-beb7-52d6bb714af1-kube-proxy\") pod \"kube-proxy-529ls\" (UID: \"27addee7-1eb3-4955-beb7-52d6bb714af1\") " pod="kube-system/kube-proxy-529ls" Sep 13 00:48:42.580225 kubelet[2740]: I0913 00:48:42.580042 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2cv2\" (UniqueName: \"kubernetes.io/projected/27addee7-1eb3-4955-beb7-52d6bb714af1-kube-api-access-f2cv2\") pod \"kube-proxy-529ls\" (UID: \"27addee7-1eb3-4955-beb7-52d6bb714af1\") " pod="kube-system/kube-proxy-529ls" Sep 13 00:48:42.580225 kubelet[2740]: I0913 00:48:42.580068 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-clustermesh-secrets\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580225 kubelet[2740]: I0913 00:48:42.580093 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/27addee7-1eb3-4955-beb7-52d6bb714af1-lib-modules\") pod \"kube-proxy-529ls\" (UID: \"27addee7-1eb3-4955-beb7-52d6bb714af1\") " pod="kube-system/kube-proxy-529ls" Sep 13 00:48:42.580225 kubelet[2740]: I0913 00:48:42.580115 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-bpf-maps\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580225 kubelet[2740]: I0913 00:48:42.580135 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-etc-cni-netd\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580507 kubelet[2740]: I0913 00:48:42.580169 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-lib-modules\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580507 kubelet[2740]: I0913 00:48:42.580193 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-net\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580507 kubelet[2740]: I0913 00:48:42.580225 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5vb5\" (UniqueName: \"kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-kube-api-access-n5vb5\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580507 kubelet[2740]: I0913 00:48:42.580254 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-kernel\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580507 kubelet[2740]: I0913 00:48:42.580290 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-run\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580507 kubelet[2740]: I0913 00:48:42.580315 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hostproc\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580761 kubelet[2740]: I0913 00:48:42.580349 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-config-path\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580761 kubelet[2740]: I0913 00:48:42.580373 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hubble-tls\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580761 kubelet[2740]: I0913 00:48:42.580411 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/27addee7-1eb3-4955-beb7-52d6bb714af1-xtables-lock\") pod \"kube-proxy-529ls\" (UID: \"27addee7-1eb3-4955-beb7-52d6bb714af1\") " pod="kube-system/kube-proxy-529ls" Sep 13 00:48:42.580761 kubelet[2740]: I0913 00:48:42.580438 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-cgroup\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.580761 kubelet[2740]: I0913 00:48:42.580474 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cni-path\") pod \"cilium-tr6vw\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " pod="kube-system/cilium-tr6vw" Sep 13 00:48:42.655256 kubelet[2740]: E0913 00:48:42.655205 2740 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[bpf-maps cilium-cgroup cilium-config-path cilium-run clustermesh-secrets cni-path etc-cni-netd host-proc-sys-kernel host-proc-sys-net hostproc hubble-tls kube-api-access-n5vb5 lib-modules xtables-lock], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-tr6vw" podUID="1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" Sep 13 00:48:42.687180 kubelet[2740]: I0913 00:48:42.687127 2740 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 13 00:48:42.782695 kubelet[2740]: I0913 00:48:42.782278 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q7zx\" (UniqueName: \"kubernetes.io/projected/ee9a8885-8284-423a-aae7-0f835d59bd7d-kube-api-access-4q7zx\") pod \"cilium-operator-5d85765b45-76z72\" (UID: \"ee9a8885-8284-423a-aae7-0f835d59bd7d\") " pod="kube-system/cilium-operator-5d85765b45-76z72" Sep 13 00:48:42.794763 kubelet[2740]: I0913 00:48:42.794717 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ee9a8885-8284-423a-aae7-0f835d59bd7d-cilium-config-path\") pod \"cilium-operator-5d85765b45-76z72\" (UID: \"ee9a8885-8284-423a-aae7-0f835d59bd7d\") " pod="kube-system/cilium-operator-5d85765b45-76z72" Sep 13 00:48:42.830649 env[1854]: time="2025-09-13T00:48:42.830529495Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-529ls,Uid:27addee7-1eb3-4955-beb7-52d6bb714af1,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:42.875081 env[1854]: time="2025-09-13T00:48:42.874852243Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:48:42.875081 env[1854]: time="2025-09-13T00:48:42.874904787Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:48:42.875081 env[1854]: time="2025-09-13T00:48:42.874919864Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:48:42.875373 env[1854]: time="2025-09-13T00:48:42.875181998Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/b45fd8009f6b4ba4ea46af2ecce0c6d03c4259466b46b3c0cd190ddba430d5e2 pid=2804 runtime=io.containerd.runc.v2 Sep 13 00:48:42.932648 env[1854]: time="2025-09-13T00:48:42.932597028Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-529ls,Uid:27addee7-1eb3-4955-beb7-52d6bb714af1,Namespace:kube-system,Attempt:0,} returns sandbox id \"b45fd8009f6b4ba4ea46af2ecce0c6d03c4259466b46b3c0cd190ddba430d5e2\"" Sep 13 00:48:42.937263 env[1854]: time="2025-09-13T00:48:42.937227589Z" level=info msg="CreateContainer within sandbox \"b45fd8009f6b4ba4ea46af2ecce0c6d03c4259466b46b3c0cd190ddba430d5e2\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 13 00:48:42.977417 env[1854]: time="2025-09-13T00:48:42.977361972Z" level=info msg="CreateContainer within sandbox \"b45fd8009f6b4ba4ea46af2ecce0c6d03c4259466b46b3c0cd190ddba430d5e2\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"b8d8578876284267f4e605e2264e38eaae4bc767b38397d5cfafb62b2eed6bf7\"" Sep 13 00:48:42.979526 env[1854]: time="2025-09-13T00:48:42.979467906Z" level=info msg="StartContainer for \"b8d8578876284267f4e605e2264e38eaae4bc767b38397d5cfafb62b2eed6bf7\"" Sep 13 00:48:43.042826 env[1854]: time="2025-09-13T00:48:43.042762221Z" level=info msg="StartContainer for \"b8d8578876284267f4e605e2264e38eaae4bc767b38397d5cfafb62b2eed6bf7\" returns successfully" Sep 13 00:48:43.078043 env[1854]: time="2025-09-13T00:48:43.077999852Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-76z72,Uid:ee9a8885-8284-423a-aae7-0f835d59bd7d,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:43.105281 env[1854]: time="2025-09-13T00:48:43.104924982Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:48:43.105281 env[1854]: time="2025-09-13T00:48:43.104970012Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:48:43.105281 env[1854]: time="2025-09-13T00:48:43.104983450Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:48:43.105904 env[1854]: time="2025-09-13T00:48:43.105238014Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa pid=2880 runtime=io.containerd.runc.v2 Sep 13 00:48:43.169634 env[1854]: time="2025-09-13T00:48:43.169169455Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-76z72,Uid:ee9a8885-8284-423a-aae7-0f835d59bd7d,Namespace:kube-system,Attempt:0,} returns sandbox id \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\"" Sep 13 00:48:43.174672 env[1854]: time="2025-09-13T00:48:43.174628508Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 13 00:48:43.399272 kubelet[2740]: I0913 00:48:43.399135 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-xtables-lock\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399272 kubelet[2740]: I0913 00:48:43.399232 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-net\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399481 kubelet[2740]: I0913 00:48:43.399456 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.399572 kubelet[2740]: I0913 00:48:43.399551 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-bpf-maps\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399768 kubelet[2740]: I0913 00:48:43.399699 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5vb5\" (UniqueName: \"kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-kube-api-access-n5vb5\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399768 kubelet[2740]: I0913 00:48:43.399727 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-kernel\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399854 kubelet[2740]: I0913 00:48:43.399771 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-config-path\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399887 kubelet[2740]: I0913 00:48:43.399876 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-cgroup\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399916 kubelet[2740]: I0913 00:48:43.399896 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-clustermesh-secrets\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399916 kubelet[2740]: I0913 00:48:43.399911 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hostproc\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.399975 kubelet[2740]: I0913 00:48:43.399960 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.400040 kubelet[2740]: I0913 00:48:43.400023 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.400090 kubelet[2740]: I0913 00:48:43.400055 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.400133 kubelet[2740]: I0913 00:48:43.400069 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.400688 kubelet[2740]: I0913 00:48:43.400667 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.400798 kubelet[2740]: I0913 00:48:43.400675 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hostproc" (OuterVolumeSpecName: "hostproc") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.400859 kubelet[2740]: I0913 00:48:43.400030 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-run\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.400938 kubelet[2740]: I0913 00:48:43.400926 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hubble-tls\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.401028 kubelet[2740]: I0913 00:48:43.401011 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-etc-cni-netd\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.401104 kubelet[2740]: I0913 00:48:43.401095 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-lib-modules\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.401186 kubelet[2740]: I0913 00:48:43.401177 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cni-path\") pod \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\" (UID: \"1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e\") " Sep 13 00:48:43.401322 kubelet[2740]: I0913 00:48:43.401313 2740 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-bpf-maps\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.401465 kubelet[2740]: I0913 00:48:43.401454 2740 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-kernel\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.401543 kubelet[2740]: I0913 00:48:43.401536 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-cgroup\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.401615 kubelet[2740]: I0913 00:48:43.401607 2740 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hostproc\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.401673 kubelet[2740]: I0913 00:48:43.401665 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-run\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.401731 kubelet[2740]: I0913 00:48:43.401716 2740 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-xtables-lock\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.401791 kubelet[2740]: I0913 00:48:43.401784 2740 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-host-proc-sys-net\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.402654 kubelet[2740]: I0913 00:48:43.402626 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 00:48:43.402730 kubelet[2740]: I0913 00:48:43.402676 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.402730 kubelet[2740]: I0913 00:48:43.402697 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.402730 kubelet[2740]: I0913 00:48:43.402712 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cni-path" (OuterVolumeSpecName: "cni-path") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:48:43.405777 kubelet[2740]: I0913 00:48:43.405732 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 00:48:43.406452 kubelet[2740]: I0913 00:48:43.406412 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-kube-api-access-n5vb5" (OuterVolumeSpecName: "kube-api-access-n5vb5") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "kube-api-access-n5vb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:48:43.408939 kubelet[2740]: I0913 00:48:43.408907 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" (UID: "1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:48:43.502610 kubelet[2740]: I0913 00:48:43.502572 2740 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5vb5\" (UniqueName: \"kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-kube-api-access-n5vb5\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.502610 kubelet[2740]: I0913 00:48:43.502611 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cilium-config-path\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.502831 kubelet[2740]: I0913 00:48:43.502624 2740 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-clustermesh-secrets\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.502831 kubelet[2740]: I0913 00:48:43.502636 2740 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-hubble-tls\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.502831 kubelet[2740]: I0913 00:48:43.502648 2740 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-cni-path\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.502831 kubelet[2740]: I0913 00:48:43.502664 2740 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-etc-cni-netd\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.502831 kubelet[2740]: I0913 00:48:43.502674 2740 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e-lib-modules\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:48:43.710142 systemd[1]: var-lib-kubelet-pods-1d7aaebb\x2d91fc\x2d44cc\x2d8c44\x2d8fb5c3188e4e-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dn5vb5.mount: Deactivated successfully. Sep 13 00:48:43.711017 systemd[1]: var-lib-kubelet-pods-1d7aaebb\x2d91fc\x2d44cc\x2d8c44\x2d8fb5c3188e4e-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:48:43.711227 systemd[1]: var-lib-kubelet-pods-1d7aaebb\x2d91fc\x2d44cc\x2d8c44\x2d8fb5c3188e4e-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:48:44.396573 kubelet[2740]: I0913 00:48:44.396313 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-529ls" podStartSLOduration=2.396276387 podStartE2EDuration="2.396276387s" podCreationTimestamp="2025-09-13 00:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:48:43.357604493 +0000 UTC m=+4.285691732" watchObservedRunningTime="2025-09-13 00:48:44.396276387 +0000 UTC m=+5.324363621" Sep 13 00:48:44.513963 kubelet[2740]: I0913 00:48:44.513924 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-config-path\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514265 kubelet[2740]: I0913 00:48:44.513966 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-run\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514265 kubelet[2740]: I0913 00:48:44.514013 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-kernel\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514265 kubelet[2740]: I0913 00:48:44.514039 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-bpf-maps\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514265 kubelet[2740]: I0913 00:48:44.514059 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-cgroup\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514265 kubelet[2740]: I0913 00:48:44.514079 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-xtables-lock\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514265 kubelet[2740]: I0913 00:48:44.514101 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3a207e55-7e53-468f-a610-f9fcb63a6c12-clustermesh-secrets\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514604 kubelet[2740]: I0913 00:48:44.514127 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-hubble-tls\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514604 kubelet[2740]: I0913 00:48:44.514165 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-hostproc\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514604 kubelet[2740]: I0913 00:48:44.514191 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnj2f\" (UniqueName: \"kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-kube-api-access-lnj2f\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514604 kubelet[2740]: I0913 00:48:44.514216 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cni-path\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514604 kubelet[2740]: I0913 00:48:44.514240 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-etc-cni-netd\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514604 kubelet[2740]: I0913 00:48:44.514262 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-lib-modules\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.514859 kubelet[2740]: I0913 00:48:44.514288 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-net\") pod \"cilium-tdqpw\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " pod="kube-system/cilium-tdqpw" Sep 13 00:48:44.561025 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount925844718.mount: Deactivated successfully. Sep 13 00:48:44.775016 env[1854]: time="2025-09-13T00:48:44.774972161Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tdqpw,Uid:3a207e55-7e53-468f-a610-f9fcb63a6c12,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:44.803455 env[1854]: time="2025-09-13T00:48:44.801610506Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:48:44.803455 env[1854]: time="2025-09-13T00:48:44.801654783Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:48:44.803455 env[1854]: time="2025-09-13T00:48:44.801670451Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:48:44.803455 env[1854]: time="2025-09-13T00:48:44.801817741Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5 pid=2959 runtime=io.containerd.runc.v2 Sep 13 00:48:44.893360 env[1854]: time="2025-09-13T00:48:44.893253039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-tdqpw,Uid:3a207e55-7e53-468f-a610-f9fcb63a6c12,Namespace:kube-system,Attempt:0,} returns sandbox id \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\"" Sep 13 00:48:45.101972 update_engine[1837]: I0913 00:48:45.101208 1837 update_attempter.cc:509] Updating boot flags... Sep 13 00:48:45.271688 kubelet[2740]: I0913 00:48:45.271649 2740 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e" path="/var/lib/kubelet/pods/1d7aaebb-91fc-44cc-8c44-8fb5c3188e4e/volumes" Sep 13 00:48:45.660711 env[1854]: time="2025-09-13T00:48:45.660652925Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:45.662600 env[1854]: time="2025-09-13T00:48:45.662551117Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:45.664782 env[1854]: time="2025-09-13T00:48:45.664741602Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:45.665459 env[1854]: time="2025-09-13T00:48:45.665395431Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 13 00:48:45.667742 env[1854]: time="2025-09-13T00:48:45.667557547Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 13 00:48:45.671092 env[1854]: time="2025-09-13T00:48:45.670901885Z" level=info msg="CreateContainer within sandbox \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 13 00:48:45.691040 env[1854]: time="2025-09-13T00:48:45.690988146Z" level=info msg="CreateContainer within sandbox \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\"" Sep 13 00:48:45.693039 env[1854]: time="2025-09-13T00:48:45.692954237Z" level=info msg="StartContainer for \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\"" Sep 13 00:48:45.724129 systemd[1]: run-containerd-runc-k8s.io-534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8-runc.NenNBp.mount: Deactivated successfully. Sep 13 00:48:45.791810 env[1854]: time="2025-09-13T00:48:45.791756331Z" level=info msg="StartContainer for \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\" returns successfully" Sep 13 00:48:47.514056 kubelet[2740]: I0913 00:48:47.508097 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-76z72" podStartSLOduration=3.013332774 podStartE2EDuration="5.508072099s" podCreationTimestamp="2025-09-13 00:48:42 +0000 UTC" firstStartedPulling="2025-09-13 00:48:43.172272968 +0000 UTC m=+4.100360184" lastFinishedPulling="2025-09-13 00:48:45.667012279 +0000 UTC m=+6.595099509" observedRunningTime="2025-09-13 00:48:46.515678323 +0000 UTC m=+7.443765563" watchObservedRunningTime="2025-09-13 00:48:47.508072099 +0000 UTC m=+8.436159338" Sep 13 00:48:50.920027 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1970852595.mount: Deactivated successfully. Sep 13 00:48:54.089819 env[1854]: time="2025-09-13T00:48:54.089776894Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:54.097289 env[1854]: time="2025-09-13T00:48:54.094050558Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:54.097485 env[1854]: time="2025-09-13T00:48:54.097454963Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 13 00:48:54.097773 env[1854]: time="2025-09-13T00:48:54.097726165Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 13 00:48:54.100801 env[1854]: time="2025-09-13T00:48:54.100760235Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:48:54.124812 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount963096030.mount: Deactivated successfully. Sep 13 00:48:54.135418 env[1854]: time="2025-09-13T00:48:54.135343761Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\"" Sep 13 00:48:54.137047 env[1854]: time="2025-09-13T00:48:54.136289521Z" level=info msg="StartContainer for \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\"" Sep 13 00:48:54.188015 env[1854]: time="2025-09-13T00:48:54.187972516Z" level=info msg="StartContainer for \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\" returns successfully" Sep 13 00:48:54.371161 env[1854]: time="2025-09-13T00:48:54.370780867Z" level=info msg="shim disconnected" id=a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89 Sep 13 00:48:54.371517 env[1854]: time="2025-09-13T00:48:54.371450403Z" level=warning msg="cleaning up after shim disconnected" id=a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89 namespace=k8s.io Sep 13 00:48:54.371517 env[1854]: time="2025-09-13T00:48:54.371479454Z" level=info msg="cleaning up dead shim" Sep 13 00:48:54.383227 env[1854]: time="2025-09-13T00:48:54.383179841Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:48:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3283 runtime=io.containerd.runc.v2\n" Sep 13 00:48:54.437420 env[1854]: time="2025-09-13T00:48:54.437381240Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 00:48:54.464885 env[1854]: time="2025-09-13T00:48:54.464823208Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\"" Sep 13 00:48:54.466923 env[1854]: time="2025-09-13T00:48:54.465621361Z" level=info msg="StartContainer for \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\"" Sep 13 00:48:54.535825 env[1854]: time="2025-09-13T00:48:54.535484184Z" level=info msg="StartContainer for \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\" returns successfully" Sep 13 00:48:54.543303 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 13 00:48:54.544401 systemd[1]: Stopped systemd-sysctl.service. Sep 13 00:48:54.544589 systemd[1]: Stopping systemd-sysctl.service... Sep 13 00:48:54.547428 systemd[1]: Starting systemd-sysctl.service... Sep 13 00:48:54.569659 systemd[1]: Finished systemd-sysctl.service. Sep 13 00:48:54.591299 env[1854]: time="2025-09-13T00:48:54.591237985Z" level=info msg="shim disconnected" id=42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc Sep 13 00:48:54.591299 env[1854]: time="2025-09-13T00:48:54.591280629Z" level=warning msg="cleaning up after shim disconnected" id=42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc namespace=k8s.io Sep 13 00:48:54.591299 env[1854]: time="2025-09-13T00:48:54.591291357Z" level=info msg="cleaning up dead shim" Sep 13 00:48:54.601043 env[1854]: time="2025-09-13T00:48:54.600986469Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:48:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3347 runtime=io.containerd.runc.v2\n" Sep 13 00:48:55.115440 systemd[1]: run-containerd-runc-k8s.io-a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89-runc.o5hHCY.mount: Deactivated successfully. Sep 13 00:48:55.115731 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89-rootfs.mount: Deactivated successfully. Sep 13 00:48:55.438492 env[1854]: time="2025-09-13T00:48:55.432682701Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 00:48:55.464291 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2292093570.mount: Deactivated successfully. Sep 13 00:48:55.478122 env[1854]: time="2025-09-13T00:48:55.478064258Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\"" Sep 13 00:48:55.479952 env[1854]: time="2025-09-13T00:48:55.478857057Z" level=info msg="StartContainer for \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\"" Sep 13 00:48:55.544456 env[1854]: time="2025-09-13T00:48:55.544409956Z" level=info msg="StartContainer for \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\" returns successfully" Sep 13 00:48:55.592643 env[1854]: time="2025-09-13T00:48:55.592601475Z" level=info msg="shim disconnected" id=71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482 Sep 13 00:48:55.592940 env[1854]: time="2025-09-13T00:48:55.592909694Z" level=warning msg="cleaning up after shim disconnected" id=71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482 namespace=k8s.io Sep 13 00:48:55.592940 env[1854]: time="2025-09-13T00:48:55.592933569Z" level=info msg="cleaning up dead shim" Sep 13 00:48:55.602337 env[1854]: time="2025-09-13T00:48:55.602278333Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:48:55Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3405 runtime=io.containerd.runc.v2\n" Sep 13 00:48:56.113973 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482-rootfs.mount: Deactivated successfully. Sep 13 00:48:56.434351 env[1854]: time="2025-09-13T00:48:56.434237534Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 00:48:56.469742 env[1854]: time="2025-09-13T00:48:56.469681033Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\"" Sep 13 00:48:56.471732 env[1854]: time="2025-09-13T00:48:56.470546119Z" level=info msg="StartContainer for \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\"" Sep 13 00:48:56.547470 env[1854]: time="2025-09-13T00:48:56.547418283Z" level=info msg="StartContainer for \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\" returns successfully" Sep 13 00:48:56.574284 env[1854]: time="2025-09-13T00:48:56.574235621Z" level=info msg="shim disconnected" id=c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a Sep 13 00:48:56.574284 env[1854]: time="2025-09-13T00:48:56.574280825Z" level=warning msg="cleaning up after shim disconnected" id=c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a namespace=k8s.io Sep 13 00:48:56.574284 env[1854]: time="2025-09-13T00:48:56.574290108Z" level=info msg="cleaning up dead shim" Sep 13 00:48:56.583491 env[1854]: time="2025-09-13T00:48:56.583445922Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:48:56Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3462 runtime=io.containerd.runc.v2\n" Sep 13 00:48:56.841681 amazon-ssm-agent[1817]: 2025-09-13 00:48:56 INFO [MessagingDeliveryService] [Association] Schedule manager refreshed with 0 associations, 0 new associations associated Sep 13 00:48:57.114800 systemd[1]: run-containerd-runc-k8s.io-c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a-runc.2DhwBB.mount: Deactivated successfully. Sep 13 00:48:57.115334 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a-rootfs.mount: Deactivated successfully. Sep 13 00:48:57.440533 env[1854]: time="2025-09-13T00:48:57.439870040Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 00:48:57.480397 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2877968142.mount: Deactivated successfully. Sep 13 00:48:57.490166 env[1854]: time="2025-09-13T00:48:57.490101010Z" level=info msg="CreateContainer within sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\"" Sep 13 00:48:57.492127 env[1854]: time="2025-09-13T00:48:57.491107113Z" level=info msg="StartContainer for \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\"" Sep 13 00:48:57.564267 env[1854]: time="2025-09-13T00:48:57.559410309Z" level=info msg="StartContainer for \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\" returns successfully" Sep 13 00:48:57.744185 kubelet[2740]: I0913 00:48:57.744068 2740 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Sep 13 00:48:57.818755 kubelet[2740]: I0913 00:48:57.818661 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d549a605-760f-4f56-94d1-303378b278ed-config-volume\") pod \"coredns-7c65d6cfc9-p5qq8\" (UID: \"d549a605-760f-4f56-94d1-303378b278ed\") " pod="kube-system/coredns-7c65d6cfc9-p5qq8" Sep 13 00:48:57.818755 kubelet[2740]: I0913 00:48:57.818702 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q5zc\" (UniqueName: \"kubernetes.io/projected/d549a605-760f-4f56-94d1-303378b278ed-kube-api-access-6q5zc\") pod \"coredns-7c65d6cfc9-p5qq8\" (UID: \"d549a605-760f-4f56-94d1-303378b278ed\") " pod="kube-system/coredns-7c65d6cfc9-p5qq8" Sep 13 00:48:57.920022 kubelet[2740]: I0913 00:48:57.919975 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxj9l\" (UniqueName: \"kubernetes.io/projected/0b8c8e23-d4bf-4dd0-853a-43820e28d451-kube-api-access-sxj9l\") pod \"coredns-7c65d6cfc9-92l8c\" (UID: \"0b8c8e23-d4bf-4dd0-853a-43820e28d451\") " pod="kube-system/coredns-7c65d6cfc9-92l8c" Sep 13 00:48:57.920222 kubelet[2740]: I0913 00:48:57.920037 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b8c8e23-d4bf-4dd0-853a-43820e28d451-config-volume\") pod \"coredns-7c65d6cfc9-92l8c\" (UID: \"0b8c8e23-d4bf-4dd0-853a-43820e28d451\") " pod="kube-system/coredns-7c65d6cfc9-92l8c" Sep 13 00:48:58.119848 env[1854]: time="2025-09-13T00:48:58.119730572Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-p5qq8,Uid:d549a605-760f-4f56-94d1-303378b278ed,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:58.122372 env[1854]: time="2025-09-13T00:48:58.122325530Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-92l8c,Uid:0b8c8e23-d4bf-4dd0-853a-43820e28d451,Namespace:kube-system,Attempt:0,}" Sep 13 00:48:59.117621 systemd[1]: run-containerd-runc-k8s.io-5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc-runc.GEVKQF.mount: Deactivated successfully. Sep 13 00:49:00.283875 systemd-networkd[1513]: cilium_host: Link UP Sep 13 00:49:00.291737 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Sep 13 00:49:00.292708 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Sep 13 00:49:00.284103 systemd-networkd[1513]: cilium_net: Link UP Sep 13 00:49:00.289299 systemd-networkd[1513]: cilium_net: Gained carrier Sep 13 00:49:00.289578 systemd-networkd[1513]: cilium_host: Gained carrier Sep 13 00:49:00.289719 systemd-networkd[1513]: cilium_net: Gained IPv6LL Sep 13 00:49:00.289894 systemd-networkd[1513]: cilium_host: Gained IPv6LL Sep 13 00:49:00.291342 (udev-worker)[3589]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:49:00.292134 (udev-worker)[3651]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:49:00.495571 (udev-worker)[3671]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:49:00.497266 systemd-networkd[1513]: cilium_vxlan: Link UP Sep 13 00:49:00.497272 systemd-networkd[1513]: cilium_vxlan: Gained carrier Sep 13 00:49:01.209603 systemd[1]: run-containerd-runc-k8s.io-5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc-runc.0n30FG.mount: Deactivated successfully. Sep 13 00:49:01.218256 kernel: NET: Registered PF_ALG protocol family Sep 13 00:49:02.249747 systemd-networkd[1513]: cilium_vxlan: Gained IPv6LL Sep 13 00:49:03.844363 systemd-networkd[1513]: lxc_health: Link UP Sep 13 00:49:03.857236 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 13 00:49:03.855753 systemd-networkd[1513]: lxc_health: Gained carrier Sep 13 00:49:04.300526 systemd-networkd[1513]: lxcc6abd94aa986: Link UP Sep 13 00:49:04.308191 kernel: eth0: renamed from tmpd1845 Sep 13 00:49:04.316297 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcc6abd94aa986: link becomes ready Sep 13 00:49:04.315724 systemd-networkd[1513]: lxcc6abd94aa986: Gained carrier Sep 13 00:49:04.318037 systemd-networkd[1513]: lxc87fd510b2910: Link UP Sep 13 00:49:04.334308 kernel: eth0: renamed from tmpe5f95 Sep 13 00:49:04.351385 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc87fd510b2910: link becomes ready Sep 13 00:49:04.348561 systemd-networkd[1513]: lxc87fd510b2910: Gained carrier Sep 13 00:49:04.352336 (udev-worker)[3670]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:49:04.821948 kubelet[2740]: I0913 00:49:04.821849 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-tdqpw" podStartSLOduration=11.61242242 podStartE2EDuration="20.816569968s" podCreationTimestamp="2025-09-13 00:48:44 +0000 UTC" firstStartedPulling="2025-09-13 00:48:44.894953391 +0000 UTC m=+5.823040613" lastFinishedPulling="2025-09-13 00:48:54.099100938 +0000 UTC m=+15.027188161" observedRunningTime="2025-09-13 00:48:58.466096028 +0000 UTC m=+19.394183336" watchObservedRunningTime="2025-09-13 00:49:04.816569968 +0000 UTC m=+25.744657208" Sep 13 00:49:05.064301 systemd-networkd[1513]: lxc_health: Gained IPv6LL Sep 13 00:49:06.278339 systemd-networkd[1513]: lxcc6abd94aa986: Gained IPv6LL Sep 13 00:49:06.278713 systemd-networkd[1513]: lxc87fd510b2910: Gained IPv6LL Sep 13 00:49:08.028916 systemd[1]: run-containerd-runc-k8s.io-5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc-runc.UVkklE.mount: Deactivated successfully. Sep 13 00:49:08.996550 env[1854]: time="2025-09-13T00:49:08.996461054Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:49:08.997022 env[1854]: time="2025-09-13T00:49:08.996968909Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:49:08.997124 env[1854]: time="2025-09-13T00:49:08.997104729Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:49:09.001352 env[1854]: time="2025-09-13T00:49:08.999892402Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:49:09.001352 env[1854]: time="2025-09-13T00:49:08.999927755Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:49:09.001352 env[1854]: time="2025-09-13T00:49:08.999938053Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:49:09.001352 env[1854]: time="2025-09-13T00:49:09.000071039Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e5f953ce026deafc5acb03eeaf429305148eaf3fe4d62d669adb1641ddde1541 pid=4116 runtime=io.containerd.runc.v2 Sep 13 00:49:09.001860 env[1854]: time="2025-09-13T00:49:09.001781400Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d1845ea08bc0934b8b404c88c5c862c7bd387d2df1dcf0de6e40dff7e9bb368e pid=4109 runtime=io.containerd.runc.v2 Sep 13 00:49:09.176619 env[1854]: time="2025-09-13T00:49:09.176532598Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-92l8c,Uid:0b8c8e23-d4bf-4dd0-853a-43820e28d451,Namespace:kube-system,Attempt:0,} returns sandbox id \"d1845ea08bc0934b8b404c88c5c862c7bd387d2df1dcf0de6e40dff7e9bb368e\"" Sep 13 00:49:09.184099 env[1854]: time="2025-09-13T00:49:09.184043023Z" level=info msg="CreateContainer within sandbox \"d1845ea08bc0934b8b404c88c5c862c7bd387d2df1dcf0de6e40dff7e9bb368e\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 00:49:09.200085 env[1854]: time="2025-09-13T00:49:09.200036508Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-p5qq8,Uid:d549a605-760f-4f56-94d1-303378b278ed,Namespace:kube-system,Attempt:0,} returns sandbox id \"e5f953ce026deafc5acb03eeaf429305148eaf3fe4d62d669adb1641ddde1541\"" Sep 13 00:49:09.227739 env[1854]: time="2025-09-13T00:49:09.227663085Z" level=info msg="CreateContainer within sandbox \"e5f953ce026deafc5acb03eeaf429305148eaf3fe4d62d669adb1641ddde1541\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 13 00:49:09.382938 env[1854]: time="2025-09-13T00:49:09.382438295Z" level=info msg="CreateContainer within sandbox \"d1845ea08bc0934b8b404c88c5c862c7bd387d2df1dcf0de6e40dff7e9bb368e\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c8cca204b9590474f9bb0568feba8e9b24f7546e48578a1f61f0eaf357280c16\"" Sep 13 00:49:09.384710 env[1854]: time="2025-09-13T00:49:09.384069967Z" level=info msg="StartContainer for \"c8cca204b9590474f9bb0568feba8e9b24f7546e48578a1f61f0eaf357280c16\"" Sep 13 00:49:09.385017 env[1854]: time="2025-09-13T00:49:09.384988165Z" level=info msg="CreateContainer within sandbox \"e5f953ce026deafc5acb03eeaf429305148eaf3fe4d62d669adb1641ddde1541\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a6442404136fdf6a5d8fd020e25b613b64f26b374143f279e9fb376c6ecc3bc9\"" Sep 13 00:49:09.385620 env[1854]: time="2025-09-13T00:49:09.385597547Z" level=info msg="StartContainer for \"a6442404136fdf6a5d8fd020e25b613b64f26b374143f279e9fb376c6ecc3bc9\"" Sep 13 00:49:09.542460 env[1854]: time="2025-09-13T00:49:09.542410234Z" level=info msg="StartContainer for \"a6442404136fdf6a5d8fd020e25b613b64f26b374143f279e9fb376c6ecc3bc9\" returns successfully" Sep 13 00:49:09.544395 env[1854]: time="2025-09-13T00:49:09.543670033Z" level=info msg="StartContainer for \"c8cca204b9590474f9bb0568feba8e9b24f7546e48578a1f61f0eaf357280c16\" returns successfully" Sep 13 00:49:10.026678 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3236401454.mount: Deactivated successfully. Sep 13 00:49:10.026820 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2896120424.mount: Deactivated successfully. Sep 13 00:49:10.174464 systemd[1]: run-containerd-runc-k8s.io-5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc-runc.Fgeu30.mount: Deactivated successfully. Sep 13 00:49:10.544049 kubelet[2740]: I0913 00:49:10.543978 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-p5qq8" podStartSLOduration=28.543940438 podStartE2EDuration="28.543940438s" podCreationTimestamp="2025-09-13 00:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:49:10.543061533 +0000 UTC m=+31.471148774" watchObservedRunningTime="2025-09-13 00:49:10.543940438 +0000 UTC m=+31.472027678" Sep 13 00:49:10.545398 kubelet[2740]: I0913 00:49:10.544590 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-92l8c" podStartSLOduration=28.544575882 podStartE2EDuration="28.544575882s" podCreationTimestamp="2025-09-13 00:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:49:10.52385191 +0000 UTC m=+31.451939148" watchObservedRunningTime="2025-09-13 00:49:10.544575882 +0000 UTC m=+31.472663122" Sep 13 00:49:11.832135 sudo[2101]: pam_unix(sudo:session): session closed for user root Sep 13 00:49:11.871847 sshd[2097]: pam_unix(sshd:session): session closed for user core Sep 13 00:49:11.886087 systemd[1]: sshd@4-172.31.19.167:22-147.75.109.163:53340.service: Deactivated successfully. Sep 13 00:49:11.887725 systemd[1]: session-5.scope: Deactivated successfully. Sep 13 00:49:11.888369 systemd-logind[1836]: Session 5 logged out. Waiting for processes to exit. Sep 13 00:49:11.889262 systemd-logind[1836]: Removed session 5. Sep 13 00:49:50.058055 amazon-ssm-agent[1817]: 2025-09-13 00:49:50 INFO [HealthCheck] HealthCheck reporting agent health. Sep 13 00:49:51.779980 systemd[1]: Started sshd@5-172.31.19.167:22-147.75.109.163:42120.service. Sep 13 00:49:51.965258 sshd[4310]: Accepted publickey for core from 147.75.109.163 port 42120 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:49:51.966873 sshd[4310]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:51.974293 systemd[1]: Started session-6.scope. Sep 13 00:49:51.975514 systemd-logind[1836]: New session 6 of user core. Sep 13 00:49:52.533432 sshd[4310]: pam_unix(sshd:session): session closed for user core Sep 13 00:49:52.536333 systemd[1]: sshd@5-172.31.19.167:22-147.75.109.163:42120.service: Deactivated successfully. Sep 13 00:49:52.537616 systemd[1]: session-6.scope: Deactivated successfully. Sep 13 00:49:52.537788 systemd-logind[1836]: Session 6 logged out. Waiting for processes to exit. Sep 13 00:49:52.538849 systemd-logind[1836]: Removed session 6. Sep 13 00:49:57.558665 systemd[1]: Started sshd@6-172.31.19.167:22-147.75.109.163:42132.service. Sep 13 00:49:57.725423 sshd[4324]: Accepted publickey for core from 147.75.109.163 port 42132 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:49:57.726898 sshd[4324]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:49:57.733273 systemd[1]: Started session-7.scope. Sep 13 00:49:57.733824 systemd-logind[1836]: New session 7 of user core. Sep 13 00:49:57.924837 sshd[4324]: pam_unix(sshd:session): session closed for user core Sep 13 00:49:57.928426 systemd[1]: sshd@6-172.31.19.167:22-147.75.109.163:42132.service: Deactivated successfully. Sep 13 00:49:57.929982 systemd[1]: session-7.scope: Deactivated successfully. Sep 13 00:49:57.930661 systemd-logind[1836]: Session 7 logged out. Waiting for processes to exit. Sep 13 00:49:57.931837 systemd-logind[1836]: Removed session 7. Sep 13 00:50:02.959044 systemd[1]: Started sshd@7-172.31.19.167:22-147.75.109.163:54046.service. Sep 13 00:50:03.220646 sshd[4338]: Accepted publickey for core from 147.75.109.163 port 54046 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:03.261555 sshd[4338]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:03.343030 systemd-logind[1836]: New session 8 of user core. Sep 13 00:50:03.343239 systemd[1]: Started session-8.scope. Sep 13 00:50:03.725781 sshd[4338]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:03.734531 systemd[1]: sshd@7-172.31.19.167:22-147.75.109.163:54046.service: Deactivated successfully. Sep 13 00:50:03.736851 systemd-logind[1836]: Session 8 logged out. Waiting for processes to exit. Sep 13 00:50:03.736961 systemd[1]: session-8.scope: Deactivated successfully. Sep 13 00:50:03.739831 systemd-logind[1836]: Removed session 8. Sep 13 00:50:08.743848 systemd[1]: Started sshd@8-172.31.19.167:22-147.75.109.163:54062.service. Sep 13 00:50:08.927195 sshd[4351]: Accepted publickey for core from 147.75.109.163 port 54062 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:08.929750 sshd[4351]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:08.937781 systemd[1]: Started session-9.scope. Sep 13 00:50:08.939274 systemd-logind[1836]: New session 9 of user core. Sep 13 00:50:09.140830 sshd[4351]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:09.144623 systemd[1]: sshd@8-172.31.19.167:22-147.75.109.163:54062.service: Deactivated successfully. Sep 13 00:50:09.145875 systemd[1]: session-9.scope: Deactivated successfully. Sep 13 00:50:09.146100 systemd-logind[1836]: Session 9 logged out. Waiting for processes to exit. Sep 13 00:50:09.148117 systemd-logind[1836]: Removed session 9. Sep 13 00:50:14.165241 systemd[1]: Started sshd@9-172.31.19.167:22-147.75.109.163:40038.service. Sep 13 00:50:14.329858 sshd[4365]: Accepted publickey for core from 147.75.109.163 port 40038 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:14.331745 sshd[4365]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:14.337927 systemd[1]: Started session-10.scope. Sep 13 00:50:14.338430 systemd-logind[1836]: New session 10 of user core. Sep 13 00:50:14.534800 sshd[4365]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:14.538504 systemd[1]: sshd@9-172.31.19.167:22-147.75.109.163:40038.service: Deactivated successfully. Sep 13 00:50:14.539872 systemd[1]: session-10.scope: Deactivated successfully. Sep 13 00:50:14.541957 systemd-logind[1836]: Session 10 logged out. Waiting for processes to exit. Sep 13 00:50:14.543447 systemd-logind[1836]: Removed session 10. Sep 13 00:50:14.558327 systemd[1]: Started sshd@10-172.31.19.167:22-147.75.109.163:40050.service. Sep 13 00:50:14.724240 sshd[4379]: Accepted publickey for core from 147.75.109.163 port 40050 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:14.725971 sshd[4379]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:14.732424 systemd[1]: Started session-11.scope. Sep 13 00:50:14.733270 systemd-logind[1836]: New session 11 of user core. Sep 13 00:50:15.008735 sshd[4379]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:15.014558 systemd-logind[1836]: Session 11 logged out. Waiting for processes to exit. Sep 13 00:50:15.015289 systemd[1]: sshd@10-172.31.19.167:22-147.75.109.163:40050.service: Deactivated successfully. Sep 13 00:50:15.019394 systemd[1]: session-11.scope: Deactivated successfully. Sep 13 00:50:15.021400 systemd-logind[1836]: Removed session 11. Sep 13 00:50:15.034024 systemd[1]: Started sshd@11-172.31.19.167:22-147.75.109.163:40064.service. Sep 13 00:50:15.210490 sshd[4392]: Accepted publickey for core from 147.75.109.163 port 40064 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:15.212580 sshd[4392]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:15.219263 systemd-logind[1836]: New session 12 of user core. Sep 13 00:50:15.219597 systemd[1]: Started session-12.scope. Sep 13 00:50:15.446652 sshd[4392]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:15.450530 systemd[1]: sshd@11-172.31.19.167:22-147.75.109.163:40064.service: Deactivated successfully. Sep 13 00:50:15.452675 systemd[1]: session-12.scope: Deactivated successfully. Sep 13 00:50:15.453246 systemd-logind[1836]: Session 12 logged out. Waiting for processes to exit. Sep 13 00:50:15.454885 systemd-logind[1836]: Removed session 12. Sep 13 00:50:20.470035 systemd[1]: Started sshd@12-172.31.19.167:22-147.75.109.163:33108.service. Sep 13 00:50:20.640717 sshd[4405]: Accepted publickey for core from 147.75.109.163 port 33108 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:20.642318 sshd[4405]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:20.648485 systemd[1]: Started session-13.scope. Sep 13 00:50:20.648963 systemd-logind[1836]: New session 13 of user core. Sep 13 00:50:20.838134 sshd[4405]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:20.841182 systemd[1]: sshd@12-172.31.19.167:22-147.75.109.163:33108.service: Deactivated successfully. Sep 13 00:50:20.842633 systemd[1]: session-13.scope: Deactivated successfully. Sep 13 00:50:20.843507 systemd-logind[1836]: Session 13 logged out. Waiting for processes to exit. Sep 13 00:50:20.844712 systemd-logind[1836]: Removed session 13. Sep 13 00:50:25.862739 systemd[1]: Started sshd@13-172.31.19.167:22-147.75.109.163:33120.service. Sep 13 00:50:26.027428 sshd[4418]: Accepted publickey for core from 147.75.109.163 port 33120 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:26.029006 sshd[4418]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:26.037195 systemd[1]: Started session-14.scope. Sep 13 00:50:26.037978 systemd-logind[1836]: New session 14 of user core. Sep 13 00:50:26.246006 sshd[4418]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:26.249070 systemd[1]: sshd@13-172.31.19.167:22-147.75.109.163:33120.service: Deactivated successfully. Sep 13 00:50:26.250115 systemd[1]: session-14.scope: Deactivated successfully. Sep 13 00:50:26.250480 systemd-logind[1836]: Session 14 logged out. Waiting for processes to exit. Sep 13 00:50:26.251948 systemd-logind[1836]: Removed session 14. Sep 13 00:50:26.271863 systemd[1]: Started sshd@14-172.31.19.167:22-147.75.109.163:33128.service. Sep 13 00:50:26.441352 sshd[4431]: Accepted publickey for core from 147.75.109.163 port 33128 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:26.442825 sshd[4431]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:26.448696 systemd[1]: Started session-15.scope. Sep 13 00:50:26.449394 systemd-logind[1836]: New session 15 of user core. Sep 13 00:50:31.278040 sshd[4431]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:31.281802 systemd-logind[1836]: Session 15 logged out. Waiting for processes to exit. Sep 13 00:50:31.281919 systemd[1]: sshd@14-172.31.19.167:22-147.75.109.163:33128.service: Deactivated successfully. Sep 13 00:50:31.282736 systemd[1]: session-15.scope: Deactivated successfully. Sep 13 00:50:31.283604 systemd-logind[1836]: Removed session 15. Sep 13 00:50:31.300964 systemd[1]: Started sshd@15-172.31.19.167:22-147.75.109.163:50478.service. Sep 13 00:50:31.481919 sshd[4442]: Accepted publickey for core from 147.75.109.163 port 50478 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:31.484044 sshd[4442]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:31.490552 systemd-logind[1836]: New session 16 of user core. Sep 13 00:50:31.490688 systemd[1]: Started session-16.scope. Sep 13 00:50:33.040656 sshd[4442]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:33.044361 systemd-logind[1836]: Session 16 logged out. Waiting for processes to exit. Sep 13 00:50:33.044580 systemd[1]: sshd@15-172.31.19.167:22-147.75.109.163:50478.service: Deactivated successfully. Sep 13 00:50:33.045739 systemd[1]: session-16.scope: Deactivated successfully. Sep 13 00:50:33.047139 systemd-logind[1836]: Removed session 16. Sep 13 00:50:33.065700 systemd[1]: Started sshd@16-172.31.19.167:22-147.75.109.163:50484.service. Sep 13 00:50:33.237488 sshd[4461]: Accepted publickey for core from 147.75.109.163 port 50484 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:33.239316 sshd[4461]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:33.245269 systemd-logind[1836]: New session 17 of user core. Sep 13 00:50:33.245510 systemd[1]: Started session-17.scope. Sep 13 00:50:33.684230 sshd[4461]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:33.688454 systemd[1]: sshd@16-172.31.19.167:22-147.75.109.163:50484.service: Deactivated successfully. Sep 13 00:50:33.689481 systemd[1]: session-17.scope: Deactivated successfully. Sep 13 00:50:33.689810 systemd-logind[1836]: Session 17 logged out. Waiting for processes to exit. Sep 13 00:50:33.690714 systemd-logind[1836]: Removed session 17. Sep 13 00:50:33.708110 systemd[1]: Started sshd@17-172.31.19.167:22-147.75.109.163:50488.service. Sep 13 00:50:33.876566 sshd[4472]: Accepted publickey for core from 147.75.109.163 port 50488 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:33.878371 sshd[4472]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:33.884066 systemd[1]: Started session-18.scope. Sep 13 00:50:33.884842 systemd-logind[1836]: New session 18 of user core. Sep 13 00:50:34.085671 sshd[4472]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:34.088707 systemd[1]: sshd@17-172.31.19.167:22-147.75.109.163:50488.service: Deactivated successfully. Sep 13 00:50:34.089716 systemd[1]: session-18.scope: Deactivated successfully. Sep 13 00:50:34.090519 systemd-logind[1836]: Session 18 logged out. Waiting for processes to exit. Sep 13 00:50:34.091766 systemd-logind[1836]: Removed session 18. Sep 13 00:50:39.110656 systemd[1]: Started sshd@18-172.31.19.167:22-147.75.109.163:50500.service. Sep 13 00:50:39.277310 sshd[4488]: Accepted publickey for core from 147.75.109.163 port 50500 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:39.278714 sshd[4488]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:39.285083 systemd[1]: Started session-19.scope. Sep 13 00:50:39.285924 systemd-logind[1836]: New session 19 of user core. Sep 13 00:50:39.476049 sshd[4488]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:39.479280 systemd[1]: sshd@18-172.31.19.167:22-147.75.109.163:50500.service: Deactivated successfully. Sep 13 00:50:39.480714 systemd[1]: session-19.scope: Deactivated successfully. Sep 13 00:50:39.481245 systemd-logind[1836]: Session 19 logged out. Waiting for processes to exit. Sep 13 00:50:39.482303 systemd-logind[1836]: Removed session 19. Sep 13 00:50:44.501161 systemd[1]: Started sshd@19-172.31.19.167:22-147.75.109.163:35920.service. Sep 13 00:50:44.672290 sshd[4503]: Accepted publickey for core from 147.75.109.163 port 35920 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:44.674110 sshd[4503]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:44.682494 systemd[1]: Started session-20.scope. Sep 13 00:50:44.683728 systemd-logind[1836]: New session 20 of user core. Sep 13 00:50:44.874198 sshd[4503]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:44.878485 systemd[1]: sshd@19-172.31.19.167:22-147.75.109.163:35920.service: Deactivated successfully. Sep 13 00:50:44.879925 systemd-logind[1836]: Session 20 logged out. Waiting for processes to exit. Sep 13 00:50:44.880037 systemd[1]: session-20.scope: Deactivated successfully. Sep 13 00:50:44.881596 systemd-logind[1836]: Removed session 20. Sep 13 00:50:49.899735 systemd[1]: Started sshd@20-172.31.19.167:22-147.75.109.163:35936.service. Sep 13 00:50:50.067191 sshd[4518]: Accepted publickey for core from 147.75.109.163 port 35936 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:50.070078 sshd[4518]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:50.078823 systemd[1]: Started session-21.scope. Sep 13 00:50:50.080241 systemd-logind[1836]: New session 21 of user core. Sep 13 00:50:50.270609 sshd[4518]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:50.274618 systemd[1]: sshd@20-172.31.19.167:22-147.75.109.163:35936.service: Deactivated successfully. Sep 13 00:50:50.275905 systemd-logind[1836]: Session 21 logged out. Waiting for processes to exit. Sep 13 00:50:50.276027 systemd[1]: session-21.scope: Deactivated successfully. Sep 13 00:50:50.277936 systemd-logind[1836]: Removed session 21. Sep 13 00:50:55.293699 systemd[1]: Started sshd@21-172.31.19.167:22-147.75.109.163:35410.service. Sep 13 00:50:55.457527 sshd[4531]: Accepted publickey for core from 147.75.109.163 port 35410 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:55.459610 sshd[4531]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:55.467992 systemd[1]: Started session-22.scope. Sep 13 00:50:55.469311 systemd-logind[1836]: New session 22 of user core. Sep 13 00:50:55.657650 sshd[4531]: pam_unix(sshd:session): session closed for user core Sep 13 00:50:55.660877 systemd[1]: sshd@21-172.31.19.167:22-147.75.109.163:35410.service: Deactivated successfully. Sep 13 00:50:55.662102 systemd[1]: session-22.scope: Deactivated successfully. Sep 13 00:50:55.662525 systemd-logind[1836]: Session 22 logged out. Waiting for processes to exit. Sep 13 00:50:55.663573 systemd-logind[1836]: Removed session 22. Sep 13 00:50:55.680871 systemd[1]: Started sshd@22-172.31.19.167:22-147.75.109.163:35426.service. Sep 13 00:50:55.846626 sshd[4544]: Accepted publickey for core from 147.75.109.163 port 35426 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:50:55.848513 sshd[4544]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:50:55.854528 systemd[1]: Started session-23.scope. Sep 13 00:50:55.855514 systemd-logind[1836]: New session 23 of user core. Sep 13 00:50:58.108889 systemd[1]: run-containerd-runc-k8s.io-5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc-runc.fkhDBM.mount: Deactivated successfully. Sep 13 00:50:58.140845 env[1854]: time="2025-09-13T00:50:58.139070580Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 13 00:50:58.148339 env[1854]: time="2025-09-13T00:50:58.148303526Z" level=info msg="StopContainer for \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\" with timeout 2 (s)" Sep 13 00:50:58.148486 env[1854]: time="2025-09-13T00:50:58.148443676Z" level=info msg="StopContainer for \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\" with timeout 30 (s)" Sep 13 00:50:58.148735 env[1854]: time="2025-09-13T00:50:58.148690252Z" level=info msg="Stop container \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\" with signal terminated" Sep 13 00:50:58.148855 env[1854]: time="2025-09-13T00:50:58.148767873Z" level=info msg="Stop container \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\" with signal terminated" Sep 13 00:50:58.164513 systemd-networkd[1513]: lxc_health: Link DOWN Sep 13 00:50:58.164521 systemd-networkd[1513]: lxc_health: Lost carrier Sep 13 00:50:58.223933 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc-rootfs.mount: Deactivated successfully. Sep 13 00:50:58.253465 env[1854]: time="2025-09-13T00:50:58.253401666Z" level=info msg="shim disconnected" id=5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc Sep 13 00:50:58.253706 env[1854]: time="2025-09-13T00:50:58.253447246Z" level=warning msg="cleaning up after shim disconnected" id=5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc namespace=k8s.io Sep 13 00:50:58.253706 env[1854]: time="2025-09-13T00:50:58.253508798Z" level=info msg="cleaning up dead shim" Sep 13 00:50:58.263071 env[1854]: time="2025-09-13T00:50:58.263020059Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:58Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4599 runtime=io.containerd.runc.v2\n" Sep 13 00:50:58.267056 env[1854]: time="2025-09-13T00:50:58.267000327Z" level=info msg="StopContainer for \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\" returns successfully" Sep 13 00:50:58.267970 env[1854]: time="2025-09-13T00:50:58.267926904Z" level=info msg="StopPodSandbox for \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\"" Sep 13 00:50:58.268192 env[1854]: time="2025-09-13T00:50:58.268007067Z" level=info msg="Container to stop \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:50:58.268192 env[1854]: time="2025-09-13T00:50:58.268049009Z" level=info msg="Container to stop \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:50:58.268192 env[1854]: time="2025-09-13T00:50:58.268069413Z" level=info msg="Container to stop \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:50:58.268192 env[1854]: time="2025-09-13T00:50:58.268096965Z" level=info msg="Container to stop \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:50:58.268192 env[1854]: time="2025-09-13T00:50:58.268115236Z" level=info msg="Container to stop \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:50:58.271832 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5-shm.mount: Deactivated successfully. Sep 13 00:50:58.305171 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5-rootfs.mount: Deactivated successfully. Sep 13 00:50:58.321744 env[1854]: time="2025-09-13T00:50:58.321678080Z" level=info msg="shim disconnected" id=031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5 Sep 13 00:50:58.321744 env[1854]: time="2025-09-13T00:50:58.321728657Z" level=warning msg="cleaning up after shim disconnected" id=031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5 namespace=k8s.io Sep 13 00:50:58.321744 env[1854]: time="2025-09-13T00:50:58.321739066Z" level=info msg="cleaning up dead shim" Sep 13 00:50:58.332782 env[1854]: time="2025-09-13T00:50:58.332731461Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:58Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4632 runtime=io.containerd.runc.v2\n" Sep 13 00:50:58.334299 env[1854]: time="2025-09-13T00:50:58.334249616Z" level=info msg="TearDown network for sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" successfully" Sep 13 00:50:58.334299 env[1854]: time="2025-09-13T00:50:58.334287479Z" level=info msg="StopPodSandbox for \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" returns successfully" Sep 13 00:50:58.465672 kubelet[2740]: I0913 00:50:58.463656 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-xtables-lock\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.465672 kubelet[2740]: I0913 00:50:58.463716 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-etc-cni-netd\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.465672 kubelet[2740]: I0913 00:50:58.463739 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-lib-modules\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.465672 kubelet[2740]: I0913 00:50:58.463772 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3a207e55-7e53-468f-a610-f9fcb63a6c12-clustermesh-secrets\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.465672 kubelet[2740]: I0913 00:50:58.463816 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnj2f\" (UniqueName: \"kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-kube-api-access-lnj2f\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.465672 kubelet[2740]: I0913 00:50:58.463841 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-cgroup\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466500 kubelet[2740]: I0913 00:50:58.463877 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-net\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466500 kubelet[2740]: I0913 00:50:58.463904 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-hostproc\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466500 kubelet[2740]: I0913 00:50:58.463926 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cni-path\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466500 kubelet[2740]: I0913 00:50:58.463968 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-run\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466500 kubelet[2740]: I0913 00:50:58.463992 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-bpf-maps\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466500 kubelet[2740]: I0913 00:50:58.464030 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-kernel\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466877 kubelet[2740]: I0913 00:50:58.464057 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-config-path\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.466877 kubelet[2740]: I0913 00:50:58.464087 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-hubble-tls\") pod \"3a207e55-7e53-468f-a610-f9fcb63a6c12\" (UID: \"3a207e55-7e53-468f-a610-f9fcb63a6c12\") " Sep 13 00:50:58.472034 kubelet[2740]: I0913 00:50:58.467426 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.472205 kubelet[2740]: I0913 00:50:58.472083 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.472338 kubelet[2740]: I0913 00:50:58.467165 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.473595 kubelet[2740]: I0913 00:50:58.473317 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.473595 kubelet[2740]: I0913 00:50:58.473360 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.473595 kubelet[2740]: I0913 00:50:58.473375 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-hostproc" (OuterVolumeSpecName: "hostproc") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.473595 kubelet[2740]: I0913 00:50:58.473389 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cni-path" (OuterVolumeSpecName: "cni-path") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.473595 kubelet[2740]: I0913 00:50:58.473404 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.473867 kubelet[2740]: I0913 00:50:58.473428 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.473867 kubelet[2740]: I0913 00:50:58.473449 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:50:58.489300 kubelet[2740]: I0913 00:50:58.489254 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 00:50:58.496780 kubelet[2740]: I0913 00:50:58.495141 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a207e55-7e53-468f-a610-f9fcb63a6c12-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 00:50:58.500537 kubelet[2740]: I0913 00:50:58.500485 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:50:58.501559 kubelet[2740]: I0913 00:50:58.501117 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-kube-api-access-lnj2f" (OuterVolumeSpecName: "kube-api-access-lnj2f") pod "3a207e55-7e53-468f-a610-f9fcb63a6c12" (UID: "3a207e55-7e53-468f-a610-f9fcb63a6c12"). InnerVolumeSpecName "kube-api-access-lnj2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:50:58.510935 env[1854]: time="2025-09-13T00:50:58.510866066Z" level=info msg="shim disconnected" id=534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8 Sep 13 00:50:58.510935 env[1854]: time="2025-09-13T00:50:58.510925812Z" level=warning msg="cleaning up after shim disconnected" id=534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8 namespace=k8s.io Sep 13 00:50:58.510935 env[1854]: time="2025-09-13T00:50:58.510940026Z" level=info msg="cleaning up dead shim" Sep 13 00:50:58.521827 env[1854]: time="2025-09-13T00:50:58.521780336Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:58Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4661 runtime=io.containerd.runc.v2\n" Sep 13 00:50:58.526455 env[1854]: time="2025-09-13T00:50:58.526408372Z" level=info msg="StopContainer for \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\" returns successfully" Sep 13 00:50:58.527129 env[1854]: time="2025-09-13T00:50:58.527089586Z" level=info msg="StopPodSandbox for \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\"" Sep 13 00:50:58.527286 env[1854]: time="2025-09-13T00:50:58.527165274Z" level=info msg="Container to stop \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:50:58.564967 kubelet[2740]: I0913 00:50:58.564920 2740 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cni-path\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.564967 kubelet[2740]: I0913 00:50:58.564955 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-run\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.564967 kubelet[2740]: I0913 00:50:58.564967 2740 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-bpf-maps\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.564977 2740 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-kernel\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.564988 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-config-path\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.564996 2740 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-hubble-tls\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.565004 2740 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-xtables-lock\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.565012 2740 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-etc-cni-netd\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.565020 2740 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-lib-modules\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.565028 2740 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnj2f\" (UniqueName: \"kubernetes.io/projected/3a207e55-7e53-468f-a610-f9fcb63a6c12-kube-api-access-lnj2f\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565281 kubelet[2740]: I0913 00:50:58.565035 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-cilium-cgroup\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565498 kubelet[2740]: I0913 00:50:58.565044 2740 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3a207e55-7e53-468f-a610-f9fcb63a6c12-clustermesh-secrets\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565498 kubelet[2740]: I0913 00:50:58.565052 2740 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-hostproc\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.565498 kubelet[2740]: I0913 00:50:58.565059 2740 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3a207e55-7e53-468f-a610-f9fcb63a6c12-host-proc-sys-net\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.569071 env[1854]: time="2025-09-13T00:50:58.569028113Z" level=info msg="shim disconnected" id=473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa Sep 13 00:50:58.569348 env[1854]: time="2025-09-13T00:50:58.569329525Z" level=warning msg="cleaning up after shim disconnected" id=473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa namespace=k8s.io Sep 13 00:50:58.569434 env[1854]: time="2025-09-13T00:50:58.569421831Z" level=info msg="cleaning up dead shim" Sep 13 00:50:58.578509 env[1854]: time="2025-09-13T00:50:58.578462574Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:50:58Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4697 runtime=io.containerd.runc.v2\n" Sep 13 00:50:58.579727 env[1854]: time="2025-09-13T00:50:58.579682971Z" level=info msg="TearDown network for sandbox \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" successfully" Sep 13 00:50:58.579727 env[1854]: time="2025-09-13T00:50:58.579721377Z" level=info msg="StopPodSandbox for \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" returns successfully" Sep 13 00:50:58.665961 kubelet[2740]: I0913 00:50:58.665910 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q7zx\" (UniqueName: \"kubernetes.io/projected/ee9a8885-8284-423a-aae7-0f835d59bd7d-kube-api-access-4q7zx\") pod \"ee9a8885-8284-423a-aae7-0f835d59bd7d\" (UID: \"ee9a8885-8284-423a-aae7-0f835d59bd7d\") " Sep 13 00:50:58.665961 kubelet[2740]: I0913 00:50:58.665964 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ee9a8885-8284-423a-aae7-0f835d59bd7d-cilium-config-path\") pod \"ee9a8885-8284-423a-aae7-0f835d59bd7d\" (UID: \"ee9a8885-8284-423a-aae7-0f835d59bd7d\") " Sep 13 00:50:58.671997 kubelet[2740]: I0913 00:50:58.671934 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9a8885-8284-423a-aae7-0f835d59bd7d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "ee9a8885-8284-423a-aae7-0f835d59bd7d" (UID: "ee9a8885-8284-423a-aae7-0f835d59bd7d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 00:50:58.675892 kubelet[2740]: I0913 00:50:58.675090 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9a8885-8284-423a-aae7-0f835d59bd7d-kube-api-access-4q7zx" (OuterVolumeSpecName: "kube-api-access-4q7zx") pod "ee9a8885-8284-423a-aae7-0f835d59bd7d" (UID: "ee9a8885-8284-423a-aae7-0f835d59bd7d"). InnerVolumeSpecName "kube-api-access-4q7zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:50:58.748089 kubelet[2740]: I0913 00:50:58.747718 2740 scope.go:117] "RemoveContainer" containerID="534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8" Sep 13 00:50:58.757112 env[1854]: time="2025-09-13T00:50:58.756898913Z" level=info msg="RemoveContainer for \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\"" Sep 13 00:50:58.764683 env[1854]: time="2025-09-13T00:50:58.764627787Z" level=info msg="RemoveContainer for \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\" returns successfully" Sep 13 00:50:58.766360 kubelet[2740]: I0913 00:50:58.766330 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ee9a8885-8284-423a-aae7-0f835d59bd7d-cilium-config-path\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.766605 kubelet[2740]: I0913 00:50:58.766571 2740 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q7zx\" (UniqueName: \"kubernetes.io/projected/ee9a8885-8284-423a-aae7-0f835d59bd7d-kube-api-access-4q7zx\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:50:58.771655 kubelet[2740]: I0913 00:50:58.771623 2740 scope.go:117] "RemoveContainer" containerID="534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8" Sep 13 00:50:58.774781 env[1854]: time="2025-09-13T00:50:58.774565718Z" level=error msg="ContainerStatus for \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\": not found" Sep 13 00:50:58.775116 kubelet[2740]: E0913 00:50:58.775085 2740 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\": not found" containerID="534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8" Sep 13 00:50:58.777221 kubelet[2740]: I0913 00:50:58.775129 2740 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8"} err="failed to get container status \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\": rpc error: code = NotFound desc = an error occurred when try to find container \"534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8\": not found" Sep 13 00:50:58.777368 kubelet[2740]: I0913 00:50:58.777226 2740 scope.go:117] "RemoveContainer" containerID="5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc" Sep 13 00:50:58.780630 env[1854]: time="2025-09-13T00:50:58.780094742Z" level=info msg="RemoveContainer for \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\"" Sep 13 00:50:58.785642 env[1854]: time="2025-09-13T00:50:58.785577860Z" level=info msg="RemoveContainer for \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\" returns successfully" Sep 13 00:50:58.785894 kubelet[2740]: I0913 00:50:58.785839 2740 scope.go:117] "RemoveContainer" containerID="c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a" Sep 13 00:50:58.790212 env[1854]: time="2025-09-13T00:50:58.789870859Z" level=info msg="RemoveContainer for \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\"" Sep 13 00:50:58.797405 env[1854]: time="2025-09-13T00:50:58.797346631Z" level=info msg="RemoveContainer for \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\" returns successfully" Sep 13 00:50:58.797671 kubelet[2740]: I0913 00:50:58.797636 2740 scope.go:117] "RemoveContainer" containerID="71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482" Sep 13 00:50:58.799661 env[1854]: time="2025-09-13T00:50:58.799365765Z" level=info msg="RemoveContainer for \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\"" Sep 13 00:50:58.805494 env[1854]: time="2025-09-13T00:50:58.805433078Z" level=info msg="RemoveContainer for \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\" returns successfully" Sep 13 00:50:58.805812 kubelet[2740]: I0913 00:50:58.805780 2740 scope.go:117] "RemoveContainer" containerID="42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc" Sep 13 00:50:58.807267 env[1854]: time="2025-09-13T00:50:58.807231565Z" level=info msg="RemoveContainer for \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\"" Sep 13 00:50:58.815055 env[1854]: time="2025-09-13T00:50:58.814957304Z" level=info msg="RemoveContainer for \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\" returns successfully" Sep 13 00:50:58.816842 kubelet[2740]: I0913 00:50:58.816802 2740 scope.go:117] "RemoveContainer" containerID="a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89" Sep 13 00:50:58.823906 env[1854]: time="2025-09-13T00:50:58.823865125Z" level=info msg="RemoveContainer for \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\"" Sep 13 00:50:58.830588 env[1854]: time="2025-09-13T00:50:58.830513384Z" level=info msg="RemoveContainer for \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\" returns successfully" Sep 13 00:50:58.831123 kubelet[2740]: I0913 00:50:58.831093 2740 scope.go:117] "RemoveContainer" containerID="5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc" Sep 13 00:50:58.831798 env[1854]: time="2025-09-13T00:50:58.831610965Z" level=error msg="ContainerStatus for \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\": not found" Sep 13 00:50:58.831890 kubelet[2740]: E0913 00:50:58.831797 2740 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\": not found" containerID="5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc" Sep 13 00:50:58.831890 kubelet[2740]: I0913 00:50:58.831824 2740 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc"} err="failed to get container status \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\": rpc error: code = NotFound desc = an error occurred when try to find container \"5a7a10af768f4ddf81ad5494cf1d39a354c94c837e6c8ff13756e266ede0e2fc\": not found" Sep 13 00:50:58.831890 kubelet[2740]: I0913 00:50:58.831846 2740 scope.go:117] "RemoveContainer" containerID="c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a" Sep 13 00:50:58.832071 env[1854]: time="2025-09-13T00:50:58.832022079Z" level=error msg="ContainerStatus for \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\": not found" Sep 13 00:50:58.832236 kubelet[2740]: E0913 00:50:58.832143 2740 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\": not found" containerID="c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a" Sep 13 00:50:58.832295 kubelet[2740]: I0913 00:50:58.832235 2740 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a"} err="failed to get container status \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\": rpc error: code = NotFound desc = an error occurred when try to find container \"c734ede589755b12461eb94f595e9610c937664614d0e8c0ec34ff9842fd7f4a\": not found" Sep 13 00:50:58.832295 kubelet[2740]: I0913 00:50:58.832251 2740 scope.go:117] "RemoveContainer" containerID="71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482" Sep 13 00:50:58.832641 env[1854]: time="2025-09-13T00:50:58.832542829Z" level=error msg="ContainerStatus for \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\": not found" Sep 13 00:50:58.832734 kubelet[2740]: E0913 00:50:58.832690 2740 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\": not found" containerID="71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482" Sep 13 00:50:58.832775 kubelet[2740]: I0913 00:50:58.832739 2740 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482"} err="failed to get container status \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\": rpc error: code = NotFound desc = an error occurred when try to find container \"71058cfd6692b4b1dcf36dcc2c0f6b7530f33858a4d349d18353d210a9d4d482\": not found" Sep 13 00:50:58.832775 kubelet[2740]: I0913 00:50:58.832755 2740 scope.go:117] "RemoveContainer" containerID="42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc" Sep 13 00:50:58.832974 env[1854]: time="2025-09-13T00:50:58.832931309Z" level=error msg="ContainerStatus for \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\": not found" Sep 13 00:50:58.833060 kubelet[2740]: E0913 00:50:58.833037 2740 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\": not found" containerID="42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc" Sep 13 00:50:58.833169 kubelet[2740]: I0913 00:50:58.833062 2740 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc"} err="failed to get container status \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\": rpc error: code = NotFound desc = an error occurred when try to find container \"42c411579d1b68624a58ed6f0e7a603f16010285a207ee9d0e61ff188505d3bc\": not found" Sep 13 00:50:58.833169 kubelet[2740]: I0913 00:50:58.833077 2740 scope.go:117] "RemoveContainer" containerID="a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89" Sep 13 00:50:58.833274 env[1854]: time="2025-09-13T00:50:58.833218483Z" level=error msg="ContainerStatus for \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\": not found" Sep 13 00:50:58.833345 kubelet[2740]: E0913 00:50:58.833325 2740 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\": not found" containerID="a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89" Sep 13 00:50:58.833384 kubelet[2740]: I0913 00:50:58.833346 2740 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89"} err="failed to get container status \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\": rpc error: code = NotFound desc = an error occurred when try to find container \"a39204fe33d48bd57c6fda524c8971667c1be4f6147df988693791b1a0213a89\": not found" Sep 13 00:50:59.098646 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-534aa5ee9dd4f963ef65499fd3c68309c29d8fa1c335c99343ab515025eed4e8-rootfs.mount: Deactivated successfully. Sep 13 00:50:59.098865 systemd[1]: var-lib-kubelet-pods-3a207e55\x2d7e53\x2d468f\x2da610\x2df9fcb63a6c12-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dlnj2f.mount: Deactivated successfully. Sep 13 00:50:59.099042 systemd[1]: var-lib-kubelet-pods-3a207e55\x2d7e53\x2d468f\x2da610\x2df9fcb63a6c12-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:50:59.099200 systemd[1]: var-lib-kubelet-pods-3a207e55\x2d7e53\x2d468f\x2da610\x2df9fcb63a6c12-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:50:59.099346 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa-rootfs.mount: Deactivated successfully. Sep 13 00:50:59.099495 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa-shm.mount: Deactivated successfully. Sep 13 00:50:59.099634 systemd[1]: var-lib-kubelet-pods-ee9a8885\x2d8284\x2d423a\x2daae7\x2d0f835d59bd7d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4q7zx.mount: Deactivated successfully. Sep 13 00:50:59.268886 kubelet[2740]: I0913 00:50:59.268820 2740 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a207e55-7e53-468f-a610-f9fcb63a6c12" path="/var/lib/kubelet/pods/3a207e55-7e53-468f-a610-f9fcb63a6c12/volumes" Sep 13 00:50:59.271343 kubelet[2740]: I0913 00:50:59.271286 2740 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9a8885-8284-423a-aae7-0f835d59bd7d" path="/var/lib/kubelet/pods/ee9a8885-8284-423a-aae7-0f835d59bd7d/volumes" Sep 13 00:50:59.431081 kubelet[2740]: E0913 00:50:59.430402 2740 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 13 00:51:00.058795 sshd[4544]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:00.103307 systemd[1]: sshd@22-172.31.19.167:22-147.75.109.163:35426.service: Deactivated successfully. Sep 13 00:51:00.134357 systemd-logind[1836]: Session 23 logged out. Waiting for processes to exit. Sep 13 00:51:00.161105 systemd[1]: Started sshd@23-172.31.19.167:22-147.75.109.163:45620.service. Sep 13 00:51:00.161787 systemd[1]: session-23.scope: Deactivated successfully. Sep 13 00:51:00.170718 systemd-logind[1836]: Removed session 23. Sep 13 00:51:00.375555 sshd[4712]: Accepted publickey for core from 147.75.109.163 port 45620 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:51:00.376705 sshd[4712]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:00.384652 systemd-logind[1836]: New session 24 of user core. Sep 13 00:51:00.385745 systemd[1]: Started session-24.scope. Sep 13 00:51:01.083545 sshd[4712]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:01.086591 systemd-logind[1836]: Session 24 logged out. Waiting for processes to exit. Sep 13 00:51:01.088346 systemd[1]: sshd@23-172.31.19.167:22-147.75.109.163:45620.service: Deactivated successfully. Sep 13 00:51:01.089326 systemd[1]: session-24.scope: Deactivated successfully. Sep 13 00:51:01.091245 systemd-logind[1836]: Removed session 24. Sep 13 00:51:01.106458 systemd[1]: Started sshd@24-172.31.19.167:22-147.75.109.163:45630.service. Sep 13 00:51:01.126019 kubelet[2740]: E0913 00:51:01.125973 2740 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3a207e55-7e53-468f-a610-f9fcb63a6c12" containerName="cilium-agent" Sep 13 00:51:01.126605 kubelet[2740]: E0913 00:51:01.126584 2740 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ee9a8885-8284-423a-aae7-0f835d59bd7d" containerName="cilium-operator" Sep 13 00:51:01.126720 kubelet[2740]: E0913 00:51:01.126709 2740 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3a207e55-7e53-468f-a610-f9fcb63a6c12" containerName="mount-cgroup" Sep 13 00:51:01.126814 kubelet[2740]: E0913 00:51:01.126802 2740 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3a207e55-7e53-468f-a610-f9fcb63a6c12" containerName="apply-sysctl-overwrites" Sep 13 00:51:01.126907 kubelet[2740]: E0913 00:51:01.126897 2740 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3a207e55-7e53-468f-a610-f9fcb63a6c12" containerName="mount-bpf-fs" Sep 13 00:51:01.126987 kubelet[2740]: E0913 00:51:01.126977 2740 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3a207e55-7e53-468f-a610-f9fcb63a6c12" containerName="clean-cilium-state" Sep 13 00:51:01.131559 kubelet[2740]: I0913 00:51:01.131524 2740 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9a8885-8284-423a-aae7-0f835d59bd7d" containerName="cilium-operator" Sep 13 00:51:01.132397 kubelet[2740]: I0913 00:51:01.132373 2740 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a207e55-7e53-468f-a610-f9fcb63a6c12" containerName="cilium-agent" Sep 13 00:51:01.187984 kubelet[2740]: I0913 00:51:01.187368 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-clustermesh-secrets\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.187984 kubelet[2740]: I0913 00:51:01.187424 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-bpf-maps\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.187984 kubelet[2740]: I0913 00:51:01.187450 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-net\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.187984 kubelet[2740]: I0913 00:51:01.187473 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-lib-modules\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.187984 kubelet[2740]: I0913 00:51:01.187498 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-etc-cni-netd\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.187984 kubelet[2740]: I0913 00:51:01.187538 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-xtables-lock\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188497 kubelet[2740]: I0913 00:51:01.187563 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/35e019cf-9262-4984-b41b-1969327e2c32-cilium-config-path\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188497 kubelet[2740]: I0913 00:51:01.187588 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-cilium-ipsec-secrets\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188497 kubelet[2740]: I0913 00:51:01.187610 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cni-path\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188497 kubelet[2740]: I0913 00:51:01.187631 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-hubble-tls\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188497 kubelet[2740]: I0913 00:51:01.187666 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-kernel\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188497 kubelet[2740]: I0913 00:51:01.187691 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-hostproc\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188751 kubelet[2740]: I0913 00:51:01.187716 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-cgroup\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188751 kubelet[2740]: I0913 00:51:01.187741 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zpmk\" (UniqueName: \"kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-kube-api-access-8zpmk\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.188751 kubelet[2740]: I0913 00:51:01.187764 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-run\") pod \"cilium-jj8d4\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " pod="kube-system/cilium-jj8d4" Sep 13 00:51:01.295053 sshd[4724]: Accepted publickey for core from 147.75.109.163 port 45630 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:51:01.298551 sshd[4724]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:01.335805 systemd-logind[1836]: New session 25 of user core. Sep 13 00:51:01.338307 systemd[1]: Started session-25.scope. Sep 13 00:51:01.466767 env[1854]: time="2025-09-13T00:51:01.465905771Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jj8d4,Uid:35e019cf-9262-4984-b41b-1969327e2c32,Namespace:kube-system,Attempt:0,}" Sep 13 00:51:01.584438 env[1854]: time="2025-09-13T00:51:01.584345973Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:51:01.584769 env[1854]: time="2025-09-13T00:51:01.584735573Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:51:01.584888 env[1854]: time="2025-09-13T00:51:01.584864112Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:51:01.585278 env[1854]: time="2025-09-13T00:51:01.585245694Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3 pid=4742 runtime=io.containerd.runc.v2 Sep 13 00:51:02.149827 kubelet[2740]: I0913 00:51:02.149755 2740 setters.go:600] "Node became not ready" node="ip-172-31-19-167" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-13T00:51:02Z","lastTransitionTime":"2025-09-13T00:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 13 00:51:02.258921 env[1854]: time="2025-09-13T00:51:02.253334546Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jj8d4,Uid:35e019cf-9262-4984-b41b-1969327e2c32,Namespace:kube-system,Attempt:0,} returns sandbox id \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\"" Sep 13 00:51:02.282452 env[1854]: time="2025-09-13T00:51:02.271277859Z" level=info msg="CreateContainer within sandbox \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:51:02.331565 env[1854]: time="2025-09-13T00:51:02.331498326Z" level=info msg="CreateContainer within sandbox \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233\"" Sep 13 00:51:02.332536 env[1854]: time="2025-09-13T00:51:02.332494377Z" level=info msg="StartContainer for \"b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233\"" Sep 13 00:51:02.363599 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3761845104.mount: Deactivated successfully. Sep 13 00:51:02.603848 env[1854]: time="2025-09-13T00:51:02.603782488Z" level=info msg="StartContainer for \"b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233\" returns successfully" Sep 13 00:51:02.725955 sshd[4724]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:02.736653 systemd[1]: sshd@24-172.31.19.167:22-147.75.109.163:45630.service: Deactivated successfully. Sep 13 00:51:02.737852 systemd[1]: session-25.scope: Deactivated successfully. Sep 13 00:51:02.739957 systemd-logind[1836]: Session 25 logged out. Waiting for processes to exit. Sep 13 00:51:02.752261 systemd[1]: Started sshd@25-172.31.19.167:22-147.75.109.163:45642.service. Sep 13 00:51:02.775034 systemd-logind[1836]: Removed session 25. Sep 13 00:51:02.999439 env[1854]: time="2025-09-13T00:51:02.999381916Z" level=info msg="shim disconnected" id=b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233 Sep 13 00:51:02.999439 env[1854]: time="2025-09-13T00:51:02.999446804Z" level=warning msg="cleaning up after shim disconnected" id=b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233 namespace=k8s.io Sep 13 00:51:02.999787 env[1854]: time="2025-09-13T00:51:02.999463148Z" level=info msg="cleaning up dead shim" Sep 13 00:51:03.084736 sshd[4813]: Accepted publickey for core from 147.75.109.163 port 45642 ssh2: RSA SHA256:9zKSfA0UBs4YCbMNRE+jf2SchYlhVPu6zl9tBdI5N0M Sep 13 00:51:03.086001 sshd[4813]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 13 00:51:03.106841 systemd[1]: Started session-26.scope. Sep 13 00:51:03.113253 systemd-logind[1836]: New session 26 of user core. Sep 13 00:51:03.144775 env[1854]: time="2025-09-13T00:51:03.144726443Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4832 runtime=io.containerd.runc.v2\n" Sep 13 00:51:03.313712 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233-rootfs.mount: Deactivated successfully. Sep 13 00:51:03.832063 env[1854]: time="2025-09-13T00:51:03.832011068Z" level=info msg="StopPodSandbox for \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\"" Sep 13 00:51:03.832582 env[1854]: time="2025-09-13T00:51:03.832082955Z" level=info msg="Container to stop \"b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 13 00:51:03.838016 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3-shm.mount: Deactivated successfully. Sep 13 00:51:03.881138 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3-rootfs.mount: Deactivated successfully. Sep 13 00:51:03.899047 env[1854]: time="2025-09-13T00:51:03.898606009Z" level=info msg="shim disconnected" id=2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3 Sep 13 00:51:03.899047 env[1854]: time="2025-09-13T00:51:03.899036025Z" level=warning msg="cleaning up after shim disconnected" id=2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3 namespace=k8s.io Sep 13 00:51:03.899047 env[1854]: time="2025-09-13T00:51:03.899048788Z" level=info msg="cleaning up dead shim" Sep 13 00:51:03.909302 env[1854]: time="2025-09-13T00:51:03.909255217Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4871 runtime=io.containerd.runc.v2\n" Sep 13 00:51:03.909649 env[1854]: time="2025-09-13T00:51:03.909614517Z" level=info msg="TearDown network for sandbox \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" successfully" Sep 13 00:51:03.909759 env[1854]: time="2025-09-13T00:51:03.909651815Z" level=info msg="StopPodSandbox for \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" returns successfully" Sep 13 00:51:04.050598 kubelet[2740]: I0913 00:51:04.050524 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-bpf-maps\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051628 kubelet[2740]: I0913 00:51:04.050906 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-xtables-lock\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051628 kubelet[2740]: I0913 00:51:04.050957 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-cilium-ipsec-secrets\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051628 kubelet[2740]: I0913 00:51:04.051000 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-run\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051628 kubelet[2740]: I0913 00:51:04.051269 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-cgroup\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051628 kubelet[2740]: I0913 00:51:04.051328 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-etc-cni-netd\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051628 kubelet[2740]: I0913 00:51:04.051352 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-kernel\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051984 kubelet[2740]: I0913 00:51:04.051376 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-hostproc\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051984 kubelet[2740]: I0913 00:51:04.051420 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-lib-modules\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051984 kubelet[2740]: I0913 00:51:04.051464 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-net\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051984 kubelet[2740]: I0913 00:51:04.051510 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/35e019cf-9262-4984-b41b-1969327e2c32-cilium-config-path\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051984 kubelet[2740]: I0913 00:51:04.051559 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zpmk\" (UniqueName: \"kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-kube-api-access-8zpmk\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.051984 kubelet[2740]: I0913 00:51:04.051588 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-clustermesh-secrets\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.052769 kubelet[2740]: I0913 00:51:04.051611 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cni-path\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.052769 kubelet[2740]: I0913 00:51:04.051649 2740 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-hubble-tls\") pod \"35e019cf-9262-4984-b41b-1969327e2c32\" (UID: \"35e019cf-9262-4984-b41b-1969327e2c32\") " Sep 13 00:51:04.052950 kubelet[2740]: I0913 00:51:04.050751 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.053003 kubelet[2740]: I0913 00:51:04.052935 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.053117 kubelet[2740]: I0913 00:51:04.053091 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.053248 kubelet[2740]: I0913 00:51:04.053230 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.053359 kubelet[2740]: I0913 00:51:04.053343 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.053454 kubelet[2740]: I0913 00:51:04.053440 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.053542 kubelet[2740]: I0913 00:51:04.053527 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.053734 kubelet[2740]: I0913 00:51:04.053620 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-hostproc" (OuterVolumeSpecName: "hostproc") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.054220 kubelet[2740]: I0913 00:51:04.054196 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.054444 kubelet[2740]: I0913 00:51:04.054395 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cni-path" (OuterVolumeSpecName: "cni-path") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 13 00:51:04.057682 kubelet[2740]: I0913 00:51:04.057638 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e019cf-9262-4984-b41b-1969327e2c32-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 13 00:51:04.062706 systemd[1]: var-lib-kubelet-pods-35e019cf\x2d9262\x2d4984\x2db41b\x2d1969327e2c32-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Sep 13 00:51:04.066077 kubelet[2740]: I0913 00:51:04.066034 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 00:51:04.070798 systemd[1]: var-lib-kubelet-pods-35e019cf\x2d9262\x2d4984\x2db41b\x2d1969327e2c32-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 13 00:51:04.073266 kubelet[2740]: I0913 00:51:04.073224 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:51:04.077814 kubelet[2740]: I0913 00:51:04.077770 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 13 00:51:04.079699 kubelet[2740]: I0913 00:51:04.079644 2740 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-kube-api-access-8zpmk" (OuterVolumeSpecName: "kube-api-access-8zpmk") pod "35e019cf-9262-4984-b41b-1969327e2c32" (UID: "35e019cf-9262-4984-b41b-1969327e2c32"). InnerVolumeSpecName "kube-api-access-8zpmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152059 2740 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-etc-cni-netd\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152094 2740 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-kernel\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152107 2740 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-hostproc\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152116 2740 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-lib-modules\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152128 2740 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-host-proc-sys-net\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152137 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/35e019cf-9262-4984-b41b-1969327e2c32-cilium-config-path\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152156 2740 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zpmk\" (UniqueName: \"kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-kube-api-access-8zpmk\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.152308 kubelet[2740]: I0913 00:51:04.152167 2740 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-clustermesh-secrets\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.153213 kubelet[2740]: I0913 00:51:04.152175 2740 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cni-path\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.153213 kubelet[2740]: I0913 00:51:04.152182 2740 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/35e019cf-9262-4984-b41b-1969327e2c32-hubble-tls\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.153213 kubelet[2740]: I0913 00:51:04.152190 2740 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-bpf-maps\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.153213 kubelet[2740]: I0913 00:51:04.152197 2740 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-xtables-lock\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.153213 kubelet[2740]: I0913 00:51:04.152206 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/35e019cf-9262-4984-b41b-1969327e2c32-cilium-ipsec-secrets\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.153213 kubelet[2740]: I0913 00:51:04.152214 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-run\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.153213 kubelet[2740]: I0913 00:51:04.152229 2740 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/35e019cf-9262-4984-b41b-1969327e2c32-cilium-cgroup\") on node \"ip-172-31-19-167\" DevicePath \"\"" Sep 13 00:51:04.308276 systemd[1]: var-lib-kubelet-pods-35e019cf\x2d9262\x2d4984\x2db41b\x2d1969327e2c32-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d8zpmk.mount: Deactivated successfully. Sep 13 00:51:04.308483 systemd[1]: var-lib-kubelet-pods-35e019cf\x2d9262\x2d4984\x2db41b\x2d1969327e2c32-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 13 00:51:04.433307 kubelet[2740]: E0913 00:51:04.433185 2740 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 13 00:51:04.836810 kubelet[2740]: I0913 00:51:04.835213 2740 scope.go:117] "RemoveContainer" containerID="b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233" Sep 13 00:51:04.838374 env[1854]: time="2025-09-13T00:51:04.838085462Z" level=info msg="RemoveContainer for \"b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233\"" Sep 13 00:51:04.844483 env[1854]: time="2025-09-13T00:51:04.844439505Z" level=info msg="RemoveContainer for \"b90d3107dbc8d34fbeb7949141ca85e71c738693dc0cfa003a2f96807a872233\" returns successfully" Sep 13 00:51:04.911009 kubelet[2740]: E0913 00:51:04.910972 2740 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="35e019cf-9262-4984-b41b-1969327e2c32" containerName="mount-cgroup" Sep 13 00:51:04.911183 kubelet[2740]: I0913 00:51:04.911026 2740 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e019cf-9262-4984-b41b-1969327e2c32" containerName="mount-cgroup" Sep 13 00:51:04.967301 kubelet[2740]: I0913 00:51:04.967234 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-hostproc\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967301 kubelet[2740]: I0913 00:51:04.967278 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-etc-cni-netd\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967301 kubelet[2740]: I0913 00:51:04.967305 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-host-proc-sys-net\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967560 kubelet[2740]: I0913 00:51:04.967323 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-cilium-run\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967560 kubelet[2740]: I0913 00:51:04.967343 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2805a3a8-20ff-43e9-85c5-224a017a7865-hubble-tls\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967560 kubelet[2740]: I0913 00:51:04.967361 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7np8f\" (UniqueName: \"kubernetes.io/projected/2805a3a8-20ff-43e9-85c5-224a017a7865-kube-api-access-7np8f\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967560 kubelet[2740]: I0913 00:51:04.967383 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-lib-modules\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967560 kubelet[2740]: I0913 00:51:04.967401 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-xtables-lock\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967560 kubelet[2740]: I0913 00:51:04.967434 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-host-proc-sys-kernel\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967730 kubelet[2740]: I0913 00:51:04.967457 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-cni-path\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967730 kubelet[2740]: I0913 00:51:04.967474 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/2805a3a8-20ff-43e9-85c5-224a017a7865-cilium-ipsec-secrets\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967730 kubelet[2740]: I0913 00:51:04.967489 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2805a3a8-20ff-43e9-85c5-224a017a7865-cilium-config-path\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967730 kubelet[2740]: I0913 00:51:04.967504 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-bpf-maps\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967730 kubelet[2740]: I0913 00:51:04.967520 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2805a3a8-20ff-43e9-85c5-224a017a7865-cilium-cgroup\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:04.967730 kubelet[2740]: I0913 00:51:04.967536 2740 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2805a3a8-20ff-43e9-85c5-224a017a7865-clustermesh-secrets\") pod \"cilium-jfn7j\" (UID: \"2805a3a8-20ff-43e9-85c5-224a017a7865\") " pod="kube-system/cilium-jfn7j" Sep 13 00:51:05.216413 env[1854]: time="2025-09-13T00:51:05.216289617Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jfn7j,Uid:2805a3a8-20ff-43e9-85c5-224a017a7865,Namespace:kube-system,Attempt:0,}" Sep 13 00:51:05.243938 env[1854]: time="2025-09-13T00:51:05.243859200Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 13 00:51:05.244138 env[1854]: time="2025-09-13T00:51:05.243903700Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 13 00:51:05.244138 env[1854]: time="2025-09-13T00:51:05.243923805Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 13 00:51:05.244138 env[1854]: time="2025-09-13T00:51:05.244098287Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab pid=4898 runtime=io.containerd.runc.v2 Sep 13 00:51:05.271815 kubelet[2740]: I0913 00:51:05.271531 2740 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e019cf-9262-4984-b41b-1969327e2c32" path="/var/lib/kubelet/pods/35e019cf-9262-4984-b41b-1969327e2c32/volumes" Sep 13 00:51:05.294973 env[1854]: time="2025-09-13T00:51:05.294915936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jfn7j,Uid:2805a3a8-20ff-43e9-85c5-224a017a7865,Namespace:kube-system,Attempt:0,} returns sandbox id \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\"" Sep 13 00:51:05.301608 env[1854]: time="2025-09-13T00:51:05.301555461Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 13 00:51:05.335608 env[1854]: time="2025-09-13T00:51:05.335549176Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b1b8add39f4a7d9ba9c22ffeb473402474ac20613a6f200219c7b587521fdefc\"" Sep 13 00:51:05.336092 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2394698479.mount: Deactivated successfully. Sep 13 00:51:05.337324 env[1854]: time="2025-09-13T00:51:05.337288065Z" level=info msg="StartContainer for \"b1b8add39f4a7d9ba9c22ffeb473402474ac20613a6f200219c7b587521fdefc\"" Sep 13 00:51:05.462975 env[1854]: time="2025-09-13T00:51:05.462923206Z" level=info msg="StartContainer for \"b1b8add39f4a7d9ba9c22ffeb473402474ac20613a6f200219c7b587521fdefc\" returns successfully" Sep 13 00:51:05.533995 env[1854]: time="2025-09-13T00:51:05.533944738Z" level=info msg="shim disconnected" id=b1b8add39f4a7d9ba9c22ffeb473402474ac20613a6f200219c7b587521fdefc Sep 13 00:51:05.534379 env[1854]: time="2025-09-13T00:51:05.534355830Z" level=warning msg="cleaning up after shim disconnected" id=b1b8add39f4a7d9ba9c22ffeb473402474ac20613a6f200219c7b587521fdefc namespace=k8s.io Sep 13 00:51:05.534480 env[1854]: time="2025-09-13T00:51:05.534466027Z" level=info msg="cleaning up dead shim" Sep 13 00:51:05.550222 env[1854]: time="2025-09-13T00:51:05.550169276Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:05Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4982 runtime=io.containerd.runc.v2\n" Sep 13 00:51:05.843338 env[1854]: time="2025-09-13T00:51:05.843205050Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 13 00:51:05.867622 env[1854]: time="2025-09-13T00:51:05.867558337Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"45a26725661e54d67b61ae67f92f124d5663bf20b1e7985cebe7bd9855df02ff\"" Sep 13 00:51:05.869702 env[1854]: time="2025-09-13T00:51:05.868633350Z" level=info msg="StartContainer for \"45a26725661e54d67b61ae67f92f124d5663bf20b1e7985cebe7bd9855df02ff\"" Sep 13 00:51:05.929810 env[1854]: time="2025-09-13T00:51:05.929757454Z" level=info msg="StartContainer for \"45a26725661e54d67b61ae67f92f124d5663bf20b1e7985cebe7bd9855df02ff\" returns successfully" Sep 13 00:51:05.971088 env[1854]: time="2025-09-13T00:51:05.971030021Z" level=info msg="shim disconnected" id=45a26725661e54d67b61ae67f92f124d5663bf20b1e7985cebe7bd9855df02ff Sep 13 00:51:05.971088 env[1854]: time="2025-09-13T00:51:05.971086659Z" level=warning msg="cleaning up after shim disconnected" id=45a26725661e54d67b61ae67f92f124d5663bf20b1e7985cebe7bd9855df02ff namespace=k8s.io Sep 13 00:51:05.971088 env[1854]: time="2025-09-13T00:51:05.971099679Z" level=info msg="cleaning up dead shim" Sep 13 00:51:05.980547 env[1854]: time="2025-09-13T00:51:05.980500698Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:05Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5047 runtime=io.containerd.runc.v2\n" Sep 13 00:51:06.308164 systemd[1]: run-containerd-runc-k8s.io-b1b8add39f4a7d9ba9c22ffeb473402474ac20613a6f200219c7b587521fdefc-runc.5N0ZRb.mount: Deactivated successfully. Sep 13 00:51:06.308324 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b1b8add39f4a7d9ba9c22ffeb473402474ac20613a6f200219c7b587521fdefc-rootfs.mount: Deactivated successfully. Sep 13 00:51:06.845830 env[1854]: time="2025-09-13T00:51:06.845792059Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 13 00:51:06.869053 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3988087869.mount: Deactivated successfully. Sep 13 00:51:06.884834 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3841812423.mount: Deactivated successfully. Sep 13 00:51:06.888635 env[1854]: time="2025-09-13T00:51:06.888588279Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"d72454598fdba10b4d544e4ab6d8461879472a41ded22442e933d029247b4aac\"" Sep 13 00:51:06.891506 env[1854]: time="2025-09-13T00:51:06.890979309Z" level=info msg="StartContainer for \"d72454598fdba10b4d544e4ab6d8461879472a41ded22442e933d029247b4aac\"" Sep 13 00:51:06.945355 env[1854]: time="2025-09-13T00:51:06.945309869Z" level=info msg="StartContainer for \"d72454598fdba10b4d544e4ab6d8461879472a41ded22442e933d029247b4aac\" returns successfully" Sep 13 00:51:07.100499 env[1854]: time="2025-09-13T00:51:07.100355939Z" level=info msg="shim disconnected" id=d72454598fdba10b4d544e4ab6d8461879472a41ded22442e933d029247b4aac Sep 13 00:51:07.100499 env[1854]: time="2025-09-13T00:51:07.100404682Z" level=warning msg="cleaning up after shim disconnected" id=d72454598fdba10b4d544e4ab6d8461879472a41ded22442e933d029247b4aac namespace=k8s.io Sep 13 00:51:07.100499 env[1854]: time="2025-09-13T00:51:07.100418174Z" level=info msg="cleaning up dead shim" Sep 13 00:51:07.110600 env[1854]: time="2025-09-13T00:51:07.110541569Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:07Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5107 runtime=io.containerd.runc.v2\n" Sep 13 00:51:07.850919 env[1854]: time="2025-09-13T00:51:07.850869263Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 13 00:51:07.876315 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2005927168.mount: Deactivated successfully. Sep 13 00:51:07.889487 env[1854]: time="2025-09-13T00:51:07.889421397Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"f8cb503c055a07522af333207313e251519b71c678b9dfca88478897a3599144\"" Sep 13 00:51:07.891596 env[1854]: time="2025-09-13T00:51:07.891226822Z" level=info msg="StartContainer for \"f8cb503c055a07522af333207313e251519b71c678b9dfca88478897a3599144\"" Sep 13 00:51:07.956422 env[1854]: time="2025-09-13T00:51:07.956366504Z" level=info msg="StartContainer for \"f8cb503c055a07522af333207313e251519b71c678b9dfca88478897a3599144\" returns successfully" Sep 13 00:51:07.991190 env[1854]: time="2025-09-13T00:51:07.991118109Z" level=info msg="shim disconnected" id=f8cb503c055a07522af333207313e251519b71c678b9dfca88478897a3599144 Sep 13 00:51:07.991190 env[1854]: time="2025-09-13T00:51:07.991187696Z" level=warning msg="cleaning up after shim disconnected" id=f8cb503c055a07522af333207313e251519b71c678b9dfca88478897a3599144 namespace=k8s.io Sep 13 00:51:07.991190 env[1854]: time="2025-09-13T00:51:07.991197455Z" level=info msg="cleaning up dead shim" Sep 13 00:51:08.000715 env[1854]: time="2025-09-13T00:51:08.000654609Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:07Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5164 runtime=io.containerd.runc.v2\n" Sep 13 00:51:08.308133 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f8cb503c055a07522af333207313e251519b71c678b9dfca88478897a3599144-rootfs.mount: Deactivated successfully. Sep 13 00:51:08.855480 env[1854]: time="2025-09-13T00:51:08.855437398Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 13 00:51:08.885949 env[1854]: time="2025-09-13T00:51:08.885875253Z" level=info msg="CreateContainer within sandbox \"c3ad82e1e6d2b59836da91bd4c14e19839e9fd692c2ca37a8b21b43e1ad516ab\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"12fe7ef74229491c4317d098348c500aecc1a96596b99abb35cdf76cea41adc0\"" Sep 13 00:51:08.888326 env[1854]: time="2025-09-13T00:51:08.887178400Z" level=info msg="StartContainer for \"12fe7ef74229491c4317d098348c500aecc1a96596b99abb35cdf76cea41adc0\"" Sep 13 00:51:08.961819 env[1854]: time="2025-09-13T00:51:08.961774217Z" level=info msg="StartContainer for \"12fe7ef74229491c4317d098348c500aecc1a96596b99abb35cdf76cea41adc0\" returns successfully" Sep 13 00:51:09.945196 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 13 00:51:11.797697 systemd[1]: run-containerd-runc-k8s.io-12fe7ef74229491c4317d098348c500aecc1a96596b99abb35cdf76cea41adc0-runc.xEiDYM.mount: Deactivated successfully. Sep 13 00:51:13.083987 systemd-networkd[1513]: lxc_health: Link UP Sep 13 00:51:13.096641 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 13 00:51:13.096205 systemd-networkd[1513]: lxc_health: Gained carrier Sep 13 00:51:13.098354 (udev-worker)[5724]: Network interface NamePolicy= disabled on kernel command line. Sep 13 00:51:13.255066 kubelet[2740]: I0913 00:51:13.254973 2740 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-jfn7j" podStartSLOduration=9.254950433 podStartE2EDuration="9.254950433s" podCreationTimestamp="2025-09-13 00:51:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-13 00:51:09.89877108 +0000 UTC m=+150.826858321" watchObservedRunningTime="2025-09-13 00:51:13.254950433 +0000 UTC m=+154.183037672" Sep 13 00:51:14.342401 systemd-networkd[1513]: lxc_health: Gained IPv6LL Sep 13 00:51:14.398237 systemd[1]: run-containerd-runc-k8s.io-12fe7ef74229491c4317d098348c500aecc1a96596b99abb35cdf76cea41adc0-runc.c3jMgF.mount: Deactivated successfully. Sep 13 00:51:21.041256 systemd[1]: run-containerd-runc-k8s.io-12fe7ef74229491c4317d098348c500aecc1a96596b99abb35cdf76cea41adc0-runc.JKsfTo.mount: Deactivated successfully. Sep 13 00:51:21.126891 sshd[4813]: pam_unix(sshd:session): session closed for user core Sep 13 00:51:21.129753 systemd[1]: sshd@25-172.31.19.167:22-147.75.109.163:45642.service: Deactivated successfully. Sep 13 00:51:21.131079 systemd[1]: session-26.scope: Deactivated successfully. Sep 13 00:51:21.131761 systemd-logind[1836]: Session 26 logged out. Waiting for processes to exit. Sep 13 00:51:21.132926 systemd-logind[1836]: Removed session 26. Sep 13 00:51:35.630810 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-74b189b03804fcdfee3a762ed74baf732bd631cc28d7367d6dbb4bd21504f283-rootfs.mount: Deactivated successfully. Sep 13 00:51:35.670356 env[1854]: time="2025-09-13T00:51:35.670306880Z" level=info msg="shim disconnected" id=74b189b03804fcdfee3a762ed74baf732bd631cc28d7367d6dbb4bd21504f283 Sep 13 00:51:35.670356 env[1854]: time="2025-09-13T00:51:35.670357206Z" level=warning msg="cleaning up after shim disconnected" id=74b189b03804fcdfee3a762ed74baf732bd631cc28d7367d6dbb4bd21504f283 namespace=k8s.io Sep 13 00:51:35.671014 env[1854]: time="2025-09-13T00:51:35.670368990Z" level=info msg="cleaning up dead shim" Sep 13 00:51:35.680277 env[1854]: time="2025-09-13T00:51:35.680228246Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:35Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5864 runtime=io.containerd.runc.v2\n" Sep 13 00:51:35.926641 kubelet[2740]: I0913 00:51:35.926139 2740 scope.go:117] "RemoveContainer" containerID="74b189b03804fcdfee3a762ed74baf732bd631cc28d7367d6dbb4bd21504f283" Sep 13 00:51:35.932971 env[1854]: time="2025-09-13T00:51:35.932928952Z" level=info msg="CreateContainer within sandbox \"fec989cef7de0aefcf0cea7bbc9e4b70c1b79e7dc0968b572dc65abc939cb360\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Sep 13 00:51:35.992211 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2761906454.mount: Deactivated successfully. Sep 13 00:51:36.006655 env[1854]: time="2025-09-13T00:51:36.006025308Z" level=info msg="CreateContainer within sandbox \"fec989cef7de0aefcf0cea7bbc9e4b70c1b79e7dc0968b572dc65abc939cb360\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"244a85e31d2de4587c71bdfd583b62a3f82b9b31236b030944527c8aa73debac\"" Sep 13 00:51:36.006904 env[1854]: time="2025-09-13T00:51:36.006856368Z" level=info msg="StartContainer for \"244a85e31d2de4587c71bdfd583b62a3f82b9b31236b030944527c8aa73debac\"" Sep 13 00:51:36.084790 env[1854]: time="2025-09-13T00:51:36.084716421Z" level=info msg="StartContainer for \"244a85e31d2de4587c71bdfd583b62a3f82b9b31236b030944527c8aa73debac\" returns successfully" Sep 13 00:51:39.337264 env[1854]: time="2025-09-13T00:51:39.337209914Z" level=info msg="StopPodSandbox for \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\"" Sep 13 00:51:39.337788 env[1854]: time="2025-09-13T00:51:39.337325060Z" level=info msg="TearDown network for sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" successfully" Sep 13 00:51:39.337788 env[1854]: time="2025-09-13T00:51:39.337369573Z" level=info msg="StopPodSandbox for \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" returns successfully" Sep 13 00:51:39.337900 env[1854]: time="2025-09-13T00:51:39.337783031Z" level=info msg="RemovePodSandbox for \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\"" Sep 13 00:51:39.337900 env[1854]: time="2025-09-13T00:51:39.337816912Z" level=info msg="Forcibly stopping sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\"" Sep 13 00:51:39.337999 env[1854]: time="2025-09-13T00:51:39.337921668Z" level=info msg="TearDown network for sandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" successfully" Sep 13 00:51:39.346885 env[1854]: time="2025-09-13T00:51:39.346837599Z" level=info msg="RemovePodSandbox \"031ec6e3a709ca76a768d40d565860e1ac5cd53b42afd45dc11a30adc7e4d7f5\" returns successfully" Sep 13 00:51:39.347544 env[1854]: time="2025-09-13T00:51:39.347508248Z" level=info msg="StopPodSandbox for \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\"" Sep 13 00:51:39.347661 env[1854]: time="2025-09-13T00:51:39.347605233Z" level=info msg="TearDown network for sandbox \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" successfully" Sep 13 00:51:39.347661 env[1854]: time="2025-09-13T00:51:39.347638621Z" level=info msg="StopPodSandbox for \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" returns successfully" Sep 13 00:51:39.347999 env[1854]: time="2025-09-13T00:51:39.347956206Z" level=info msg="RemovePodSandbox for \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\"" Sep 13 00:51:39.347999 env[1854]: time="2025-09-13T00:51:39.347987417Z" level=info msg="Forcibly stopping sandbox \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\"" Sep 13 00:51:39.348179 env[1854]: time="2025-09-13T00:51:39.348069213Z" level=info msg="TearDown network for sandbox \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" successfully" Sep 13 00:51:39.353760 env[1854]: time="2025-09-13T00:51:39.353706367Z" level=info msg="RemovePodSandbox \"2ca4d3eb9f725d0fa30260597a2c74ac21fe873d28337d05d5f36bfd2125fce3\" returns successfully" Sep 13 00:51:39.356174 env[1854]: time="2025-09-13T00:51:39.354642437Z" level=info msg="StopPodSandbox for \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\"" Sep 13 00:51:39.356174 env[1854]: time="2025-09-13T00:51:39.354774849Z" level=info msg="TearDown network for sandbox \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" successfully" Sep 13 00:51:39.356174 env[1854]: time="2025-09-13T00:51:39.354819270Z" level=info msg="StopPodSandbox for \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" returns successfully" Sep 13 00:51:39.359291 env[1854]: time="2025-09-13T00:51:39.359249170Z" level=info msg="RemovePodSandbox for \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\"" Sep 13 00:51:39.359438 env[1854]: time="2025-09-13T00:51:39.359297619Z" level=info msg="Forcibly stopping sandbox \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\"" Sep 13 00:51:39.359438 env[1854]: time="2025-09-13T00:51:39.359423735Z" level=info msg="TearDown network for sandbox \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" successfully" Sep 13 00:51:39.365997 env[1854]: time="2025-09-13T00:51:39.365934409Z" level=info msg="RemovePodSandbox \"473e40331e3c822d4760c2d19baa5d9405555c5bc3de7c885bcd1c8e2992e4fa\" returns successfully" Sep 13 00:51:41.564097 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-dae12014b4cd4a8cdb9364f499f957163f235bd00597fda1eb6478b54f440430-rootfs.mount: Deactivated successfully. Sep 13 00:51:41.592255 env[1854]: time="2025-09-13T00:51:41.592213318Z" level=info msg="shim disconnected" id=dae12014b4cd4a8cdb9364f499f957163f235bd00597fda1eb6478b54f440430 Sep 13 00:51:41.592773 env[1854]: time="2025-09-13T00:51:41.592737056Z" level=warning msg="cleaning up after shim disconnected" id=dae12014b4cd4a8cdb9364f499f957163f235bd00597fda1eb6478b54f440430 namespace=k8s.io Sep 13 00:51:41.592773 env[1854]: time="2025-09-13T00:51:41.592763195Z" level=info msg="cleaning up dead shim" Sep 13 00:51:41.601014 env[1854]: time="2025-09-13T00:51:41.600971784Z" level=warning msg="cleanup warnings time=\"2025-09-13T00:51:41Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5928 runtime=io.containerd.runc.v2\n" Sep 13 00:51:41.944731 kubelet[2740]: I0913 00:51:41.944341 2740 scope.go:117] "RemoveContainer" containerID="dae12014b4cd4a8cdb9364f499f957163f235bd00597fda1eb6478b54f440430" Sep 13 00:51:41.947194 env[1854]: time="2025-09-13T00:51:41.947134276Z" level=info msg="CreateContainer within sandbox \"cb7a499ed5522a7253b700a2bcf6a6918813badde9aa26906a4028e1c1673a1c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Sep 13 00:51:41.964570 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount548295159.mount: Deactivated successfully. Sep 13 00:51:41.971974 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1647028982.mount: Deactivated successfully. Sep 13 00:51:41.981602 env[1854]: time="2025-09-13T00:51:41.981525096Z" level=info msg="CreateContainer within sandbox \"cb7a499ed5522a7253b700a2bcf6a6918813badde9aa26906a4028e1c1673a1c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"3015209061b0db5897d957c55f9406cc653e706b84d9187dd22b0fed568fd510\"" Sep 13 00:51:41.982134 env[1854]: time="2025-09-13T00:51:41.982106623Z" level=info msg="StartContainer for \"3015209061b0db5897d957c55f9406cc653e706b84d9187dd22b0fed568fd510\"" Sep 13 00:51:42.056404 env[1854]: time="2025-09-13T00:51:42.056352031Z" level=info msg="StartContainer for \"3015209061b0db5897d957c55f9406cc653e706b84d9187dd22b0fed568fd510\" returns successfully" Sep 13 00:51:42.642086 kubelet[2740]: E0913 00:51:42.642024 2740 controller.go:195] "Failed to update lease" err="Put \"https://172.31.19.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-19-167?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 13 00:51:52.643008 kubelet[2740]: E0913 00:51:52.642929 2740 controller.go:195] "Failed to update lease" err="Put \"https://172.31.19.167:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-19-167?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"