May 27 03:26:25.927178 kernel: Linux version 6.12.30-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Tue May 27 01:09:43 -00 2025 May 27 03:26:25.927210 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=f6c186658a19d5a08471ef76df75f82494b37b46908f9237b2c3cf497da860c6 May 27 03:26:25.927220 kernel: BIOS-provided physical RAM map: May 27 03:26:25.927227 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable May 27 03:26:25.927233 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable May 27 03:26:25.927240 kernel: BIOS-e820: [mem 0x00000000786ce000-0x000000007894dfff] reserved May 27 03:26:25.927248 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data May 27 03:26:25.927255 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS May 27 03:26:25.927279 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable May 27 03:26:25.927285 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved May 27 03:26:25.927292 kernel: NX (Execute Disable) protection: active May 27 03:26:25.927299 kernel: APIC: Static calls initialized May 27 03:26:25.927306 kernel: e820: update [mem 0x768c0018-0x768c8e57] usable ==> usable May 27 03:26:25.927313 kernel: extended physical RAM map: May 27 03:26:25.927324 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable May 27 03:26:25.927332 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000768c0017] usable May 27 03:26:25.927340 kernel: reserve setup_data: [mem 0x00000000768c0018-0x00000000768c8e57] usable May 27 03:26:25.927347 kernel: reserve setup_data: [mem 0x00000000768c8e58-0x00000000786cdfff] usable May 27 03:26:25.927355 kernel: reserve setup_data: [mem 0x00000000786ce000-0x000000007894dfff] reserved May 27 03:26:25.927362 kernel: reserve setup_data: [mem 0x000000007894e000-0x000000007895dfff] ACPI data May 27 03:26:25.927370 kernel: reserve setup_data: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS May 27 03:26:25.927378 kernel: reserve setup_data: [mem 0x00000000789de000-0x000000007c97bfff] usable May 27 03:26:25.927385 kernel: reserve setup_data: [mem 0x000000007c97c000-0x000000007c9fffff] reserved May 27 03:26:25.927393 kernel: efi: EFI v2.7 by EDK II May 27 03:26:25.927403 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77003518 May 27 03:26:25.927410 kernel: secureboot: Secure boot disabled May 27 03:26:25.927418 kernel: SMBIOS 2.7 present. May 27 03:26:25.927425 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 May 27 03:26:25.927433 kernel: DMI: Memory slots populated: 1/1 May 27 03:26:25.927440 kernel: Hypervisor detected: KVM May 27 03:26:25.927448 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 27 03:26:25.927456 kernel: kvm-clock: using sched offset of 5005334276 cycles May 27 03:26:25.927464 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 27 03:26:25.927472 kernel: tsc: Detected 2499.996 MHz processor May 27 03:26:25.927479 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 27 03:26:25.927490 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 27 03:26:25.927498 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 May 27 03:26:25.927506 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs May 27 03:26:25.927513 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 27 03:26:25.927521 kernel: Using GB pages for direct mapping May 27 03:26:25.927533 kernel: ACPI: Early table checksum verification disabled May 27 03:26:25.927541 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) May 27 03:26:25.927551 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) May 27 03:26:25.927560 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) May 27 03:26:25.927568 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) May 27 03:26:25.927576 kernel: ACPI: FACS 0x00000000789D0000 000040 May 27 03:26:25.927584 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) May 27 03:26:25.927592 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) May 27 03:26:25.927600 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) May 27 03:26:25.927611 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) May 27 03:26:25.927619 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) May 27 03:26:25.927627 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) May 27 03:26:25.927635 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) May 27 03:26:25.927644 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) May 27 03:26:25.927652 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] May 27 03:26:25.927661 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] May 27 03:26:25.927669 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] May 27 03:26:25.927677 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] May 27 03:26:25.927687 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] May 27 03:26:25.927695 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] May 27 03:26:25.927704 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] May 27 03:26:25.927712 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] May 27 03:26:25.927720 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] May 27 03:26:25.927728 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] May 27 03:26:25.927736 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] May 27 03:26:25.927744 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] May 27 03:26:25.927752 kernel: NUMA: Initialized distance table, cnt=1 May 27 03:26:25.927763 kernel: NODE_DATA(0) allocated [mem 0x7a8eddc0-0x7a8f4fff] May 27 03:26:25.927771 kernel: Zone ranges: May 27 03:26:25.927779 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 27 03:26:25.927787 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] May 27 03:26:25.927796 kernel: Normal empty May 27 03:26:25.927804 kernel: Device empty May 27 03:26:25.927812 kernel: Movable zone start for each node May 27 03:26:25.927820 kernel: Early memory node ranges May 27 03:26:25.927828 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] May 27 03:26:25.927836 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] May 27 03:26:25.927847 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] May 27 03:26:25.927855 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] May 27 03:26:25.927863 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 27 03:26:25.927871 kernel: On node 0, zone DMA: 96 pages in unavailable ranges May 27 03:26:25.927879 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges May 27 03:26:25.927888 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges May 27 03:26:25.927896 kernel: ACPI: PM-Timer IO Port: 0xb008 May 27 03:26:25.927904 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 27 03:26:25.927912 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 May 27 03:26:25.927923 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 27 03:26:25.927931 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 27 03:26:25.929267 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 27 03:26:25.929280 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 27 03:26:25.929289 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 27 03:26:25.929298 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 27 03:26:25.929307 kernel: TSC deadline timer available May 27 03:26:25.929315 kernel: CPU topo: Max. logical packages: 1 May 27 03:26:25.929324 kernel: CPU topo: Max. logical dies: 1 May 27 03:26:25.929337 kernel: CPU topo: Max. dies per package: 1 May 27 03:26:25.929345 kernel: CPU topo: Max. threads per core: 2 May 27 03:26:25.929353 kernel: CPU topo: Num. cores per package: 1 May 27 03:26:25.929362 kernel: CPU topo: Num. threads per package: 2 May 27 03:26:25.929370 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs May 27 03:26:25.929379 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 27 03:26:25.929388 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices May 27 03:26:25.929396 kernel: Booting paravirtualized kernel on KVM May 27 03:26:25.929405 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 27 03:26:25.929417 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 May 27 03:26:25.929425 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 May 27 03:26:25.929434 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 May 27 03:26:25.929442 kernel: pcpu-alloc: [0] 0 1 May 27 03:26:25.929451 kernel: kvm-guest: PV spinlocks enabled May 27 03:26:25.929459 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) May 27 03:26:25.929469 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=f6c186658a19d5a08471ef76df75f82494b37b46908f9237b2c3cf497da860c6 May 27 03:26:25.929479 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 27 03:26:25.929489 kernel: random: crng init done May 27 03:26:25.929497 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 27 03:26:25.929506 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 27 03:26:25.929514 kernel: Fallback order for Node 0: 0 May 27 03:26:25.929522 kernel: Built 1 zonelists, mobility grouping on. Total pages: 509451 May 27 03:26:25.929531 kernel: Policy zone: DMA32 May 27 03:26:25.929547 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 27 03:26:25.929558 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 27 03:26:25.929567 kernel: Kernel/User page tables isolation: enabled May 27 03:26:25.929576 kernel: ftrace: allocating 40081 entries in 157 pages May 27 03:26:25.929585 kernel: ftrace: allocated 157 pages with 5 groups May 27 03:26:25.929594 kernel: Dynamic Preempt: voluntary May 27 03:26:25.929605 kernel: rcu: Preemptible hierarchical RCU implementation. May 27 03:26:25.929614 kernel: rcu: RCU event tracing is enabled. May 27 03:26:25.929623 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 27 03:26:25.929632 kernel: Trampoline variant of Tasks RCU enabled. May 27 03:26:25.929641 kernel: Rude variant of Tasks RCU enabled. May 27 03:26:25.929652 kernel: Tracing variant of Tasks RCU enabled. May 27 03:26:25.929661 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 27 03:26:25.929670 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 27 03:26:25.929679 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 27 03:26:25.929688 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 27 03:26:25.929697 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 27 03:26:25.929706 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 27 03:26:25.929715 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 27 03:26:25.929726 kernel: Console: colour dummy device 80x25 May 27 03:26:25.929735 kernel: printk: legacy console [tty0] enabled May 27 03:26:25.929744 kernel: printk: legacy console [ttyS0] enabled May 27 03:26:25.929753 kernel: ACPI: Core revision 20240827 May 27 03:26:25.929762 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns May 27 03:26:25.929771 kernel: APIC: Switch to symmetric I/O mode setup May 27 03:26:25.929780 kernel: x2apic enabled May 27 03:26:25.929789 kernel: APIC: Switched APIC routing to: physical x2apic May 27 03:26:25.929798 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns May 27 03:26:25.929807 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499996) May 27 03:26:25.929819 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 May 27 03:26:25.929828 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 May 27 03:26:25.929837 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 27 03:26:25.929845 kernel: Spectre V2 : Mitigation: Retpolines May 27 03:26:25.929854 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT May 27 03:26:25.929862 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! May 27 03:26:25.929872 kernel: RETBleed: Vulnerable May 27 03:26:25.929880 kernel: Speculative Store Bypass: Vulnerable May 27 03:26:25.929889 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode May 27 03:26:25.929898 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 27 03:26:25.929908 kernel: GDS: Unknown: Dependent on hypervisor status May 27 03:26:25.929917 kernel: ITS: Mitigation: Aligned branch/return thunks May 27 03:26:25.929926 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 27 03:26:25.929934 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 27 03:26:25.931934 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 27 03:26:25.931967 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' May 27 03:26:25.931977 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' May 27 03:26:25.931985 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' May 27 03:26:25.931994 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' May 27 03:26:25.932003 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' May 27 03:26:25.932012 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' May 27 03:26:25.932025 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 27 03:26:25.932034 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 May 27 03:26:25.932043 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 May 27 03:26:25.932051 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 May 27 03:26:25.932060 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 May 27 03:26:25.932069 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 May 27 03:26:25.932077 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 May 27 03:26:25.932086 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. May 27 03:26:25.932095 kernel: Freeing SMP alternatives memory: 32K May 27 03:26:25.932103 kernel: pid_max: default: 32768 minimum: 301 May 27 03:26:25.932112 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima May 27 03:26:25.932124 kernel: landlock: Up and running. May 27 03:26:25.932132 kernel: SELinux: Initializing. May 27 03:26:25.932141 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 27 03:26:25.932150 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 27 03:26:25.932162 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) May 27 03:26:25.932171 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. May 27 03:26:25.932180 kernel: signal: max sigframe size: 3632 May 27 03:26:25.932189 kernel: rcu: Hierarchical SRCU implementation. May 27 03:26:25.932199 kernel: rcu: Max phase no-delay instances is 400. May 27 03:26:25.932208 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level May 27 03:26:25.932220 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 27 03:26:25.932229 kernel: smp: Bringing up secondary CPUs ... May 27 03:26:25.932238 kernel: smpboot: x86: Booting SMP configuration: May 27 03:26:25.932247 kernel: .... node #0, CPUs: #1 May 27 03:26:25.932256 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. May 27 03:26:25.932266 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. May 27 03:26:25.932275 kernel: smp: Brought up 1 node, 2 CPUs May 27 03:26:25.932284 kernel: smpboot: Total of 2 processors activated (9999.98 BogoMIPS) May 27 03:26:25.932293 kernel: Memory: 1908048K/2037804K available (14336K kernel code, 2430K rwdata, 9952K rodata, 54416K init, 2552K bss, 125192K reserved, 0K cma-reserved) May 27 03:26:25.932304 kernel: devtmpfs: initialized May 27 03:26:25.932313 kernel: x86/mm: Memory block size: 128MB May 27 03:26:25.932322 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) May 27 03:26:25.932332 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 27 03:26:25.932341 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 27 03:26:25.932350 kernel: pinctrl core: initialized pinctrl subsystem May 27 03:26:25.932359 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 27 03:26:25.932368 kernel: audit: initializing netlink subsys (disabled) May 27 03:26:25.932379 kernel: audit: type=2000 audit(1748316383.127:1): state=initialized audit_enabled=0 res=1 May 27 03:26:25.932392 kernel: thermal_sys: Registered thermal governor 'step_wise' May 27 03:26:25.932405 kernel: thermal_sys: Registered thermal governor 'user_space' May 27 03:26:25.932415 kernel: cpuidle: using governor menu May 27 03:26:25.932430 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 27 03:26:25.932444 kernel: dca service started, version 1.12.1 May 27 03:26:25.932457 kernel: PCI: Using configuration type 1 for base access May 27 03:26:25.932470 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 27 03:26:25.932484 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages May 27 03:26:25.932500 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page May 27 03:26:25.932513 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 27 03:26:25.932523 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 27 03:26:25.932532 kernel: ACPI: Added _OSI(Module Device) May 27 03:26:25.932541 kernel: ACPI: Added _OSI(Processor Device) May 27 03:26:25.932549 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 27 03:26:25.932558 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 27 03:26:25.932567 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded May 27 03:26:25.932576 kernel: ACPI: Interpreter enabled May 27 03:26:25.932585 kernel: ACPI: PM: (supports S0 S5) May 27 03:26:25.932596 kernel: ACPI: Using IOAPIC for interrupt routing May 27 03:26:25.932605 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 27 03:26:25.932614 kernel: PCI: Using E820 reservations for host bridge windows May 27 03:26:25.932623 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F May 27 03:26:25.932632 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 27 03:26:25.932806 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 27 03:26:25.932902 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] May 27 03:26:25.933451 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge May 27 03:26:25.933469 kernel: acpiphp: Slot [3] registered May 27 03:26:25.933479 kernel: acpiphp: Slot [4] registered May 27 03:26:25.933488 kernel: acpiphp: Slot [5] registered May 27 03:26:25.933497 kernel: acpiphp: Slot [6] registered May 27 03:26:25.933506 kernel: acpiphp: Slot [7] registered May 27 03:26:25.933515 kernel: acpiphp: Slot [8] registered May 27 03:26:25.933524 kernel: acpiphp: Slot [9] registered May 27 03:26:25.933533 kernel: acpiphp: Slot [10] registered May 27 03:26:25.933547 kernel: acpiphp: Slot [11] registered May 27 03:26:25.933556 kernel: acpiphp: Slot [12] registered May 27 03:26:25.933565 kernel: acpiphp: Slot [13] registered May 27 03:26:25.933574 kernel: acpiphp: Slot [14] registered May 27 03:26:25.933583 kernel: acpiphp: Slot [15] registered May 27 03:26:25.933592 kernel: acpiphp: Slot [16] registered May 27 03:26:25.933601 kernel: acpiphp: Slot [17] registered May 27 03:26:25.933610 kernel: acpiphp: Slot [18] registered May 27 03:26:25.933619 kernel: acpiphp: Slot [19] registered May 27 03:26:25.933630 kernel: acpiphp: Slot [20] registered May 27 03:26:25.933639 kernel: acpiphp: Slot [21] registered May 27 03:26:25.933648 kernel: acpiphp: Slot [22] registered May 27 03:26:25.933656 kernel: acpiphp: Slot [23] registered May 27 03:26:25.933665 kernel: acpiphp: Slot [24] registered May 27 03:26:25.933674 kernel: acpiphp: Slot [25] registered May 27 03:26:25.933683 kernel: acpiphp: Slot [26] registered May 27 03:26:25.933692 kernel: acpiphp: Slot [27] registered May 27 03:26:25.933701 kernel: acpiphp: Slot [28] registered May 27 03:26:25.933710 kernel: acpiphp: Slot [29] registered May 27 03:26:25.933721 kernel: acpiphp: Slot [30] registered May 27 03:26:25.933730 kernel: acpiphp: Slot [31] registered May 27 03:26:25.933738 kernel: PCI host bridge to bus 0000:00 May 27 03:26:25.933841 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 27 03:26:25.933925 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 27 03:26:25.934440 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 27 03:26:25.934531 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] May 27 03:26:25.934611 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] May 27 03:26:25.934694 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 27 03:26:25.934799 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint May 27 03:26:25.934899 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint May 27 03:26:25.935030 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 conventional PCI endpoint May 27 03:26:25.935122 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI May 27 03:26:25.935216 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff May 27 03:26:25.935305 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff May 27 03:26:25.935393 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff May 27 03:26:25.935480 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff May 27 03:26:25.935827 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff May 27 03:26:25.935930 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff May 27 03:26:25.936052 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 conventional PCI endpoint May 27 03:26:25.936143 kernel: pci 0000:00:03.0: BAR 0 [mem 0x80000000-0x803fffff pref] May 27 03:26:25.936238 kernel: pci 0000:00:03.0: ROM [mem 0xffff0000-0xffffffff pref] May 27 03:26:25.936340 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 27 03:26:25.936446 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 PCIe Endpoint May 27 03:26:25.936538 kernel: pci 0000:00:04.0: BAR 0 [mem 0x80404000-0x80407fff] May 27 03:26:25.936634 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 PCIe Endpoint May 27 03:26:25.936723 kernel: pci 0000:00:05.0: BAR 0 [mem 0x80400000-0x80403fff] May 27 03:26:25.936739 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 27 03:26:25.936748 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 27 03:26:25.936757 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 27 03:26:25.936766 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 27 03:26:25.936775 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 27 03:26:25.936784 kernel: iommu: Default domain type: Translated May 27 03:26:25.936793 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 27 03:26:25.936802 kernel: efivars: Registered efivars operations May 27 03:26:25.936811 kernel: PCI: Using ACPI for IRQ routing May 27 03:26:25.936823 kernel: PCI: pci_cache_line_size set to 64 bytes May 27 03:26:25.936832 kernel: e820: reserve RAM buffer [mem 0x768c0018-0x77ffffff] May 27 03:26:25.936841 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] May 27 03:26:25.936849 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] May 27 03:26:25.937009 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device May 27 03:26:25.937104 kernel: pci 0000:00:03.0: vgaarb: bridge control possible May 27 03:26:25.937194 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 27 03:26:25.937206 kernel: vgaarb: loaded May 27 03:26:25.937219 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 May 27 03:26:25.937229 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter May 27 03:26:25.937238 kernel: clocksource: Switched to clocksource kvm-clock May 27 03:26:25.937247 kernel: VFS: Disk quotas dquot_6.6.0 May 27 03:26:25.937256 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 27 03:26:25.937265 kernel: pnp: PnP ACPI init May 27 03:26:25.937274 kernel: pnp: PnP ACPI: found 5 devices May 27 03:26:25.937283 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 27 03:26:25.937292 kernel: NET: Registered PF_INET protocol family May 27 03:26:25.937303 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) May 27 03:26:25.937312 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) May 27 03:26:25.937321 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 27 03:26:25.937330 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) May 27 03:26:25.937339 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) May 27 03:26:25.937348 kernel: TCP: Hash tables configured (established 16384 bind 16384) May 27 03:26:25.937357 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) May 27 03:26:25.937366 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) May 27 03:26:25.937375 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 27 03:26:25.937387 kernel: NET: Registered PF_XDP protocol family May 27 03:26:25.937474 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 27 03:26:25.937554 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 27 03:26:25.937634 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 27 03:26:25.937718 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] May 27 03:26:25.937797 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] May 27 03:26:25.937891 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 27 03:26:25.937904 kernel: PCI: CLS 0 bytes, default 64 May 27 03:26:25.937917 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 27 03:26:25.937926 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns May 27 03:26:25.937955 kernel: clocksource: Switched to clocksource tsc May 27 03:26:25.937965 kernel: Initialise system trusted keyrings May 27 03:26:25.937974 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 May 27 03:26:25.937983 kernel: Key type asymmetric registered May 27 03:26:25.937992 kernel: Asymmetric key parser 'x509' registered May 27 03:26:25.938001 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 27 03:26:25.938010 kernel: io scheduler mq-deadline registered May 27 03:26:25.938022 kernel: io scheduler kyber registered May 27 03:26:25.938031 kernel: io scheduler bfq registered May 27 03:26:25.938040 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 27 03:26:25.938049 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 27 03:26:25.938058 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 27 03:26:25.938067 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 27 03:26:25.938076 kernel: i8042: Warning: Keylock active May 27 03:26:25.938085 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 27 03:26:25.938094 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 27 03:26:25.938199 kernel: rtc_cmos 00:00: RTC can wake from S4 May 27 03:26:25.938283 kernel: rtc_cmos 00:00: registered as rtc0 May 27 03:26:25.938366 kernel: rtc_cmos 00:00: setting system clock to 2025-05-27T03:26:25 UTC (1748316385) May 27 03:26:25.938448 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram May 27 03:26:25.938479 kernel: intel_pstate: CPU model not supported May 27 03:26:25.938491 kernel: efifb: probing for efifb May 27 03:26:25.938501 kernel: efifb: framebuffer at 0x80000000, using 1876k, total 1875k May 27 03:26:25.938512 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 May 27 03:26:25.938522 kernel: efifb: scrolling: redraw May 27 03:26:25.938531 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 May 27 03:26:25.938540 kernel: Console: switching to colour frame buffer device 100x37 May 27 03:26:25.938550 kernel: fb0: EFI VGA frame buffer device May 27 03:26:25.938559 kernel: pstore: Using crash dump compression: deflate May 27 03:26:25.938569 kernel: pstore: Registered efi_pstore as persistent store backend May 27 03:26:25.938578 kernel: NET: Registered PF_INET6 protocol family May 27 03:26:25.938588 kernel: Segment Routing with IPv6 May 27 03:26:25.938597 kernel: In-situ OAM (IOAM) with IPv6 May 27 03:26:25.938609 kernel: NET: Registered PF_PACKET protocol family May 27 03:26:25.938618 kernel: Key type dns_resolver registered May 27 03:26:25.938627 kernel: IPI shorthand broadcast: enabled May 27 03:26:25.938637 kernel: sched_clock: Marking stable (2666001944, 156864941)->(2922561613, -99694728) May 27 03:26:25.938646 kernel: registered taskstats version 1 May 27 03:26:25.938656 kernel: Loading compiled-in X.509 certificates May 27 03:26:25.938665 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.30-flatcar: ba9eddccb334a70147f3ddfe4fbde029feaa991d' May 27 03:26:25.938674 kernel: Demotion targets for Node 0: null May 27 03:26:25.938684 kernel: Key type .fscrypt registered May 27 03:26:25.938695 kernel: Key type fscrypt-provisioning registered May 27 03:26:25.938704 kernel: ima: No TPM chip found, activating TPM-bypass! May 27 03:26:25.938713 kernel: ima: Allocated hash algorithm: sha1 May 27 03:26:25.938723 kernel: ima: No architecture policies found May 27 03:26:25.938732 kernel: clk: Disabling unused clocks May 27 03:26:25.938741 kernel: Warning: unable to open an initial console. May 27 03:26:25.938751 kernel: Freeing unused kernel image (initmem) memory: 54416K May 27 03:26:25.938760 kernel: Write protecting the kernel read-only data: 24576k May 27 03:26:25.938772 kernel: Freeing unused kernel image (rodata/data gap) memory: 288K May 27 03:26:25.938784 kernel: Run /init as init process May 27 03:26:25.938793 kernel: with arguments: May 27 03:26:25.938802 kernel: /init May 27 03:26:25.938811 kernel: with environment: May 27 03:26:25.938823 kernel: HOME=/ May 27 03:26:25.938834 kernel: TERM=linux May 27 03:26:25.938844 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 27 03:26:25.938854 systemd[1]: Successfully made /usr/ read-only. May 27 03:26:25.938867 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 27 03:26:25.938877 systemd[1]: Detected virtualization amazon. May 27 03:26:25.938887 systemd[1]: Detected architecture x86-64. May 27 03:26:25.938896 systemd[1]: Running in initrd. May 27 03:26:25.938908 systemd[1]: No hostname configured, using default hostname. May 27 03:26:25.938918 systemd[1]: Hostname set to . May 27 03:26:25.938928 systemd[1]: Initializing machine ID from VM UUID. May 27 03:26:25.941262 systemd[1]: Queued start job for default target initrd.target. May 27 03:26:25.941281 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 27 03:26:25.941293 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 27 03:26:25.941306 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 27 03:26:25.941317 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 27 03:26:25.941333 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 27 03:26:25.941344 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 27 03:26:25.941356 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 27 03:26:25.941366 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 27 03:26:25.941376 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 27 03:26:25.941386 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 27 03:26:25.941396 systemd[1]: Reached target paths.target - Path Units. May 27 03:26:25.941409 systemd[1]: Reached target slices.target - Slice Units. May 27 03:26:25.941419 systemd[1]: Reached target swap.target - Swaps. May 27 03:26:25.941429 systemd[1]: Reached target timers.target - Timer Units. May 27 03:26:25.941439 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 27 03:26:25.941449 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 27 03:26:25.941459 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 27 03:26:25.941469 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 27 03:26:25.941479 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 27 03:26:25.941489 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 27 03:26:25.941501 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 27 03:26:25.941510 systemd[1]: Reached target sockets.target - Socket Units. May 27 03:26:25.941520 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 27 03:26:25.941530 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 27 03:26:25.941540 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 27 03:26:25.941550 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). May 27 03:26:25.941560 systemd[1]: Starting systemd-fsck-usr.service... May 27 03:26:25.941570 systemd[1]: Starting systemd-journald.service - Journal Service... May 27 03:26:25.941582 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 27 03:26:25.941592 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:26:25.941602 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 27 03:26:25.941612 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 27 03:26:25.941624 systemd[1]: Finished systemd-fsck-usr.service. May 27 03:26:25.941635 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 27 03:26:25.941682 systemd-journald[207]: Collecting audit messages is disabled. May 27 03:26:25.941707 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:26:25.941721 systemd-journald[207]: Journal started May 27 03:26:25.941743 systemd-journald[207]: Runtime Journal (/run/log/journal/ec2fb34c2679c45c75604177d1a7e24a) is 4.8M, max 38.4M, 33.6M free. May 27 03:26:25.924584 systemd-modules-load[208]: Inserted module 'overlay' May 27 03:26:25.945967 systemd[1]: Started systemd-journald.service - Journal Service. May 27 03:26:25.948209 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 27 03:26:25.957060 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 27 03:26:25.957089 kernel: Bridge firewalling registered May 27 03:26:25.953000 systemd-modules-load[208]: Inserted module 'br_netfilter' May 27 03:26:25.959876 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 27 03:26:25.962097 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 27 03:26:25.962657 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 27 03:26:25.964650 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 27 03:26:25.967112 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 27 03:26:25.972235 systemd-tmpfiles[225]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. May 27 03:26:25.978247 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 27 03:26:25.987575 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 27 03:26:25.989135 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 27 03:26:25.991527 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 27 03:26:25.992699 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 27 03:26:25.995061 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 27 03:26:26.003981 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 27 03:26:26.021964 dracut-cmdline[247]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=f6c186658a19d5a08471ef76df75f82494b37b46908f9237b2c3cf497da860c6 May 27 03:26:26.052022 systemd-resolved[246]: Positive Trust Anchors: May 27 03:26:26.053045 systemd-resolved[246]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 27 03:26:26.053115 systemd-resolved[246]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 27 03:26:26.062190 systemd-resolved[246]: Defaulting to hostname 'linux'. May 27 03:26:26.063590 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 27 03:26:26.064626 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 27 03:26:26.118980 kernel: SCSI subsystem initialized May 27 03:26:26.128984 kernel: Loading iSCSI transport class v2.0-870. May 27 03:26:26.140965 kernel: iscsi: registered transport (tcp) May 27 03:26:26.163037 kernel: iscsi: registered transport (qla4xxx) May 27 03:26:26.163128 kernel: QLogic iSCSI HBA Driver May 27 03:26:26.182297 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 27 03:26:26.198631 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 27 03:26:26.199521 systemd[1]: Reached target network-pre.target - Preparation for Network. May 27 03:26:26.246525 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 27 03:26:26.248709 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 27 03:26:26.303985 kernel: raid6: avx512x4 gen() 17831 MB/s May 27 03:26:26.321973 kernel: raid6: avx512x2 gen() 17872 MB/s May 27 03:26:26.339971 kernel: raid6: avx512x1 gen() 17802 MB/s May 27 03:26:26.357985 kernel: raid6: avx2x4 gen() 17446 MB/s May 27 03:26:26.375973 kernel: raid6: avx2x2 gen() 17518 MB/s May 27 03:26:26.394213 kernel: raid6: avx2x1 gen() 13656 MB/s May 27 03:26:26.394285 kernel: raid6: using algorithm avx512x2 gen() 17872 MB/s May 27 03:26:26.413228 kernel: raid6: .... xor() 24522 MB/s, rmw enabled May 27 03:26:26.413304 kernel: raid6: using avx512x2 recovery algorithm May 27 03:26:26.434975 kernel: xor: automatically using best checksumming function avx May 27 03:26:26.603977 kernel: Btrfs loaded, zoned=no, fsverity=no May 27 03:26:26.610465 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 27 03:26:26.612691 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 27 03:26:26.640622 systemd-udevd[456]: Using default interface naming scheme 'v255'. May 27 03:26:26.647467 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 27 03:26:26.651491 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 27 03:26:26.681676 dracut-pre-trigger[462]: rd.md=0: removing MD RAID activation May 27 03:26:26.690053 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input3 May 27 03:26:26.713558 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 27 03:26:26.715532 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 27 03:26:26.777371 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 27 03:26:26.781876 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 27 03:26:26.876776 kernel: nvme nvme0: pci function 0000:00:04.0 May 27 03:26:26.877091 kernel: ena 0000:00:05.0: ENA device version: 0.10 May 27 03:26:26.877300 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 May 27 03:26:26.877325 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 May 27 03:26:26.883591 kernel: cryptd: max_cpu_qlen set to 1000 May 27 03:26:26.883661 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. May 27 03:26:26.888961 kernel: nvme nvme0: 2/0/0 default/read/poll queues May 27 03:26:26.895977 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:cd:cc:fe:6c:13 May 27 03:26:26.902697 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 27 03:26:26.902758 kernel: GPT:9289727 != 16777215 May 27 03:26:26.902777 kernel: GPT:Alternate GPT header not at the end of the disk. May 27 03:26:26.905086 kernel: GPT:9289727 != 16777215 May 27 03:26:26.906374 kernel: GPT: Use GNU Parted to correct GPT errors. May 27 03:26:26.908749 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 27 03:26:26.915073 (udev-worker)[518]: Network interface NamePolicy= disabled on kernel command line. May 27 03:26:26.923967 kernel: AES CTR mode by8 optimization enabled May 27 03:26:26.923792 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 27 03:26:26.924922 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:26:26.927597 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:26:26.934113 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:26:26.946545 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 27 03:26:26.971965 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 27 03:26:26.972106 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:26:26.973927 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 27 03:26:26.978276 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:26:26.989955 kernel: nvme nvme0: using unchecked data buffer May 27 03:26:27.010371 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:26:27.089001 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. May 27 03:26:27.107958 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. May 27 03:26:27.108825 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 27 03:26:27.128620 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. May 27 03:26:27.137698 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. May 27 03:26:27.138385 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. May 27 03:26:27.139620 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 27 03:26:27.140825 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 27 03:26:27.141861 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 27 03:26:27.143401 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 27 03:26:27.144558 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 27 03:26:27.162080 disk-uuid[696]: Primary Header is updated. May 27 03:26:27.162080 disk-uuid[696]: Secondary Entries is updated. May 27 03:26:27.162080 disk-uuid[696]: Secondary Header is updated. May 27 03:26:27.166763 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 27 03:26:27.170968 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 27 03:26:27.178007 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 27 03:26:28.187184 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 27 03:26:28.187240 disk-uuid[700]: The operation has completed successfully. May 27 03:26:28.335795 systemd[1]: disk-uuid.service: Deactivated successfully. May 27 03:26:28.335957 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 27 03:26:28.364640 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 27 03:26:28.388651 sh[962]: Success May 27 03:26:28.410470 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 27 03:26:28.410571 kernel: device-mapper: uevent: version 1.0.3 May 27 03:26:28.411346 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev May 27 03:26:28.423993 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" May 27 03:26:28.543224 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 27 03:26:28.545966 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 27 03:26:28.558861 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 27 03:26:28.584977 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' May 27 03:26:28.587968 kernel: BTRFS: device fsid f0f66fe8-3990-49eb-980e-559a3dfd3522 devid 1 transid 40 /dev/mapper/usr (254:0) scanned by mount (986) May 27 03:26:28.591076 kernel: BTRFS info (device dm-0): first mount of filesystem f0f66fe8-3990-49eb-980e-559a3dfd3522 May 27 03:26:28.591129 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 27 03:26:28.593600 kernel: BTRFS info (device dm-0): using free-space-tree May 27 03:26:28.669680 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 27 03:26:28.670596 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. May 27 03:26:28.671137 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 27 03:26:28.671901 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 27 03:26:28.673819 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 27 03:26:28.709966 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 (259:5) scanned by mount (1019) May 27 03:26:28.718884 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:26:28.718985 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm May 27 03:26:28.719009 kernel: BTRFS info (device nvme0n1p6): using free-space-tree May 27 03:26:28.732054 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:26:28.733047 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 27 03:26:28.736202 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 27 03:26:28.778420 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 27 03:26:28.784134 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 27 03:26:28.828400 systemd-networkd[1155]: lo: Link UP May 27 03:26:28.828470 systemd-networkd[1155]: lo: Gained carrier May 27 03:26:28.830434 systemd-networkd[1155]: Enumeration completed May 27 03:26:28.830880 systemd-networkd[1155]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 27 03:26:28.830886 systemd-networkd[1155]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 27 03:26:28.831993 systemd[1]: Started systemd-networkd.service - Network Configuration. May 27 03:26:28.833401 systemd[1]: Reached target network.target - Network. May 27 03:26:28.834921 systemd-networkd[1155]: eth0: Link UP May 27 03:26:28.834926 systemd-networkd[1155]: eth0: Gained carrier May 27 03:26:28.834961 systemd-networkd[1155]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 27 03:26:28.848043 systemd-networkd[1155]: eth0: DHCPv4 address 172.31.17.111/20, gateway 172.31.16.1 acquired from 172.31.16.1 May 27 03:26:28.977540 ignition[1105]: Ignition 2.21.0 May 27 03:26:28.977557 ignition[1105]: Stage: fetch-offline May 27 03:26:28.977735 ignition[1105]: no configs at "/usr/lib/ignition/base.d" May 27 03:26:28.977743 ignition[1105]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 27 03:26:28.978074 ignition[1105]: Ignition finished successfully May 27 03:26:28.979579 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 27 03:26:28.981664 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 27 03:26:29.002511 ignition[1165]: Ignition 2.21.0 May 27 03:26:29.002529 ignition[1165]: Stage: fetch May 27 03:26:29.002921 ignition[1165]: no configs at "/usr/lib/ignition/base.d" May 27 03:26:29.002934 ignition[1165]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 27 03:26:29.003075 ignition[1165]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 27 03:26:29.062203 ignition[1165]: PUT result: OK May 27 03:26:29.066174 ignition[1165]: parsed url from cmdline: "" May 27 03:26:29.066186 ignition[1165]: no config URL provided May 27 03:26:29.066196 ignition[1165]: reading system config file "/usr/lib/ignition/user.ign" May 27 03:26:29.066211 ignition[1165]: no config at "/usr/lib/ignition/user.ign" May 27 03:26:29.066234 ignition[1165]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 27 03:26:29.071219 ignition[1165]: PUT result: OK May 27 03:26:29.071331 ignition[1165]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 May 27 03:26:29.076095 ignition[1165]: GET result: OK May 27 03:26:29.076210 ignition[1165]: parsing config with SHA512: 63b2dda9c1aab20939e03ad0372d1e2501bf7cd30aea270d3ace33f2e34a81fccdde573c023785a953dff81b7e167ce94b25049f148bbf664a8814c034122a06 May 27 03:26:29.081586 unknown[1165]: fetched base config from "system" May 27 03:26:29.081705 unknown[1165]: fetched base config from "system" May 27 03:26:29.082091 ignition[1165]: fetch: fetch complete May 27 03:26:29.081713 unknown[1165]: fetched user config from "aws" May 27 03:26:29.082108 ignition[1165]: fetch: fetch passed May 27 03:26:29.082155 ignition[1165]: Ignition finished successfully May 27 03:26:29.084638 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 27 03:26:29.086191 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 27 03:26:29.118016 ignition[1171]: Ignition 2.21.0 May 27 03:26:29.118030 ignition[1171]: Stage: kargs May 27 03:26:29.118436 ignition[1171]: no configs at "/usr/lib/ignition/base.d" May 27 03:26:29.118449 ignition[1171]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 27 03:26:29.118587 ignition[1171]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 27 03:26:29.119564 ignition[1171]: PUT result: OK May 27 03:26:29.123128 ignition[1171]: kargs: kargs passed May 27 03:26:29.123208 ignition[1171]: Ignition finished successfully May 27 03:26:29.125403 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 27 03:26:29.126862 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 27 03:26:29.152643 ignition[1177]: Ignition 2.21.0 May 27 03:26:29.152660 ignition[1177]: Stage: disks May 27 03:26:29.153058 ignition[1177]: no configs at "/usr/lib/ignition/base.d" May 27 03:26:29.153071 ignition[1177]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 27 03:26:29.153195 ignition[1177]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 27 03:26:29.154223 ignition[1177]: PUT result: OK May 27 03:26:29.157813 ignition[1177]: disks: disks passed May 27 03:26:29.158383 ignition[1177]: Ignition finished successfully May 27 03:26:29.160078 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 27 03:26:29.160763 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 27 03:26:29.161210 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 27 03:26:29.161748 systemd[1]: Reached target local-fs.target - Local File Systems. May 27 03:26:29.162390 systemd[1]: Reached target sysinit.target - System Initialization. May 27 03:26:29.162954 systemd[1]: Reached target basic.target - Basic System. May 27 03:26:29.164718 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 27 03:26:29.209960 systemd-fsck[1185]: ROOT: clean, 15/553520 files, 52789/553472 blocks May 27 03:26:29.212749 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 27 03:26:29.214407 systemd[1]: Mounting sysroot.mount - /sysroot... May 27 03:26:29.378966 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 18301365-b380-45d7-9677-e42472a122bc r/w with ordered data mode. Quota mode: none. May 27 03:26:29.379785 systemd[1]: Mounted sysroot.mount - /sysroot. May 27 03:26:29.380793 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 27 03:26:29.382835 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 27 03:26:29.384672 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 27 03:26:29.386657 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. May 27 03:26:29.386708 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 27 03:26:29.386734 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 27 03:26:29.393174 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 27 03:26:29.395180 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 27 03:26:29.410976 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 (259:5) scanned by mount (1204) May 27 03:26:29.414112 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:26:29.414184 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm May 27 03:26:29.417149 kernel: BTRFS info (device nvme0n1p6): using free-space-tree May 27 03:26:29.424632 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 27 03:26:29.602433 initrd-setup-root[1228]: cut: /sysroot/etc/passwd: No such file or directory May 27 03:26:29.608581 initrd-setup-root[1235]: cut: /sysroot/etc/group: No such file or directory May 27 03:26:29.615229 initrd-setup-root[1242]: cut: /sysroot/etc/shadow: No such file or directory May 27 03:26:29.620133 initrd-setup-root[1249]: cut: /sysroot/etc/gshadow: No such file or directory May 27 03:26:29.799443 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 27 03:26:29.801565 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 27 03:26:29.805047 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 27 03:26:29.820960 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 27 03:26:29.823244 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:26:29.851526 ignition[1317]: INFO : Ignition 2.21.0 May 27 03:26:29.851526 ignition[1317]: INFO : Stage: mount May 27 03:26:29.851526 ignition[1317]: INFO : no configs at "/usr/lib/ignition/base.d" May 27 03:26:29.851526 ignition[1317]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 27 03:26:29.851526 ignition[1317]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 27 03:26:29.854616 ignition[1317]: INFO : PUT result: OK May 27 03:26:29.851669 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 27 03:26:29.855533 ignition[1317]: INFO : mount: mount passed May 27 03:26:29.855890 ignition[1317]: INFO : Ignition finished successfully May 27 03:26:29.856900 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 27 03:26:29.858813 systemd[1]: Starting ignition-files.service - Ignition (files)... May 27 03:26:29.874826 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 27 03:26:29.908973 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 (259:5) scanned by mount (1330) May 27 03:26:29.913225 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem fd7bb961-7a0f-4c90-a609-3bffeb956d05 May 27 03:26:29.913293 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm May 27 03:26:29.913309 kernel: BTRFS info (device nvme0n1p6): using free-space-tree May 27 03:26:29.922674 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 27 03:26:29.949887 ignition[1346]: INFO : Ignition 2.21.0 May 27 03:26:29.949887 ignition[1346]: INFO : Stage: files May 27 03:26:29.951022 ignition[1346]: INFO : no configs at "/usr/lib/ignition/base.d" May 27 03:26:29.951022 ignition[1346]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 27 03:26:29.951022 ignition[1346]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 27 03:26:29.951933 ignition[1346]: INFO : PUT result: OK May 27 03:26:29.954539 ignition[1346]: DEBUG : files: compiled without relabeling support, skipping May 27 03:26:29.956712 ignition[1346]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 27 03:26:29.956712 ignition[1346]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 27 03:26:29.960215 ignition[1346]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 27 03:26:29.961030 ignition[1346]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 27 03:26:29.961030 ignition[1346]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 27 03:26:29.960697 unknown[1346]: wrote ssh authorized keys file for user: core May 27 03:26:29.964662 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" May 27 03:26:29.965361 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 May 27 03:26:30.038931 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 27 03:26:30.319781 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" May 27 03:26:30.321089 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 27 03:26:30.321089 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 May 27 03:26:30.721140 systemd-networkd[1155]: eth0: Gained IPv6LL May 27 03:26:30.762291 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 27 03:26:30.874854 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 27 03:26:30.874854 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" May 27 03:26:30.876723 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" May 27 03:26:30.876723 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" May 27 03:26:30.876723 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" May 27 03:26:30.876723 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 27 03:26:30.876723 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 27 03:26:30.876723 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 27 03:26:30.876723 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 27 03:26:30.881977 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" May 27 03:26:30.881977 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 27 03:26:30.881977 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" May 27 03:26:30.884629 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" May 27 03:26:30.884629 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" May 27 03:26:30.886510 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 May 27 03:26:31.527337 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK May 27 03:26:31.988129 ignition[1346]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" May 27 03:26:31.988129 ignition[1346]: INFO : files: op(c): [started] processing unit "prepare-helm.service" May 27 03:26:31.990274 ignition[1346]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 27 03:26:31.995215 ignition[1346]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 27 03:26:31.995215 ignition[1346]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" May 27 03:26:31.995215 ignition[1346]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" May 27 03:26:31.997805 ignition[1346]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" May 27 03:26:31.997805 ignition[1346]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" May 27 03:26:31.997805 ignition[1346]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" May 27 03:26:31.997805 ignition[1346]: INFO : files: files passed May 27 03:26:31.997805 ignition[1346]: INFO : Ignition finished successfully May 27 03:26:31.997530 systemd[1]: Finished ignition-files.service - Ignition (files). May 27 03:26:32.001073 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 27 03:26:32.003134 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 27 03:26:32.011299 systemd[1]: ignition-quench.service: Deactivated successfully. May 27 03:26:32.011406 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 27 03:26:32.027403 initrd-setup-root-after-ignition[1377]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 27 03:26:32.027403 initrd-setup-root-after-ignition[1377]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 27 03:26:32.029819 initrd-setup-root-after-ignition[1381]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 27 03:26:32.031422 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 27 03:26:32.032019 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 27 03:26:32.033782 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 27 03:26:32.094673 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 27 03:26:32.094810 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 27 03:26:32.096108 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 27 03:26:32.097452 systemd[1]: Reached target initrd.target - Initrd Default Target. May 27 03:26:32.098362 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 27 03:26:32.099543 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 27 03:26:32.124988 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 27 03:26:32.126976 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 27 03:26:32.146346 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 27 03:26:32.147074 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 27 03:26:32.148179 systemd[1]: Stopped target timers.target - Timer Units. May 27 03:26:32.149265 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 27 03:26:32.149511 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 27 03:26:32.150664 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 27 03:26:32.151626 systemd[1]: Stopped target basic.target - Basic System. May 27 03:26:32.152577 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 27 03:26:32.153370 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 27 03:26:32.154166 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 27 03:26:32.154974 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. May 27 03:26:32.155763 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 27 03:26:32.156652 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 27 03:26:32.157524 systemd[1]: Stopped target sysinit.target - System Initialization. May 27 03:26:32.158718 systemd[1]: Stopped target local-fs.target - Local File Systems. May 27 03:26:32.159505 systemd[1]: Stopped target swap.target - Swaps. May 27 03:26:32.160251 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 27 03:26:32.160566 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 27 03:26:32.161652 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 27 03:26:32.162474 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 27 03:26:32.163128 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 27 03:26:32.163463 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 27 03:26:32.163882 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 27 03:26:32.164076 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 27 03:26:32.165689 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 27 03:26:32.165960 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 27 03:26:32.166671 systemd[1]: ignition-files.service: Deactivated successfully. May 27 03:26:32.166871 systemd[1]: Stopped ignition-files.service - Ignition (files). May 27 03:26:32.168809 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 27 03:26:32.172185 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 27 03:26:32.173491 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 27 03:26:32.174329 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 27 03:26:32.175830 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 27 03:26:32.176587 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 27 03:26:32.182607 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 27 03:26:32.185852 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 27 03:26:32.204744 ignition[1401]: INFO : Ignition 2.21.0 May 27 03:26:32.206116 ignition[1401]: INFO : Stage: umount May 27 03:26:32.207571 ignition[1401]: INFO : no configs at "/usr/lib/ignition/base.d" May 27 03:26:32.207571 ignition[1401]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 27 03:26:32.207571 ignition[1401]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 27 03:26:32.210334 ignition[1401]: INFO : PUT result: OK May 27 03:26:32.213147 ignition[1401]: INFO : umount: umount passed May 27 03:26:32.213147 ignition[1401]: INFO : Ignition finished successfully May 27 03:26:32.214459 systemd[1]: ignition-mount.service: Deactivated successfully. May 27 03:26:32.214573 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 27 03:26:32.215749 systemd[1]: ignition-disks.service: Deactivated successfully. May 27 03:26:32.215842 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 27 03:26:32.216359 systemd[1]: ignition-kargs.service: Deactivated successfully. May 27 03:26:32.216545 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 27 03:26:32.217055 systemd[1]: ignition-fetch.service: Deactivated successfully. May 27 03:26:32.217125 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 27 03:26:32.217474 systemd[1]: Stopped target network.target - Network. May 27 03:26:32.217765 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 27 03:26:32.217809 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 27 03:26:32.218435 systemd[1]: Stopped target paths.target - Path Units. May 27 03:26:32.219141 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 27 03:26:32.226006 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 27 03:26:32.226431 systemd[1]: Stopped target slices.target - Slice Units. May 27 03:26:32.226746 systemd[1]: Stopped target sockets.target - Socket Units. May 27 03:26:32.227137 systemd[1]: iscsid.socket: Deactivated successfully. May 27 03:26:32.227181 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 27 03:26:32.227919 systemd[1]: iscsiuio.socket: Deactivated successfully. May 27 03:26:32.227972 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 27 03:26:32.228300 systemd[1]: ignition-setup.service: Deactivated successfully. May 27 03:26:32.228357 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 27 03:26:32.228844 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 27 03:26:32.228885 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 27 03:26:32.229530 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 27 03:26:32.230146 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 27 03:26:32.231996 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 27 03:26:32.235921 systemd[1]: systemd-resolved.service: Deactivated successfully. May 27 03:26:32.236082 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 27 03:26:32.239234 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 27 03:26:32.239510 systemd[1]: systemd-networkd.service: Deactivated successfully. May 27 03:26:32.239628 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 27 03:26:32.241454 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 27 03:26:32.242014 systemd[1]: Stopped target network-pre.target - Preparation for Network. May 27 03:26:32.242837 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 27 03:26:32.242886 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 27 03:26:32.244340 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 27 03:26:32.244791 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 27 03:26:32.244843 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 27 03:26:32.245269 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 27 03:26:32.245309 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 27 03:26:32.245698 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 27 03:26:32.245734 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 27 03:26:32.248535 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 27 03:26:32.248587 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 27 03:26:32.249589 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 27 03:26:32.253159 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 27 03:26:32.253236 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 27 03:26:32.258319 systemd[1]: systemd-udevd.service: Deactivated successfully. May 27 03:26:32.258475 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 27 03:26:32.259400 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 27 03:26:32.259468 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 27 03:26:32.261404 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 27 03:26:32.261441 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 27 03:26:32.261776 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 27 03:26:32.261825 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 27 03:26:32.262388 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 27 03:26:32.262432 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 27 03:26:32.263681 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 27 03:26:32.263731 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 27 03:26:32.267539 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 27 03:26:32.268268 systemd[1]: systemd-network-generator.service: Deactivated successfully. May 27 03:26:32.268329 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. May 27 03:26:32.269977 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 27 03:26:32.270026 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 27 03:26:32.271504 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. May 27 03:26:32.271548 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 27 03:26:32.273056 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 27 03:26:32.273101 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 27 03:26:32.273500 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 27 03:26:32.273540 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:26:32.275705 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. May 27 03:26:32.275766 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev\x2dearly.service.mount: Deactivated successfully. May 27 03:26:32.275799 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 27 03:26:32.275834 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 27 03:26:32.276538 systemd[1]: network-cleanup.service: Deactivated successfully. May 27 03:26:32.278087 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 27 03:26:32.284499 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 27 03:26:32.284613 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 27 03:26:32.347590 systemd[1]: sysroot-boot.service: Deactivated successfully. May 27 03:26:32.347703 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 27 03:26:32.349049 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 27 03:26:32.349605 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 27 03:26:32.349668 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 27 03:26:32.351402 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 27 03:26:32.378303 systemd[1]: Switching root. May 27 03:26:32.409541 systemd-journald[207]: Journal stopped May 27 03:26:34.076868 systemd-journald[207]: Received SIGTERM from PID 1 (systemd). May 27 03:26:34.088101 kernel: SELinux: policy capability network_peer_controls=1 May 27 03:26:34.088154 kernel: SELinux: policy capability open_perms=1 May 27 03:26:34.088177 kernel: SELinux: policy capability extended_socket_class=1 May 27 03:26:34.088204 kernel: SELinux: policy capability always_check_network=0 May 27 03:26:34.088225 kernel: SELinux: policy capability cgroup_seclabel=1 May 27 03:26:34.088253 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 27 03:26:34.088276 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 27 03:26:34.088304 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 27 03:26:34.088325 kernel: SELinux: policy capability userspace_initial_context=0 May 27 03:26:34.088348 kernel: audit: type=1403 audit(1748316392.716:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 27 03:26:34.088384 systemd[1]: Successfully loaded SELinux policy in 63.922ms. May 27 03:26:34.088419 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 10.332ms. May 27 03:26:34.088450 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 27 03:26:34.088475 systemd[1]: Detected virtualization amazon. May 27 03:26:34.088499 systemd[1]: Detected architecture x86-64. May 27 03:26:34.088521 systemd[1]: Detected first boot. May 27 03:26:34.088546 systemd[1]: Initializing machine ID from VM UUID. May 27 03:26:34.088570 zram_generator::config[1445]: No configuration found. May 27 03:26:34.088595 kernel: Guest personality initialized and is inactive May 27 03:26:34.088618 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 27 03:26:34.088644 kernel: Initialized host personality May 27 03:26:34.088667 kernel: NET: Registered PF_VSOCK protocol family May 27 03:26:34.088692 systemd[1]: Populated /etc with preset unit settings. May 27 03:26:34.088718 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 27 03:26:34.088742 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 27 03:26:34.088771 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 27 03:26:34.088797 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 27 03:26:34.088820 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 27 03:26:34.088846 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 27 03:26:34.088874 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 27 03:26:34.088896 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 27 03:26:34.088921 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 27 03:26:34.091014 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 27 03:26:34.091053 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 27 03:26:34.091077 systemd[1]: Created slice user.slice - User and Session Slice. May 27 03:26:34.091103 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 27 03:26:34.091128 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 27 03:26:34.091151 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 27 03:26:34.091183 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 27 03:26:34.091215 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 27 03:26:34.091238 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 27 03:26:34.091262 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 27 03:26:34.091288 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 27 03:26:34.091311 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 27 03:26:34.091337 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 27 03:26:34.091362 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 27 03:26:34.091390 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 27 03:26:34.091415 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 27 03:26:34.091438 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 27 03:26:34.091462 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 27 03:26:34.091489 systemd[1]: Reached target slices.target - Slice Units. May 27 03:26:34.091514 systemd[1]: Reached target swap.target - Swaps. May 27 03:26:34.091539 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 27 03:26:34.091564 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 27 03:26:34.091587 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 27 03:26:34.091616 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 27 03:26:34.091640 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 27 03:26:34.091665 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 27 03:26:34.091690 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 27 03:26:34.091714 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 27 03:26:34.091737 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 27 03:26:34.091761 systemd[1]: Mounting media.mount - External Media Directory... May 27 03:26:34.091787 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:34.091811 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 27 03:26:34.091839 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 27 03:26:34.091863 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 27 03:26:34.091887 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 27 03:26:34.091912 systemd[1]: Reached target machines.target - Containers. May 27 03:26:34.100280 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 27 03:26:34.100341 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:26:34.100367 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 27 03:26:34.100405 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 27 03:26:34.100441 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 27 03:26:34.100466 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 27 03:26:34.100491 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 27 03:26:34.100515 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 27 03:26:34.100539 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 27 03:26:34.100564 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 27 03:26:34.100588 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 27 03:26:34.100612 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 27 03:26:34.100639 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 27 03:26:34.100664 systemd[1]: Stopped systemd-fsck-usr.service. May 27 03:26:34.100688 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:26:34.100710 systemd[1]: Starting systemd-journald.service - Journal Service... May 27 03:26:34.100734 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 27 03:26:34.103995 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 27 03:26:34.104043 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 27 03:26:34.104070 kernel: loop: module loaded May 27 03:26:34.104095 kernel: fuse: init (API version 7.41) May 27 03:26:34.104119 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 27 03:26:34.104144 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 27 03:26:34.104174 systemd[1]: verity-setup.service: Deactivated successfully. May 27 03:26:34.104198 systemd[1]: Stopped verity-setup.service. May 27 03:26:34.104223 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:34.104248 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 27 03:26:34.104273 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 27 03:26:34.104297 systemd[1]: Mounted media.mount - External Media Directory. May 27 03:26:34.104321 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 27 03:26:34.104345 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 27 03:26:34.104383 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 27 03:26:34.104410 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 27 03:26:34.104433 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 27 03:26:34.104458 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 27 03:26:34.104477 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 27 03:26:34.104497 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 27 03:26:34.112297 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 27 03:26:34.112324 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 27 03:26:34.112347 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 27 03:26:34.112386 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 27 03:26:34.112407 systemd[1]: modprobe@loop.service: Deactivated successfully. May 27 03:26:34.112428 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 27 03:26:34.112449 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 27 03:26:34.112471 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 27 03:26:34.112493 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 27 03:26:34.112513 systemd[1]: Reached target network-pre.target - Preparation for Network. May 27 03:26:34.112535 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 27 03:26:34.112556 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 27 03:26:34.112581 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 27 03:26:34.112602 systemd[1]: Reached target local-fs.target - Local File Systems. May 27 03:26:34.112623 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 27 03:26:34.112645 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 27 03:26:34.112667 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:26:34.112691 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 27 03:26:34.112713 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 27 03:26:34.112734 kernel: ACPI: bus type drm_connector registered May 27 03:26:34.112756 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 27 03:26:34.112777 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 27 03:26:34.112798 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 27 03:26:34.112820 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 27 03:26:34.112884 systemd-journald[1528]: Collecting audit messages is disabled. May 27 03:26:34.112927 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 27 03:26:34.116899 systemd[1]: modprobe@drm.service: Deactivated successfully. May 27 03:26:34.117431 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 27 03:26:34.117474 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 27 03:26:34.117497 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 27 03:26:34.117519 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 27 03:26:34.117541 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 27 03:26:34.117572 systemd-journald[1528]: Journal started May 27 03:26:34.117620 systemd-journald[1528]: Runtime Journal (/run/log/journal/ec2fb34c2679c45c75604177d1a7e24a) is 4.8M, max 38.4M, 33.6M free. May 27 03:26:33.599914 systemd[1]: Queued start job for default target multi-user.target. May 27 03:26:34.124833 systemd[1]: Started systemd-journald.service - Journal Service. May 27 03:26:33.613832 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. May 27 03:26:33.614389 systemd[1]: systemd-journald.service: Deactivated successfully. May 27 03:26:34.124139 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 27 03:26:34.126034 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 27 03:26:34.162097 kernel: loop0: detected capacity change from 0 to 146240 May 27 03:26:34.171686 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 27 03:26:34.175694 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 27 03:26:34.183193 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 27 03:26:34.197614 systemd-tmpfiles[1561]: ACLs are not supported, ignoring. May 27 03:26:34.197641 systemd-tmpfiles[1561]: ACLs are not supported, ignoring. May 27 03:26:34.211596 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 27 03:26:34.217857 systemd-journald[1528]: Time spent on flushing to /var/log/journal/ec2fb34c2679c45c75604177d1a7e24a is 69.760ms for 1032 entries. May 27 03:26:34.217857 systemd-journald[1528]: System Journal (/var/log/journal/ec2fb34c2679c45c75604177d1a7e24a) is 8M, max 195.6M, 187.6M free. May 27 03:26:34.301366 systemd-journald[1528]: Received client request to flush runtime journal. May 27 03:26:34.301430 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 27 03:26:34.301454 kernel: loop1: detected capacity change from 0 to 72352 May 27 03:26:34.226171 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 27 03:26:34.239128 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 27 03:26:34.303834 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 27 03:26:34.327877 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 27 03:26:34.343339 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 27 03:26:34.348161 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 27 03:26:34.353983 kernel: loop2: detected capacity change from 0 to 113872 May 27 03:26:34.378877 systemd-tmpfiles[1600]: ACLs are not supported, ignoring. May 27 03:26:34.378898 systemd-tmpfiles[1600]: ACLs are not supported, ignoring. May 27 03:26:34.384244 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 27 03:26:34.463973 kernel: loop3: detected capacity change from 0 to 229808 May 27 03:26:34.610970 kernel: loop4: detected capacity change from 0 to 146240 May 27 03:26:34.616504 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 27 03:26:34.640049 kernel: loop5: detected capacity change from 0 to 72352 May 27 03:26:34.662984 kernel: loop6: detected capacity change from 0 to 113872 May 27 03:26:34.685256 kernel: loop7: detected capacity change from 0 to 229808 May 27 03:26:34.718547 (sd-merge)[1605]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. May 27 03:26:34.719492 (sd-merge)[1605]: Merged extensions into '/usr'. May 27 03:26:34.731702 systemd[1]: Reload requested from client PID 1560 ('systemd-sysext') (unit systemd-sysext.service)... May 27 03:26:34.731853 systemd[1]: Reloading... May 27 03:26:34.846063 zram_generator::config[1631]: No configuration found. May 27 03:26:35.013427 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:26:35.182749 systemd[1]: Reloading finished in 450 ms. May 27 03:26:35.207538 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 27 03:26:35.218800 systemd[1]: Starting ensure-sysext.service... May 27 03:26:35.221595 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 27 03:26:35.261038 systemd[1]: Reload requested from client PID 1682 ('systemctl') (unit ensure-sysext.service)... May 27 03:26:35.261193 systemd[1]: Reloading... May 27 03:26:35.299723 systemd-tmpfiles[1683]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. May 27 03:26:35.299761 systemd-tmpfiles[1683]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. May 27 03:26:35.300191 systemd-tmpfiles[1683]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 27 03:26:35.300536 systemd-tmpfiles[1683]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 27 03:26:35.301917 systemd-tmpfiles[1683]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 27 03:26:35.302551 systemd-tmpfiles[1683]: ACLs are not supported, ignoring. May 27 03:26:35.302667 systemd-tmpfiles[1683]: ACLs are not supported, ignoring. May 27 03:26:35.321008 systemd-tmpfiles[1683]: Detected autofs mount point /boot during canonicalization of boot. May 27 03:26:35.321028 systemd-tmpfiles[1683]: Skipping /boot May 27 03:26:35.350975 zram_generator::config[1708]: No configuration found. May 27 03:26:35.368303 systemd-tmpfiles[1683]: Detected autofs mount point /boot during canonicalization of boot. May 27 03:26:35.368322 systemd-tmpfiles[1683]: Skipping /boot May 27 03:26:35.540249 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:26:35.572789 ldconfig[1553]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 27 03:26:35.637598 systemd[1]: Reloading finished in 375 ms. May 27 03:26:35.664452 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 27 03:26:35.665284 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 27 03:26:35.671059 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 27 03:26:35.679693 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 27 03:26:35.682423 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 27 03:26:35.690096 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 27 03:26:35.693001 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 27 03:26:35.696419 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 27 03:26:35.701098 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 27 03:26:35.705767 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:35.706436 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:26:35.710200 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 27 03:26:35.714441 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 27 03:26:35.720451 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 27 03:26:35.721270 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:26:35.721396 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:26:35.721488 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:35.727212 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 27 03:26:35.728873 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:35.729491 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:26:35.729667 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:26:35.729748 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:26:35.729833 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:35.735594 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:35.735854 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 27 03:26:35.738065 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 27 03:26:35.739135 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 27 03:26:35.739251 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 27 03:26:35.739414 systemd[1]: Reached target time-set.target - System Time Set. May 27 03:26:35.740092 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 27 03:26:35.749986 systemd[1]: Finished ensure-sysext.service. May 27 03:26:35.761058 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 27 03:26:35.764239 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 27 03:26:35.764429 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 27 03:26:35.776757 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 27 03:26:35.776976 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 27 03:26:35.777718 systemd[1]: modprobe@loop.service: Deactivated successfully. May 27 03:26:35.777865 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 27 03:26:35.778524 systemd-udevd[1770]: Using default interface naming scheme 'v255'. May 27 03:26:35.778931 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 27 03:26:35.780070 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 27 03:26:35.784373 systemd[1]: modprobe@drm.service: Deactivated successfully. May 27 03:26:35.784932 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 27 03:26:35.797215 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 27 03:26:35.799068 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 27 03:26:35.831367 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 27 03:26:35.832617 augenrules[1804]: No rules May 27 03:26:35.833533 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 27 03:26:35.835057 systemd[1]: audit-rules.service: Deactivated successfully. May 27 03:26:35.835262 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 27 03:26:35.850535 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 27 03:26:35.852612 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 27 03:26:35.856339 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 27 03:26:35.857995 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 27 03:26:35.942121 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 27 03:26:35.949539 (udev-worker)[1840]: Network interface NamePolicy= disabled on kernel command line. May 27 03:26:36.028445 systemd-networkd[1825]: lo: Link UP May 27 03:26:36.028755 systemd-networkd[1825]: lo: Gained carrier May 27 03:26:36.031474 systemd-networkd[1825]: Enumeration completed May 27 03:26:36.031590 systemd[1]: Started systemd-networkd.service - Network Configuration. May 27 03:26:36.032621 systemd-networkd[1825]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 27 03:26:36.035062 systemd-networkd[1825]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 27 03:26:36.035539 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 27 03:26:36.039004 systemd-networkd[1825]: eth0: Link UP May 27 03:26:36.039041 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 27 03:26:36.039302 systemd-networkd[1825]: eth0: Gained carrier May 27 03:26:36.039372 systemd-networkd[1825]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 27 03:26:36.046092 systemd-resolved[1769]: Positive Trust Anchors: May 27 03:26:36.046399 systemd-resolved[1769]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 27 03:26:36.046479 systemd-resolved[1769]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 27 03:26:36.048027 systemd-networkd[1825]: eth0: DHCPv4 address 172.31.17.111/20, gateway 172.31.16.1 acquired from 172.31.16.1 May 27 03:26:36.055592 systemd-resolved[1769]: Defaulting to hostname 'linux'. May 27 03:26:36.059491 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 27 03:26:36.060250 systemd[1]: Reached target network.target - Network. May 27 03:26:36.061041 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 27 03:26:36.062007 systemd[1]: Reached target sysinit.target - System Initialization. May 27 03:26:36.062444 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 27 03:26:36.063250 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 27 03:26:36.063948 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. May 27 03:26:36.065152 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 27 03:26:36.065589 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 27 03:26:36.066189 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 27 03:26:36.066572 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 27 03:26:36.066609 systemd[1]: Reached target paths.target - Path Units. May 27 03:26:36.067226 systemd[1]: Reached target timers.target - Timer Units. May 27 03:26:36.068887 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 27 03:26:36.071536 systemd[1]: Starting docker.socket - Docker Socket for the API... May 27 03:26:36.078222 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 27 03:26:36.079197 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 27 03:26:36.081038 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 27 03:26:36.090866 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 27 03:26:36.092370 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 27 03:26:36.093708 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 27 03:26:36.094974 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input4 May 27 03:26:36.097000 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 27 03:26:36.101130 systemd[1]: Reached target sockets.target - Socket Units. May 27 03:26:36.101561 systemd[1]: Reached target basic.target - Basic System. May 27 03:26:36.103066 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 27 03:26:36.103096 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 27 03:26:36.105095 systemd[1]: Starting containerd.service - containerd container runtime... May 27 03:26:36.107855 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 27 03:26:36.114370 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 27 03:26:36.114966 kernel: mousedev: PS/2 mouse device common for all mice May 27 03:26:36.116434 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 27 03:26:36.119322 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 27 03:26:36.126068 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 27 03:26:36.126486 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 27 03:26:36.131077 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... May 27 03:26:36.138004 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 27 03:26:36.143146 systemd[1]: Started ntpd.service - Network Time Service. May 27 03:26:36.151117 kernel: ACPI: button: Power Button [PWRF] May 27 03:26:36.155960 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input5 May 27 03:26:36.156844 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 27 03:26:36.161130 systemd[1]: Starting setup-oem.service - Setup OEM... May 27 03:26:36.164593 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 27 03:26:36.178898 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 27 03:26:36.190727 systemd[1]: Starting systemd-logind.service - User Login Management... May 27 03:26:36.192262 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 27 03:26:36.194251 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 27 03:26:36.195768 jq[1865]: false May 27 03:26:36.196155 systemd[1]: Starting update-engine.service - Update Engine... May 27 03:26:36.198884 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 27 03:26:36.220299 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 27 03:26:36.221136 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 27 03:26:36.222210 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 27 03:26:36.227117 kernel: ACPI: button: Sleep Button [SLPF] May 27 03:26:36.232335 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 27 03:26:36.232580 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 27 03:26:36.247006 google_oslogin_nss_cache[1867]: oslogin_cache_refresh[1867]: Refreshing passwd entry cache May 27 03:26:36.243256 oslogin_cache_refresh[1867]: Refreshing passwd entry cache May 27 03:26:36.247420 jq[1877]: true May 27 03:26:36.271469 (ntainerd)[1895]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 27 03:26:36.272142 oslogin_cache_refresh[1867]: Failure getting users, quitting May 27 03:26:36.272668 google_oslogin_nss_cache[1867]: oslogin_cache_refresh[1867]: Failure getting users, quitting May 27 03:26:36.272668 google_oslogin_nss_cache[1867]: oslogin_cache_refresh[1867]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 27 03:26:36.272161 oslogin_cache_refresh[1867]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 27 03:26:36.273192 google_oslogin_nss_cache[1867]: oslogin_cache_refresh[1867]: Refreshing group entry cache May 27 03:26:36.272854 oslogin_cache_refresh[1867]: Refreshing group entry cache May 27 03:26:36.276130 google_oslogin_nss_cache[1867]: oslogin_cache_refresh[1867]: Failure getting groups, quitting May 27 03:26:36.276130 google_oslogin_nss_cache[1867]: oslogin_cache_refresh[1867]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 27 03:26:36.274683 oslogin_cache_refresh[1867]: Failure getting groups, quitting May 27 03:26:36.274694 oslogin_cache_refresh[1867]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 27 03:26:36.279449 systemd[1]: google-oslogin-cache.service: Deactivated successfully. May 27 03:26:36.280588 jq[1893]: true May 27 03:26:36.279661 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. May 27 03:26:36.295964 update_engine[1876]: I20250527 03:26:36.288139 1876 main.cc:92] Flatcar Update Engine starting May 27 03:26:36.352703 tar[1884]: linux-amd64/LICENSE May 27 03:26:36.352703 tar[1884]: linux-amd64/helm May 27 03:26:36.388730 dbus-daemon[1863]: [system] SELinux support is enabled May 27 03:26:36.396566 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 27 03:26:36.403714 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 27 03:26:36.403760 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 27 03:26:36.405481 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 27 03:26:36.405518 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 27 03:26:36.410113 dbus-daemon[1863]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.1' (uid=244 pid=1825 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") May 27 03:26:36.418610 dbus-daemon[1863]: [system] Successfully activated service 'org.freedesktop.systemd1' May 27 03:26:36.423600 systemd[1]: Finished setup-oem.service - Setup OEM. May 27 03:26:36.437165 extend-filesystems[1866]: Found loop4 May 27 03:26:36.437165 extend-filesystems[1866]: Found loop5 May 27 03:26:36.437165 extend-filesystems[1866]: Found loop6 May 27 03:26:36.437165 extend-filesystems[1866]: Found loop7 May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1 May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1p1 May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1p2 May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1p3 May 27 03:26:36.437165 extend-filesystems[1866]: Found usr May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1p4 May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1p6 May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1p7 May 27 03:26:36.437165 extend-filesystems[1866]: Found nvme0n1p9 May 27 03:26:36.494048 bash[1950]: Updated "/home/core/.ssh/authorized_keys" May 27 03:26:36.439573 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... May 27 03:26:36.494283 update_engine[1876]: I20250527 03:26:36.447135 1876 update_check_scheduler.cc:74] Next update check in 6m16s May 27 03:26:36.503535 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr May 27 03:26:36.445080 systemd[1]: extend-filesystems.service: Deactivated successfully. May 27 03:26:36.445583 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 27 03:26:36.459525 systemd[1]: Started update-engine.service - Update Engine. May 27 03:26:36.464148 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 27 03:26:36.466518 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 27 03:26:36.475855 systemd[1]: Starting sshkeys.service... May 27 03:26:36.512593 systemd[1]: motdgen.service: Deactivated successfully. May 27 03:26:36.512931 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 27 03:26:36.528460 coreos-metadata[1862]: May 27 03:26:36.528 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 May 27 03:26:36.547143 coreos-metadata[1862]: May 27 03:26:36.533 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 May 27 03:26:36.549267 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. May 27 03:26:36.559970 coreos-metadata[1862]: May 27 03:26:36.551 INFO Fetch successful May 27 03:26:36.559970 coreos-metadata[1862]: May 27 03:26:36.552 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 May 27 03:26:36.559970 coreos-metadata[1862]: May 27 03:26:36.553 INFO Fetch successful May 27 03:26:36.559970 coreos-metadata[1862]: May 27 03:26:36.553 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 May 27 03:26:36.555069 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... May 27 03:26:36.578032 coreos-metadata[1862]: May 27 03:26:36.577 INFO Fetch successful May 27 03:26:36.578032 coreos-metadata[1862]: May 27 03:26:36.577 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 May 27 03:26:36.597452 coreos-metadata[1862]: May 27 03:26:36.597 INFO Fetch successful May 27 03:26:36.597452 coreos-metadata[1862]: May 27 03:26:36.597 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 May 27 03:26:36.607691 coreos-metadata[1862]: May 27 03:26:36.606 INFO Fetch failed with 404: resource not found May 27 03:26:36.607691 coreos-metadata[1862]: May 27 03:26:36.606 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 May 27 03:26:36.608320 coreos-metadata[1862]: May 27 03:26:36.608 INFO Fetch successful May 27 03:26:36.608320 coreos-metadata[1862]: May 27 03:26:36.608 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 May 27 03:26:36.609230 coreos-metadata[1862]: May 27 03:26:36.609 INFO Fetch successful May 27 03:26:36.609230 coreos-metadata[1862]: May 27 03:26:36.609 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 May 27 03:26:36.609804 coreos-metadata[1862]: May 27 03:26:36.609 INFO Fetch successful May 27 03:26:36.609888 coreos-metadata[1862]: May 27 03:26:36.609 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 May 27 03:26:36.630631 coreos-metadata[1862]: May 27 03:26:36.611 INFO Fetch successful May 27 03:26:36.630631 coreos-metadata[1862]: May 27 03:26:36.611 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 May 27 03:26:36.630631 coreos-metadata[1862]: May 27 03:26:36.628 INFO Fetch successful May 27 03:26:36.757784 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 27 03:26:36.759033 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 27 03:26:36.800664 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:26:36.932154 coreos-metadata[1976]: May 27 03:26:36.931 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 May 27 03:26:36.936167 coreos-metadata[1976]: May 27 03:26:36.934 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 May 27 03:26:36.940019 coreos-metadata[1976]: May 27 03:26:36.937 INFO Fetch successful May 27 03:26:36.940019 coreos-metadata[1976]: May 27 03:26:36.937 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 May 27 03:26:36.941314 coreos-metadata[1976]: May 27 03:26:36.941 INFO Fetch successful May 27 03:26:36.941572 locksmithd[1960]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 27 03:26:36.941832 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 27 03:26:36.942165 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:26:36.955204 unknown[1976]: wrote ssh authorized keys file for user: core May 27 03:26:36.972248 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. May 27 03:26:36.995031 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 27 03:26:37.002108 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 27 03:26:37.040817 update-ssh-keys[2044]: Updated "/home/core/.ssh/authorized_keys" May 27 03:26:37.043690 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). May 27 03:26:37.049013 systemd[1]: Finished sshkeys.service. May 27 03:26:37.124840 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 27 03:26:37.161254 containerd[1895]: time="2025-05-27T03:26:37Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 27 03:26:37.164683 ntpd[1869]: ntpd 4.2.8p17@1.4004-o Tue May 27 00:37:40 UTC 2025 (1): Starting May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: ntpd 4.2.8p17@1.4004-o Tue May 27 00:37:40 UTC 2025 (1): Starting May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: ---------------------------------------------------- May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: ntp-4 is maintained by Network Time Foundation, May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Inc. (NTF), a non-profit 501(c)(3) public-benefit May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: corporation. Support and training for ntp-4 are May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: available at https://www.nwtime.org/support May 27 03:26:37.165563 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: ---------------------------------------------------- May 27 03:26:37.164716 ntpd[1869]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp May 27 03:26:37.164727 ntpd[1869]: ---------------------------------------------------- May 27 03:26:37.164736 ntpd[1869]: ntp-4 is maintained by Network Time Foundation, May 27 03:26:37.164746 ntpd[1869]: Inc. (NTF), a non-profit 501(c)(3) public-benefit May 27 03:26:37.164755 ntpd[1869]: corporation. Support and training for ntp-4 are May 27 03:26:37.164764 ntpd[1869]: available at https://www.nwtime.org/support May 27 03:26:37.164773 ntpd[1869]: ---------------------------------------------------- May 27 03:26:37.174311 ntpd[1869]: proto: precision = 0.097 usec (-23) May 27 03:26:37.174662 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: proto: precision = 0.097 usec (-23) May 27 03:26:37.176532 containerd[1895]: time="2025-05-27T03:26:37.174828905Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 May 27 03:26:37.181994 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: basedate set to 2025-05-15 May 27 03:26:37.181994 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: gps base set to 2025-05-18 (week 2367) May 27 03:26:37.181204 ntpd[1869]: basedate set to 2025-05-15 May 27 03:26:37.181224 ntpd[1869]: gps base set to 2025-05-18 (week 2367) May 27 03:26:37.185056 systemd-networkd[1825]: eth0: Gained IPv6LL May 27 03:26:37.189773 sshd_keygen[1911]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 27 03:26:37.190021 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 27 03:26:37.190773 systemd[1]: Reached target network-online.target - Network is Online. May 27 03:26:37.191598 ntpd[1869]: Listen and drop on 0 v6wildcard [::]:123 May 27 03:26:37.191714 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Listen and drop on 0 v6wildcard [::]:123 May 27 03:26:37.191802 ntpd[1869]: Listen and drop on 1 v4wildcard 0.0.0.0:123 May 27 03:26:37.193395 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Listen and drop on 1 v4wildcard 0.0.0.0:123 May 27 03:26:37.193395 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Listen normally on 2 lo 127.0.0.1:123 May 27 03:26:37.193395 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Listen normally on 3 eth0 172.31.17.111:123 May 27 03:26:37.193395 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Listen normally on 4 lo [::1]:123 May 27 03:26:37.193395 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Listen normally on 5 eth0 [fe80::4cd:ccff:fefe:6c13%2]:123 May 27 03:26:37.193395 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: Listening on routing socket on fd #22 for interface updates May 27 03:26:37.193147 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. May 27 03:26:37.192196 ntpd[1869]: Listen normally on 2 lo 127.0.0.1:123 May 27 03:26:37.192238 ntpd[1869]: Listen normally on 3 eth0 172.31.17.111:123 May 27 03:26:37.192283 ntpd[1869]: Listen normally on 4 lo [::1]:123 May 27 03:26:37.192330 ntpd[1869]: Listen normally on 5 eth0 [fe80::4cd:ccff:fefe:6c13%2]:123 May 27 03:26:37.192379 ntpd[1869]: Listening on routing socket on fd #22 for interface updates May 27 03:26:37.196493 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:26:37.208091 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 27 03:26:37.210732 ntpd[1869]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 27 03:26:37.214542 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 27 03:26:37.214542 ntpd[1869]: 27 May 03:26:37 ntpd[1869]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 27 03:26:37.210769 ntpd[1869]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 27 03:26:37.237490 containerd[1895]: time="2025-05-27T03:26:37.237226571Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="12.668µs" May 27 03:26:37.237490 containerd[1895]: time="2025-05-27T03:26:37.237270425Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 27 03:26:37.237490 containerd[1895]: time="2025-05-27T03:26:37.237298577Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 27 03:26:37.238477 containerd[1895]: time="2025-05-27T03:26:37.238438223Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 27 03:26:37.238567 containerd[1895]: time="2025-05-27T03:26:37.238483444Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 27 03:26:37.238567 containerd[1895]: time="2025-05-27T03:26:37.238518571Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 27 03:26:37.238638 containerd[1895]: time="2025-05-27T03:26:37.238593986Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 27 03:26:37.238638 containerd[1895]: time="2025-05-27T03:26:37.238614170Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 27 03:26:37.241683 containerd[1895]: time="2025-05-27T03:26:37.238877516Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 27 03:26:37.241683 containerd[1895]: time="2025-05-27T03:26:37.238909859Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 27 03:26:37.241683 containerd[1895]: time="2025-05-27T03:26:37.238926200Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 27 03:26:37.241969 containerd[1895]: time="2025-05-27T03:26:37.241899791Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 27 03:26:37.242546 containerd[1895]: time="2025-05-27T03:26:37.242148208Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 27 03:26:37.242546 containerd[1895]: time="2025-05-27T03:26:37.242441117Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 27 03:26:37.242546 containerd[1895]: time="2025-05-27T03:26:37.242483784Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 27 03:26:37.242546 containerd[1895]: time="2025-05-27T03:26:37.242500725Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 27 03:26:37.242724 containerd[1895]: time="2025-05-27T03:26:37.242564314Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 27 03:26:37.244007 containerd[1895]: time="2025-05-27T03:26:37.243061648Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 27 03:26:37.244007 containerd[1895]: time="2025-05-27T03:26:37.243154275Z" level=info msg="metadata content store policy set" policy=shared May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251591125Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251690915Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251712334Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251776776Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251797353Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251817706Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251837878Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251856799Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251882794Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.251898844Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.252424744Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 27 03:26:37.253704 containerd[1895]: time="2025-05-27T03:26:37.252503132Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 27 03:26:37.255324 containerd[1895]: time="2025-05-27T03:26:37.255072028Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 27 03:26:37.255324 containerd[1895]: time="2025-05-27T03:26:37.255148757Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 27 03:26:37.255324 containerd[1895]: time="2025-05-27T03:26:37.255180952Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 27 03:26:37.255324 containerd[1895]: time="2025-05-27T03:26:37.255304473Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 27 03:26:37.255543 containerd[1895]: time="2025-05-27T03:26:37.255333321Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.255355947Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.256216649Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.256259945Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.256303712Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.256330220Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.256350261Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.256860908Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 27 03:26:37.258968 containerd[1895]: time="2025-05-27T03:26:37.256992135Z" level=info msg="Start snapshots syncer" May 27 03:26:37.259351 containerd[1895]: time="2025-05-27T03:26:37.259055788Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 27 03:26:37.266508 containerd[1895]: time="2025-05-27T03:26:37.262819538Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 27 03:26:37.266508 containerd[1895]: time="2025-05-27T03:26:37.265268445Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 27 03:26:37.267025 containerd[1895]: time="2025-05-27T03:26:37.266967040Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.267162052Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270463851Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270495243Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270520397Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270548969Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270568439Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270592275Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270641647Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270665553Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270691666Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270741933Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270773462Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 27 03:26:37.271958 containerd[1895]: time="2025-05-27T03:26:37.270794745Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 27 03:26:37.272558 containerd[1895]: time="2025-05-27T03:26:37.270816920Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 27 03:26:37.272558 containerd[1895]: time="2025-05-27T03:26:37.270831510Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 27 03:26:37.272558 containerd[1895]: time="2025-05-27T03:26:37.270863625Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 27 03:26:37.272558 containerd[1895]: time="2025-05-27T03:26:37.270886010Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 27 03:26:37.272558 containerd[1895]: time="2025-05-27T03:26:37.270913026Z" level=info msg="runtime interface created" May 27 03:26:37.272558 containerd[1895]: time="2025-05-27T03:26:37.270926702Z" level=info msg="created NRI interface" May 27 03:26:37.276575 containerd[1895]: time="2025-05-27T03:26:37.271739973Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 27 03:26:37.276575 containerd[1895]: time="2025-05-27T03:26:37.273923129Z" level=info msg="Connect containerd service" May 27 03:26:37.276575 containerd[1895]: time="2025-05-27T03:26:37.274012715Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 27 03:26:37.276575 containerd[1895]: time="2025-05-27T03:26:37.275244472Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 27 03:26:37.342260 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 27 03:26:37.349308 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 27 03:26:37.373381 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 27 03:26:37.379879 systemd[1]: Starting issuegen.service - Generate /run/issue... May 27 03:26:37.437329 systemd-logind[1875]: Watching system buttons on /dev/input/event2 (Power Button) May 27 03:26:37.437363 systemd-logind[1875]: Watching system buttons on /dev/input/event3 (Sleep Button) May 27 03:26:37.437386 systemd-logind[1875]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 27 03:26:37.442628 systemd-logind[1875]: New seat seat0. May 27 03:26:37.445638 systemd[1]: Started systemd-logind.service - User Login Management. May 27 03:26:37.457797 amazon-ssm-agent[2060]: Initializing new seelog logger May 27 03:26:37.458177 amazon-ssm-agent[2060]: New Seelog Logger Creation Complete May 27 03:26:37.458177 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.458177 amazon-ssm-agent[2060]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.461133 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 processing appconfig overrides May 27 03:26:37.462120 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.462120 amazon-ssm-agent[2060]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.462120 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 processing appconfig overrides May 27 03:26:37.462120 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.462120 amazon-ssm-agent[2060]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.462120 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 processing appconfig overrides May 27 03:26:37.462381 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.4604 INFO Proxy environment variables: May 27 03:26:37.467352 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.467352 amazon-ssm-agent[2060]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:37.467352 amazon-ssm-agent[2060]: 2025/05/27 03:26:37 processing appconfig overrides May 27 03:26:37.471225 systemd[1]: issuegen.service: Deactivated successfully. May 27 03:26:37.472003 systemd[1]: Finished issuegen.service - Generate /run/issue. May 27 03:26:37.480765 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 27 03:26:37.569765 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.4612 INFO https_proxy: May 27 03:26:37.576349 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 27 03:26:37.582363 systemd[1]: Started getty@tty1.service - Getty on tty1. May 27 03:26:37.589362 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 27 03:26:37.590285 systemd[1]: Reached target getty.target - Login Prompts. May 27 03:26:37.673525 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.4612 INFO http_proxy: May 27 03:26:37.770141 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.4612 INFO no_proxy: May 27 03:26:37.777892 systemd[1]: Started systemd-hostnamed.service - Hostname Service. May 27 03:26:37.779991 dbus-daemon[1863]: [system] Successfully activated service 'org.freedesktop.hostname1' May 27 03:26:37.793844 dbus-daemon[1863]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.7' (uid=0 pid=1955 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") May 27 03:26:37.800811 systemd[1]: Starting polkit.service - Authorization Manager... May 27 03:26:37.813663 containerd[1895]: time="2025-05-27T03:26:37.813614932Z" level=info msg="Start subscribing containerd event" May 27 03:26:37.815964 containerd[1895]: time="2025-05-27T03:26:37.814241688Z" level=info msg="Start recovering state" May 27 03:26:37.816094 containerd[1895]: time="2025-05-27T03:26:37.815982354Z" level=info msg="Start event monitor" May 27 03:26:37.816094 containerd[1895]: time="2025-05-27T03:26:37.816011414Z" level=info msg="Start cni network conf syncer for default" May 27 03:26:37.816094 containerd[1895]: time="2025-05-27T03:26:37.816081705Z" level=info msg="Start streaming server" May 27 03:26:37.816202 containerd[1895]: time="2025-05-27T03:26:37.816101241Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 27 03:26:37.816202 containerd[1895]: time="2025-05-27T03:26:37.816166652Z" level=info msg="runtime interface starting up..." May 27 03:26:37.818020 containerd[1895]: time="2025-05-27T03:26:37.816334918Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 27 03:26:37.818110 containerd[1895]: time="2025-05-27T03:26:37.817963501Z" level=info msg="starting plugins..." May 27 03:26:37.818110 containerd[1895]: time="2025-05-27T03:26:37.818094445Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 27 03:26:37.818204 containerd[1895]: time="2025-05-27T03:26:37.818190072Z" level=info msg=serving... address=/run/containerd/containerd.sock May 27 03:26:37.818509 systemd[1]: Started containerd.service - containerd container runtime. May 27 03:26:37.819189 containerd[1895]: time="2025-05-27T03:26:37.819163766Z" level=info msg="containerd successfully booted in 0.660614s" May 27 03:26:37.870868 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.4613 INFO Checking if agent identity type OnPrem can be assumed May 27 03:26:37.971024 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.4615 INFO Checking if agent identity type EC2 can be assumed May 27 03:26:38.055847 polkitd[2111]: Started polkitd version 126 May 27 03:26:38.068320 polkitd[2111]: Loading rules from directory /etc/polkit-1/rules.d May 27 03:26:38.068853 polkitd[2111]: Loading rules from directory /run/polkit-1/rules.d May 27 03:26:38.068904 polkitd[2111]: Error opening rules directory: Error opening directory “/run/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) May 27 03:26:38.070103 polkitd[2111]: Loading rules from directory /usr/local/share/polkit-1/rules.d May 27 03:26:38.070146 polkitd[2111]: Error opening rules directory: Error opening directory “/usr/local/share/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) May 27 03:26:38.070203 polkitd[2111]: Loading rules from directory /usr/share/polkit-1/rules.d May 27 03:26:38.071001 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7562 INFO Agent will take identity from EC2 May 27 03:26:38.072191 polkitd[2111]: Finished loading, compiling and executing 2 rules May 27 03:26:38.072991 systemd[1]: Started polkit.service - Authorization Manager. May 27 03:26:38.074244 dbus-daemon[1863]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' May 27 03:26:38.076216 polkitd[2111]: Acquired the name org.freedesktop.PolicyKit1 on the system bus May 27 03:26:38.102883 systemd-hostnamed[1955]: Hostname set to (transient) May 27 03:26:38.103026 systemd-resolved[1769]: System hostname changed to 'ip-172-31-17-111'. May 27 03:26:38.152315 tar[1884]: linux-amd64/README.md May 27 03:26:38.155696 amazon-ssm-agent[2060]: 2025/05/27 03:26:38 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:38.155696 amazon-ssm-agent[2060]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 27 03:26:38.155894 amazon-ssm-agent[2060]: 2025/05/27 03:26:38 processing appconfig overrides May 27 03:26:38.169426 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 27 03:26:38.170844 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7584 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.3.0.0 May 27 03:26:38.182294 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7585 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 May 27 03:26:38.182294 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7585 INFO [amazon-ssm-agent] Starting Core Agent May 27 03:26:38.182294 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7585 INFO [amazon-ssm-agent] Registrar detected. Attempting registration May 27 03:26:38.182294 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7585 INFO [Registrar] Starting registrar module May 27 03:26:38.182294 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7619 INFO [EC2Identity] Checking disk for registration info May 27 03:26:38.182294 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7620 INFO [EC2Identity] No registration info found for ec2 instance, attempting registration May 27 03:26:38.182294 amazon-ssm-agent[2060]: 2025-05-27 03:26:37.7620 INFO [EC2Identity] Generating registration keypair May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1183 INFO [EC2Identity] Checking write access before registering May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1187 INFO [EC2Identity] Registering EC2 instance with Systems Manager May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1554 INFO [EC2Identity] EC2 registration was successful. May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1554 INFO [amazon-ssm-agent] Registration attempted. Resuming core agent startup. May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1555 INFO [CredentialRefresher] credentialRefresher has started May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1555 INFO [CredentialRefresher] Starting credentials refresher loop May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1819 INFO EC2RoleProvider Successfully connected with instance profile role credentials May 27 03:26:38.182576 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1821 INFO [CredentialRefresher] Credentials ready May 27 03:26:38.269201 amazon-ssm-agent[2060]: 2025-05-27 03:26:38.1823 INFO [CredentialRefresher] Next credential rotation will be in 29.999993636316667 minutes May 27 03:26:39.195113 amazon-ssm-agent[2060]: 2025-05-27 03:26:39.1949 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process May 27 03:26:39.295553 amazon-ssm-agent[2060]: 2025-05-27 03:26:39.1974 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2126) started May 27 03:26:39.396606 amazon-ssm-agent[2060]: 2025-05-27 03:26:39.1975 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds May 27 03:26:39.788559 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:26:39.789880 systemd[1]: Reached target multi-user.target - Multi-User System. May 27 03:26:39.792096 systemd[1]: Startup finished in 2.771s (kernel) + 7.030s (initrd) + 7.137s (userspace) = 16.940s. May 27 03:26:39.802901 (kubelet)[2143]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 27 03:26:39.921771 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 27 03:26:39.923519 systemd[1]: Started sshd@0-172.31.17.111:22-139.178.68.195:58860.service - OpenSSH per-connection server daemon (139.178.68.195:58860). May 27 03:26:40.117166 sshd[2149]: Accepted publickey for core from 139.178.68.195 port 58860 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:26:40.119037 sshd-session[2149]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:26:40.125969 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 27 03:26:40.127088 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 27 03:26:40.138785 systemd-logind[1875]: New session 1 of user core. May 27 03:26:40.149726 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 27 03:26:40.153310 systemd[1]: Starting user@500.service - User Manager for UID 500... May 27 03:26:40.170563 (systemd)[2157]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 27 03:26:40.174160 systemd-logind[1875]: New session c1 of user core. May 27 03:26:40.347720 systemd[2157]: Queued start job for default target default.target. May 27 03:26:40.358125 systemd[2157]: Created slice app.slice - User Application Slice. May 27 03:26:40.358159 systemd[2157]: Reached target paths.target - Paths. May 27 03:26:40.358304 systemd[2157]: Reached target timers.target - Timers. May 27 03:26:40.359668 systemd[2157]: Starting dbus.socket - D-Bus User Message Bus Socket... May 27 03:26:40.391107 systemd[2157]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 27 03:26:40.391225 systemd[2157]: Reached target sockets.target - Sockets. May 27 03:26:40.391275 systemd[2157]: Reached target basic.target - Basic System. May 27 03:26:40.391316 systemd[2157]: Reached target default.target - Main User Target. May 27 03:26:40.391347 systemd[2157]: Startup finished in 207ms. May 27 03:26:40.392053 systemd[1]: Started user@500.service - User Manager for UID 500. May 27 03:26:40.397213 systemd[1]: Started session-1.scope - Session 1 of User core. May 27 03:26:40.543117 systemd[1]: Started sshd@1-172.31.17.111:22-139.178.68.195:58866.service - OpenSSH per-connection server daemon (139.178.68.195:58866). May 27 03:26:40.730865 sshd[2168]: Accepted publickey for core from 139.178.68.195 port 58866 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:26:40.733165 sshd-session[2168]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:26:40.740681 systemd-logind[1875]: New session 2 of user core. May 27 03:26:40.745451 systemd[1]: Started session-2.scope - Session 2 of User core. May 27 03:26:40.866169 sshd[2171]: Connection closed by 139.178.68.195 port 58866 May 27 03:26:40.866712 sshd-session[2168]: pam_unix(sshd:session): session closed for user core May 27 03:26:40.870766 systemd-logind[1875]: Session 2 logged out. Waiting for processes to exit. May 27 03:26:40.871580 systemd[1]: sshd@1-172.31.17.111:22-139.178.68.195:58866.service: Deactivated successfully. May 27 03:26:40.874575 systemd[1]: session-2.scope: Deactivated successfully. May 27 03:26:40.877109 systemd-logind[1875]: Removed session 2. May 27 03:26:40.898591 systemd[1]: Started sshd@2-172.31.17.111:22-139.178.68.195:58876.service - OpenSSH per-connection server daemon (139.178.68.195:58876). May 27 03:26:40.929770 kubelet[2143]: E0527 03:26:40.929731 2143 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 27 03:26:40.932647 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 27 03:26:40.932793 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 27 03:26:40.933262 systemd[1]: kubelet.service: Consumed 1.110s CPU time, 266.4M memory peak. May 27 03:26:41.069129 sshd[2177]: Accepted publickey for core from 139.178.68.195 port 58876 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:26:41.070749 sshd-session[2177]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:26:41.076847 systemd-logind[1875]: New session 3 of user core. May 27 03:26:41.082220 systemd[1]: Started session-3.scope - Session 3 of User core. May 27 03:26:41.211235 sshd[2180]: Connection closed by 139.178.68.195 port 58876 May 27 03:26:41.211757 sshd-session[2177]: pam_unix(sshd:session): session closed for user core May 27 03:26:41.215651 systemd[1]: sshd@2-172.31.17.111:22-139.178.68.195:58876.service: Deactivated successfully. May 27 03:26:41.218338 systemd[1]: session-3.scope: Deactivated successfully. May 27 03:26:41.219298 systemd-logind[1875]: Session 3 logged out. Waiting for processes to exit. May 27 03:26:41.221773 systemd-logind[1875]: Removed session 3. May 27 03:26:41.243032 systemd[1]: Started sshd@3-172.31.17.111:22-139.178.68.195:58884.service - OpenSSH per-connection server daemon (139.178.68.195:58884). May 27 03:26:41.412173 sshd[2186]: Accepted publickey for core from 139.178.68.195 port 58884 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:26:41.413986 sshd-session[2186]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:26:41.419341 systemd-logind[1875]: New session 4 of user core. May 27 03:26:41.433190 systemd[1]: Started session-4.scope - Session 4 of User core. May 27 03:26:41.552754 sshd[2188]: Connection closed by 139.178.68.195 port 58884 May 27 03:26:41.553334 sshd-session[2186]: pam_unix(sshd:session): session closed for user core May 27 03:26:41.557581 systemd[1]: sshd@3-172.31.17.111:22-139.178.68.195:58884.service: Deactivated successfully. May 27 03:26:41.559791 systemd[1]: session-4.scope: Deactivated successfully. May 27 03:26:41.560825 systemd-logind[1875]: Session 4 logged out. Waiting for processes to exit. May 27 03:26:41.562666 systemd-logind[1875]: Removed session 4. May 27 03:26:41.589974 systemd[1]: Started sshd@4-172.31.17.111:22-139.178.68.195:58892.service - OpenSSH per-connection server daemon (139.178.68.195:58892). May 27 03:26:41.774395 sshd[2194]: Accepted publickey for core from 139.178.68.195 port 58892 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:26:41.777731 sshd-session[2194]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:26:41.788206 systemd-logind[1875]: New session 5 of user core. May 27 03:26:41.795612 systemd[1]: Started session-5.scope - Session 5 of User core. May 27 03:26:41.930532 sudo[2197]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 27 03:26:41.930908 sudo[2197]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:26:41.943678 sudo[2197]: pam_unix(sudo:session): session closed for user root May 27 03:26:41.966914 sshd[2196]: Connection closed by 139.178.68.195 port 58892 May 27 03:26:41.967661 sshd-session[2194]: pam_unix(sshd:session): session closed for user core May 27 03:26:41.972258 systemd[1]: sshd@4-172.31.17.111:22-139.178.68.195:58892.service: Deactivated successfully. May 27 03:26:41.977880 systemd[1]: session-5.scope: Deactivated successfully. May 27 03:26:41.984020 systemd-logind[1875]: Session 5 logged out. Waiting for processes to exit. May 27 03:26:41.987732 systemd-logind[1875]: Removed session 5. May 27 03:26:41.999161 systemd[1]: Started sshd@5-172.31.17.111:22-139.178.68.195:58906.service - OpenSSH per-connection server daemon (139.178.68.195:58906). May 27 03:26:42.166139 sshd[2203]: Accepted publickey for core from 139.178.68.195 port 58906 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:26:42.167320 sshd-session[2203]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:26:42.173977 systemd-logind[1875]: New session 6 of user core. May 27 03:26:42.179185 systemd[1]: Started session-6.scope - Session 6 of User core. May 27 03:26:42.288853 sudo[2207]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 27 03:26:42.289247 sudo[2207]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:26:42.305785 sudo[2207]: pam_unix(sudo:session): session closed for user root May 27 03:26:42.311858 sudo[2206]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 27 03:26:42.312171 sudo[2206]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:26:42.322493 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 27 03:26:42.368240 augenrules[2229]: No rules May 27 03:26:42.369916 systemd[1]: audit-rules.service: Deactivated successfully. May 27 03:26:42.370299 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 27 03:26:42.372265 sudo[2206]: pam_unix(sudo:session): session closed for user root May 27 03:26:42.395012 sshd[2205]: Connection closed by 139.178.68.195 port 58906 May 27 03:26:42.395562 sshd-session[2203]: pam_unix(sshd:session): session closed for user core May 27 03:26:42.399479 systemd[1]: sshd@5-172.31.17.111:22-139.178.68.195:58906.service: Deactivated successfully. May 27 03:26:42.401314 systemd[1]: session-6.scope: Deactivated successfully. May 27 03:26:42.402100 systemd-logind[1875]: Session 6 logged out. Waiting for processes to exit. May 27 03:26:42.403788 systemd-logind[1875]: Removed session 6. May 27 03:26:42.425966 systemd[1]: Started sshd@6-172.31.17.111:22-139.178.68.195:58918.service - OpenSSH per-connection server daemon (139.178.68.195:58918). May 27 03:26:42.595071 sshd[2238]: Accepted publickey for core from 139.178.68.195 port 58918 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:26:42.596629 sshd-session[2238]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:26:42.601974 systemd-logind[1875]: New session 7 of user core. May 27 03:26:42.611177 systemd[1]: Started session-7.scope - Session 7 of User core. May 27 03:26:42.706509 sudo[2241]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 27 03:26:42.706880 sudo[2241]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 27 03:26:43.303181 systemd[1]: Starting docker.service - Docker Application Container Engine... May 27 03:26:43.313479 (dockerd)[2260]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 27 03:26:43.751820 dockerd[2260]: time="2025-05-27T03:26:43.751499913Z" level=info msg="Starting up" May 27 03:26:43.757451 dockerd[2260]: time="2025-05-27T03:26:43.757114455Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 27 03:26:43.886426 dockerd[2260]: time="2025-05-27T03:26:43.886207141Z" level=info msg="Loading containers: start." May 27 03:26:43.899252 kernel: Initializing XFRM netlink socket May 27 03:26:44.111583 (udev-worker)[2281]: Network interface NamePolicy= disabled on kernel command line. May 27 03:26:44.156173 systemd-networkd[1825]: docker0: Link UP May 27 03:26:44.162191 dockerd[2260]: time="2025-05-27T03:26:44.162146326Z" level=info msg="Loading containers: done." May 27 03:26:44.645110 systemd-resolved[1769]: Clock change detected. Flushing caches. May 27 03:26:44.660396 dockerd[2260]: time="2025-05-27T03:26:44.660345468Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 27 03:26:44.660577 dockerd[2260]: time="2025-05-27T03:26:44.660430578Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 May 27 03:26:44.660577 dockerd[2260]: time="2025-05-27T03:26:44.660535386Z" level=info msg="Initializing buildkit" May 27 03:26:44.689474 dockerd[2260]: time="2025-05-27T03:26:44.689429387Z" level=info msg="Completed buildkit initialization" May 27 03:26:44.696699 dockerd[2260]: time="2025-05-27T03:26:44.696648933Z" level=info msg="Daemon has completed initialization" May 27 03:26:44.696699 dockerd[2260]: time="2025-05-27T03:26:44.696712732Z" level=info msg="API listen on /run/docker.sock" May 27 03:26:44.696893 systemd[1]: Started docker.service - Docker Application Container Engine. May 27 03:26:45.781029 containerd[1895]: time="2025-05-27T03:26:45.780986298Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.1\"" May 27 03:26:46.345712 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3372482589.mount: Deactivated successfully. May 27 03:26:47.716950 containerd[1895]: time="2025-05-27T03:26:47.716897023Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:47.719056 containerd[1895]: time="2025-05-27T03:26:47.719010076Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.1: active requests=0, bytes read=30075403" May 27 03:26:47.720423 containerd[1895]: time="2025-05-27T03:26:47.720373723Z" level=info msg="ImageCreate event name:\"sha256:c6ab243b29f82a6ce269a5342bfd9ea3d0d4ef0f2bb7e98c6ac0bde1aeafab66\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:47.724158 containerd[1895]: time="2025-05-27T03:26:47.724102477Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:d8ae2fb01c39aa1c7add84f3d54425cf081c24c11e3946830292a8cfa4293548\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:47.725060 containerd[1895]: time="2025-05-27T03:26:47.724872458Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.1\" with image id \"sha256:c6ab243b29f82a6ce269a5342bfd9ea3d0d4ef0f2bb7e98c6ac0bde1aeafab66\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.1\", repo digest \"registry.k8s.io/kube-apiserver@sha256:d8ae2fb01c39aa1c7add84f3d54425cf081c24c11e3946830292a8cfa4293548\", size \"30072203\" in 1.943849389s" May 27 03:26:47.725060 containerd[1895]: time="2025-05-27T03:26:47.724909352Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.1\" returns image reference \"sha256:c6ab243b29f82a6ce269a5342bfd9ea3d0d4ef0f2bb7e98c6ac0bde1aeafab66\"" May 27 03:26:47.725691 containerd[1895]: time="2025-05-27T03:26:47.725662508Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.1\"" May 27 03:26:49.326526 containerd[1895]: time="2025-05-27T03:26:49.326469590Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:49.327384 containerd[1895]: time="2025-05-27T03:26:49.327341687Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.1: active requests=0, bytes read=26011390" May 27 03:26:49.328435 containerd[1895]: time="2025-05-27T03:26:49.328389833Z" level=info msg="ImageCreate event name:\"sha256:ef43894fa110c389f7286f4d5a3ea176072c95280efeca60d6a79617cdbbf3e4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:49.331517 containerd[1895]: time="2025-05-27T03:26:49.331447446Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7c9bea694e3a3c01ed6a5ee02d55a6124cc08e0b2eec6caa33f2c396b8cbc3f8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:49.332772 containerd[1895]: time="2025-05-27T03:26:49.332593470Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.1\" with image id \"sha256:ef43894fa110c389f7286f4d5a3ea176072c95280efeca60d6a79617cdbbf3e4\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.1\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7c9bea694e3a3c01ed6a5ee02d55a6124cc08e0b2eec6caa33f2c396b8cbc3f8\", size \"27638910\" in 1.606802754s" May 27 03:26:49.332772 containerd[1895]: time="2025-05-27T03:26:49.332635449Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.1\" returns image reference \"sha256:ef43894fa110c389f7286f4d5a3ea176072c95280efeca60d6a79617cdbbf3e4\"" May 27 03:26:49.333508 containerd[1895]: time="2025-05-27T03:26:49.333467517Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.1\"" May 27 03:26:50.841266 containerd[1895]: time="2025-05-27T03:26:50.841215997Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:50.842165 containerd[1895]: time="2025-05-27T03:26:50.842120170Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.1: active requests=0, bytes read=20148960" May 27 03:26:50.843794 containerd[1895]: time="2025-05-27T03:26:50.843610778Z" level=info msg="ImageCreate event name:\"sha256:398c985c0d950becc8dcdab5877a8a517ffeafca0792b3fe5f1acff218aeac49\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:50.849083 containerd[1895]: time="2025-05-27T03:26:50.848844670Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:395b7de7cdbdcc3c3a3db270844a3f71d757e2447a1e4db76b4cce46fba7fd55\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:50.850663 containerd[1895]: time="2025-05-27T03:26:50.850622077Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.1\" with image id \"sha256:398c985c0d950becc8dcdab5877a8a517ffeafca0792b3fe5f1acff218aeac49\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.1\", repo digest \"registry.k8s.io/kube-scheduler@sha256:395b7de7cdbdcc3c3a3db270844a3f71d757e2447a1e4db76b4cce46fba7fd55\", size \"21776498\" in 1.517122024s" May 27 03:26:50.850663 containerd[1895]: time="2025-05-27T03:26:50.850654863Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.1\" returns image reference \"sha256:398c985c0d950becc8dcdab5877a8a517ffeafca0792b3fe5f1acff218aeac49\"" May 27 03:26:50.851132 containerd[1895]: time="2025-05-27T03:26:50.851107222Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.1\"" May 27 03:26:51.490451 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 27 03:26:51.493651 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:26:51.765457 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:26:51.779028 (kubelet)[2535]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 27 03:26:51.853490 kubelet[2535]: E0527 03:26:51.853436 2535 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 27 03:26:51.861570 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 27 03:26:51.861775 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 27 03:26:51.863357 systemd[1]: kubelet.service: Consumed 228ms CPU time, 108.7M memory peak. May 27 03:26:52.040825 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount588970356.mount: Deactivated successfully. May 27 03:26:52.639932 containerd[1895]: time="2025-05-27T03:26:52.639881454Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:52.640843 containerd[1895]: time="2025-05-27T03:26:52.640804844Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.1: active requests=0, bytes read=31889075" May 27 03:26:52.641968 containerd[1895]: time="2025-05-27T03:26:52.641914807Z" level=info msg="ImageCreate event name:\"sha256:b79c189b052cdbe0e837d0caa6faf1d9fd696d8664fcc462f67d9ea51f26fef2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:52.644620 containerd[1895]: time="2025-05-27T03:26:52.644572936Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:7ddf379897139ae8ade8b33cb9373b70c632a4d5491da6e234f5d830e0a50807\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:52.645202 containerd[1895]: time="2025-05-27T03:26:52.644988161Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.1\" with image id \"sha256:b79c189b052cdbe0e837d0caa6faf1d9fd696d8664fcc462f67d9ea51f26fef2\", repo tag \"registry.k8s.io/kube-proxy:v1.33.1\", repo digest \"registry.k8s.io/kube-proxy@sha256:7ddf379897139ae8ade8b33cb9373b70c632a4d5491da6e234f5d830e0a50807\", size \"31888094\" in 1.793840333s" May 27 03:26:52.645202 containerd[1895]: time="2025-05-27T03:26:52.645018968Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.1\" returns image reference \"sha256:b79c189b052cdbe0e837d0caa6faf1d9fd696d8664fcc462f67d9ea51f26fef2\"" May 27 03:26:52.645625 containerd[1895]: time="2025-05-27T03:26:52.645603916Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" May 27 03:26:53.178978 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1384227278.mount: Deactivated successfully. May 27 03:26:54.351373 containerd[1895]: time="2025-05-27T03:26:54.351298497Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:54.352734 containerd[1895]: time="2025-05-27T03:26:54.352691475Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" May 27 03:26:54.353829 containerd[1895]: time="2025-05-27T03:26:54.353761831Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:54.357079 containerd[1895]: time="2025-05-27T03:26:54.357022224Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:54.358302 containerd[1895]: time="2025-05-27T03:26:54.358112765Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.712476626s" May 27 03:26:54.358302 containerd[1895]: time="2025-05-27T03:26:54.358154391Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" May 27 03:26:54.358942 containerd[1895]: time="2025-05-27T03:26:54.358897095Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 27 03:26:54.831510 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount801366998.mount: Deactivated successfully. May 27 03:26:54.838074 containerd[1895]: time="2025-05-27T03:26:54.838002563Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 27 03:26:54.838942 containerd[1895]: time="2025-05-27T03:26:54.838891693Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" May 27 03:26:54.840097 containerd[1895]: time="2025-05-27T03:26:54.840044429Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 27 03:26:54.842124 containerd[1895]: time="2025-05-27T03:26:54.842075225Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 27 03:26:54.843011 containerd[1895]: time="2025-05-27T03:26:54.842563395Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 483.573608ms" May 27 03:26:54.843011 containerd[1895]: time="2025-05-27T03:26:54.842592787Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" May 27 03:26:54.843144 containerd[1895]: time="2025-05-27T03:26:54.843099797Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" May 27 03:26:57.463794 containerd[1895]: time="2025-05-27T03:26:57.463555627Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:57.464747 containerd[1895]: time="2025-05-27T03:26:57.464709211Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58142739" May 27 03:26:57.465817 containerd[1895]: time="2025-05-27T03:26:57.465773676Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:57.468156 containerd[1895]: time="2025-05-27T03:26:57.468109778Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:26:57.469796 containerd[1895]: time="2025-05-27T03:26:57.469486593Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 2.626364814s" May 27 03:26:57.469796 containerd[1895]: time="2025-05-27T03:26:57.469518488Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" May 27 03:27:02.112933 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 27 03:27:02.117578 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:27:02.290142 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 27 03:27:02.290274 systemd[1]: kubelet.service: Failed with result 'signal'. May 27 03:27:02.291036 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:27:02.294852 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:27:02.332266 systemd[1]: Reload requested from client PID 2646 ('systemctl') (unit session-7.scope)... May 27 03:27:02.332288 systemd[1]: Reloading... May 27 03:27:02.485337 zram_generator::config[2691]: No configuration found. May 27 03:27:02.615865 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:27:02.763377 systemd[1]: Reloading finished in 430 ms. May 27 03:27:02.813733 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 27 03:27:02.813849 systemd[1]: kubelet.service: Failed with result 'signal'. May 27 03:27:02.814193 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:27:02.814253 systemd[1]: kubelet.service: Consumed 146ms CPU time, 98.3M memory peak. May 27 03:27:02.817754 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:27:03.547938 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:27:03.559950 (kubelet)[2755]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 27 03:27:03.610705 kubelet[2755]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:27:03.610705 kubelet[2755]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 27 03:27:03.610705 kubelet[2755]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:27:03.615359 kubelet[2755]: I0527 03:27:03.615269 2755 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 27 03:27:04.366372 kubelet[2755]: I0527 03:27:04.366294 2755 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" May 27 03:27:04.366372 kubelet[2755]: I0527 03:27:04.366355 2755 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 27 03:27:04.366631 kubelet[2755]: I0527 03:27:04.366612 2755 server.go:956] "Client rotation is on, will bootstrap in background" May 27 03:27:04.421648 kubelet[2755]: I0527 03:27:04.421533 2755 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 27 03:27:04.431847 kubelet[2755]: E0527 03:27:04.431789 2755 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.31.17.111:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" May 27 03:27:04.459833 kubelet[2755]: I0527 03:27:04.459367 2755 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 27 03:27:04.476347 kubelet[2755]: I0527 03:27:04.476114 2755 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 27 03:27:04.479756 kubelet[2755]: I0527 03:27:04.479639 2755 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 27 03:27:04.483869 kubelet[2755]: I0527 03:27:04.479704 2755 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-17-111","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 27 03:27:04.485276 kubelet[2755]: I0527 03:27:04.485239 2755 topology_manager.go:138] "Creating topology manager with none policy" May 27 03:27:04.485276 kubelet[2755]: I0527 03:27:04.485277 2755 container_manager_linux.go:303] "Creating device plugin manager" May 27 03:27:04.485653 kubelet[2755]: I0527 03:27:04.485439 2755 state_mem.go:36] "Initialized new in-memory state store" May 27 03:27:04.491450 kubelet[2755]: I0527 03:27:04.491378 2755 kubelet.go:480] "Attempting to sync node with API server" May 27 03:27:04.491450 kubelet[2755]: I0527 03:27:04.491442 2755 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" May 27 03:27:04.493274 kubelet[2755]: I0527 03:27:04.493238 2755 kubelet.go:386] "Adding apiserver pod source" May 27 03:27:04.496027 kubelet[2755]: I0527 03:27:04.495839 2755 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 27 03:27:04.507771 kubelet[2755]: E0527 03:27:04.507724 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.17.111:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-17-111&limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" May 27 03:27:04.507946 kubelet[2755]: E0527 03:27:04.507862 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.17.111:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" May 27 03:27:04.508394 kubelet[2755]: I0527 03:27:04.508369 2755 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 27 03:27:04.509180 kubelet[2755]: I0527 03:27:04.509157 2755 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" May 27 03:27:04.510911 kubelet[2755]: W0527 03:27:04.510890 2755 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 27 03:27:04.516366 kubelet[2755]: I0527 03:27:04.516309 2755 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 27 03:27:04.516507 kubelet[2755]: I0527 03:27:04.516420 2755 server.go:1289] "Started kubelet" May 27 03:27:04.518828 kubelet[2755]: I0527 03:27:04.518771 2755 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 May 27 03:27:04.525340 kubelet[2755]: I0527 03:27:04.524828 2755 server.go:317] "Adding debug handlers to kubelet server" May 27 03:27:04.525722 kubelet[2755]: I0527 03:27:04.525662 2755 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 27 03:27:04.526415 kubelet[2755]: I0527 03:27:04.526083 2755 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 27 03:27:04.536346 kubelet[2755]: E0527 03:27:04.526242 2755 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.17.111:6443/api/v1/namespaces/default/events\": dial tcp 172.31.17.111:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-17-111.184344866e476a33 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-17-111,UID:ip-172-31-17-111,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-17-111,},FirstTimestamp:2025-05-27 03:27:04.516364851 +0000 UTC m=+0.952026873,LastTimestamp:2025-05-27 03:27:04.516364851 +0000 UTC m=+0.952026873,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-17-111,}" May 27 03:27:04.537366 kubelet[2755]: I0527 03:27:04.537305 2755 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 27 03:27:04.545339 kubelet[2755]: I0527 03:27:04.544973 2755 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 27 03:27:04.550799 kubelet[2755]: E0527 03:27:04.550766 2755 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:04.550799 kubelet[2755]: I0527 03:27:04.550807 2755 volume_manager.go:297] "Starting Kubelet Volume Manager" May 27 03:27:04.551068 kubelet[2755]: I0527 03:27:04.551048 2755 desired_state_of_world_populator.go:150] "Desired state populator starts to run" May 27 03:27:04.551127 kubelet[2755]: I0527 03:27:04.551118 2755 reconciler.go:26] "Reconciler: start to sync state" May 27 03:27:04.551727 kubelet[2755]: E0527 03:27:04.551693 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.17.111:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" May 27 03:27:04.561105 kubelet[2755]: E0527 03:27:04.560742 2755 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.17.111:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-17-111?timeout=10s\": dial tcp 172.31.17.111:6443: connect: connection refused" interval="200ms" May 27 03:27:04.566204 kubelet[2755]: I0527 03:27:04.566178 2755 factory.go:223] Registration of the containerd container factory successfully May 27 03:27:04.566366 kubelet[2755]: I0527 03:27:04.566357 2755 factory.go:223] Registration of the systemd container factory successfully May 27 03:27:04.566500 kubelet[2755]: I0527 03:27:04.566484 2755 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 27 03:27:04.571947 kubelet[2755]: I0527 03:27:04.571902 2755 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" May 27 03:27:04.575338 kubelet[2755]: I0527 03:27:04.575252 2755 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" May 27 03:27:04.575338 kubelet[2755]: I0527 03:27:04.575277 2755 status_manager.go:230] "Starting to sync pod status with apiserver" May 27 03:27:04.575338 kubelet[2755]: I0527 03:27:04.575295 2755 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 27 03:27:04.575338 kubelet[2755]: I0527 03:27:04.575304 2755 kubelet.go:2436] "Starting kubelet main sync loop" May 27 03:27:04.575515 kubelet[2755]: E0527 03:27:04.575368 2755 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 27 03:27:04.595551 kubelet[2755]: E0527 03:27:04.594907 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.17.111:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" May 27 03:27:04.597009 kubelet[2755]: I0527 03:27:04.596981 2755 cpu_manager.go:221] "Starting CPU manager" policy="none" May 27 03:27:04.597009 kubelet[2755]: I0527 03:27:04.596999 2755 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 27 03:27:04.597009 kubelet[2755]: I0527 03:27:04.597016 2755 state_mem.go:36] "Initialized new in-memory state store" May 27 03:27:04.603662 kubelet[2755]: I0527 03:27:04.603523 2755 policy_none.go:49] "None policy: Start" May 27 03:27:04.603797 kubelet[2755]: I0527 03:27:04.603779 2755 memory_manager.go:186] "Starting memorymanager" policy="None" May 27 03:27:04.603827 kubelet[2755]: I0527 03:27:04.603812 2755 state_mem.go:35] "Initializing new in-memory state store" May 27 03:27:04.612087 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 27 03:27:04.633172 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 27 03:27:04.637413 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 27 03:27:04.646382 kubelet[2755]: E0527 03:27:04.646338 2755 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" May 27 03:27:04.646718 kubelet[2755]: I0527 03:27:04.646518 2755 eviction_manager.go:189] "Eviction manager: starting control loop" May 27 03:27:04.646718 kubelet[2755]: I0527 03:27:04.646528 2755 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 27 03:27:04.647019 kubelet[2755]: I0527 03:27:04.646999 2755 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 27 03:27:04.648611 kubelet[2755]: E0527 03:27:04.648581 2755 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 27 03:27:04.648825 kubelet[2755]: E0527 03:27:04.648814 2755 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-17-111\" not found" May 27 03:27:04.690985 systemd[1]: Created slice kubepods-burstable-pod0a9ff9cbfcbe58b31618c7e9f09efe42.slice - libcontainer container kubepods-burstable-pod0a9ff9cbfcbe58b31618c7e9f09efe42.slice. May 27 03:27:04.704544 kubelet[2755]: E0527 03:27:04.704338 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:04.707721 systemd[1]: Created slice kubepods-burstable-pod46f2cad2ef124d0ed9bf2d351de54cf0.slice - libcontainer container kubepods-burstable-pod46f2cad2ef124d0ed9bf2d351de54cf0.slice. May 27 03:27:04.720171 kubelet[2755]: E0527 03:27:04.719998 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:04.723075 systemd[1]: Created slice kubepods-burstable-podd5ffa022a545bafe8ff473dc5bf49c84.slice - libcontainer container kubepods-burstable-podd5ffa022a545bafe8ff473dc5bf49c84.slice. May 27 03:27:04.725404 kubelet[2755]: E0527 03:27:04.725352 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:04.749336 kubelet[2755]: I0527 03:27:04.749295 2755 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-17-111" May 27 03:27:04.749703 kubelet[2755]: E0527 03:27:04.749671 2755 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.17.111:6443/api/v1/nodes\": dial tcp 172.31.17.111:6443: connect: connection refused" node="ip-172-31-17-111" May 27 03:27:04.753019 kubelet[2755]: I0527 03:27:04.752984 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0a9ff9cbfcbe58b31618c7e9f09efe42-ca-certs\") pod \"kube-apiserver-ip-172-31-17-111\" (UID: \"0a9ff9cbfcbe58b31618c7e9f09efe42\") " pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:04.753019 kubelet[2755]: I0527 03:27:04.753027 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0a9ff9cbfcbe58b31618c7e9f09efe42-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-17-111\" (UID: \"0a9ff9cbfcbe58b31618c7e9f09efe42\") " pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:04.753019 kubelet[2755]: I0527 03:27:04.753058 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-ca-certs\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:04.753411 kubelet[2755]: I0527 03:27:04.753091 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:04.753411 kubelet[2755]: I0527 03:27:04.753117 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-k8s-certs\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:04.753411 kubelet[2755]: I0527 03:27:04.753143 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-kubeconfig\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:04.753411 kubelet[2755]: I0527 03:27:04.753172 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:04.753411 kubelet[2755]: I0527 03:27:04.753197 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0a9ff9cbfcbe58b31618c7e9f09efe42-k8s-certs\") pod \"kube-apiserver-ip-172-31-17-111\" (UID: \"0a9ff9cbfcbe58b31618c7e9f09efe42\") " pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:04.753530 kubelet[2755]: I0527 03:27:04.753223 2755 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d5ffa022a545bafe8ff473dc5bf49c84-kubeconfig\") pod \"kube-scheduler-ip-172-31-17-111\" (UID: \"d5ffa022a545bafe8ff473dc5bf49c84\") " pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:04.761825 kubelet[2755]: E0527 03:27:04.761785 2755 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.17.111:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-17-111?timeout=10s\": dial tcp 172.31.17.111:6443: connect: connection refused" interval="400ms" May 27 03:27:04.916015 kubelet[2755]: E0527 03:27:04.915832 2755 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.17.111:6443/api/v1/namespaces/default/events\": dial tcp 172.31.17.111:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-17-111.184344866e476a33 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-17-111,UID:ip-172-31-17-111,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-17-111,},FirstTimestamp:2025-05-27 03:27:04.516364851 +0000 UTC m=+0.952026873,LastTimestamp:2025-05-27 03:27:04.516364851 +0000 UTC m=+0.952026873,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-17-111,}" May 27 03:27:04.951809 kubelet[2755]: I0527 03:27:04.951764 2755 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-17-111" May 27 03:27:04.952138 kubelet[2755]: E0527 03:27:04.952104 2755 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.17.111:6443/api/v1/nodes\": dial tcp 172.31.17.111:6443: connect: connection refused" node="ip-172-31-17-111" May 27 03:27:05.006309 containerd[1895]: time="2025-05-27T03:27:05.006267958Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-17-111,Uid:0a9ff9cbfcbe58b31618c7e9f09efe42,Namespace:kube-system,Attempt:0,}" May 27 03:27:05.021633 containerd[1895]: time="2025-05-27T03:27:05.021562370Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-17-111,Uid:46f2cad2ef124d0ed9bf2d351de54cf0,Namespace:kube-system,Attempt:0,}" May 27 03:27:05.026778 containerd[1895]: time="2025-05-27T03:27:05.026730039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-17-111,Uid:d5ffa022a545bafe8ff473dc5bf49c84,Namespace:kube-system,Attempt:0,}" May 27 03:27:05.154530 containerd[1895]: time="2025-05-27T03:27:05.154455129Z" level=info msg="connecting to shim 6039a1b8b68cd09a566570ffb2f19b1a61b13d3f7555245d5b1bad95d089a4a8" address="unix:///run/containerd/s/3060c7d25f758a3d46df2da0568d523fc64a615ad467933730c9b6b7e480d00a" namespace=k8s.io protocol=ttrpc version=3 May 27 03:27:05.161875 containerd[1895]: time="2025-05-27T03:27:05.161827582Z" level=info msg="connecting to shim 26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba" address="unix:///run/containerd/s/3e530b6cfa15efd8d23f031e713f5ea1a754b2facde1d92a0676ffa47cc590f5" namespace=k8s.io protocol=ttrpc version=3 May 27 03:27:05.163140 kubelet[2755]: E0527 03:27:05.163084 2755 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.17.111:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-17-111?timeout=10s\": dial tcp 172.31.17.111:6443: connect: connection refused" interval="800ms" May 27 03:27:05.169533 containerd[1895]: time="2025-05-27T03:27:05.168529666Z" level=info msg="connecting to shim 9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805" address="unix:///run/containerd/s/80f20d15877b2529fc9ca266ce1ef8e61815738865c79d923fc688cf4d70633e" namespace=k8s.io protocol=ttrpc version=3 May 27 03:27:05.291677 systemd[1]: Started cri-containerd-26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba.scope - libcontainer container 26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba. May 27 03:27:05.294442 systemd[1]: Started cri-containerd-6039a1b8b68cd09a566570ffb2f19b1a61b13d3f7555245d5b1bad95d089a4a8.scope - libcontainer container 6039a1b8b68cd09a566570ffb2f19b1a61b13d3f7555245d5b1bad95d089a4a8. May 27 03:27:05.297146 systemd[1]: Started cri-containerd-9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805.scope - libcontainer container 9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805. May 27 03:27:05.355923 kubelet[2755]: I0527 03:27:05.355878 2755 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-17-111" May 27 03:27:05.357594 kubelet[2755]: E0527 03:27:05.357459 2755 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.17.111:6443/api/v1/nodes\": dial tcp 172.31.17.111:6443: connect: connection refused" node="ip-172-31-17-111" May 27 03:27:05.407617 containerd[1895]: time="2025-05-27T03:27:05.407492145Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-17-111,Uid:d5ffa022a545bafe8ff473dc5bf49c84,Namespace:kube-system,Attempt:0,} returns sandbox id \"26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba\"" May 27 03:27:05.414610 containerd[1895]: time="2025-05-27T03:27:05.414546986Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-17-111,Uid:0a9ff9cbfcbe58b31618c7e9f09efe42,Namespace:kube-system,Attempt:0,} returns sandbox id \"6039a1b8b68cd09a566570ffb2f19b1a61b13d3f7555245d5b1bad95d089a4a8\"" May 27 03:27:05.422548 containerd[1895]: time="2025-05-27T03:27:05.421867819Z" level=info msg="CreateContainer within sandbox \"26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 27 03:27:05.426901 containerd[1895]: time="2025-05-27T03:27:05.426859422Z" level=info msg="CreateContainer within sandbox \"6039a1b8b68cd09a566570ffb2f19b1a61b13d3f7555245d5b1bad95d089a4a8\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 27 03:27:05.442839 containerd[1895]: time="2025-05-27T03:27:05.442797468Z" level=info msg="Container 436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:05.444440 containerd[1895]: time="2025-05-27T03:27:05.444395590Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-17-111,Uid:46f2cad2ef124d0ed9bf2d351de54cf0,Namespace:kube-system,Attempt:0,} returns sandbox id \"9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805\"" May 27 03:27:05.450563 containerd[1895]: time="2025-05-27T03:27:05.450478769Z" level=info msg="Container f3e02d8557ab7149c5e39c31c6d15c77f543e963535f559aaa83cc4b5004dbef: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:05.451284 containerd[1895]: time="2025-05-27T03:27:05.450944371Z" level=info msg="CreateContainer within sandbox \"9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 27 03:27:05.472336 containerd[1895]: time="2025-05-27T03:27:05.472252866Z" level=info msg="CreateContainer within sandbox \"26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187\"" May 27 03:27:05.473133 containerd[1895]: time="2025-05-27T03:27:05.473111599Z" level=info msg="CreateContainer within sandbox \"6039a1b8b68cd09a566570ffb2f19b1a61b13d3f7555245d5b1bad95d089a4a8\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"f3e02d8557ab7149c5e39c31c6d15c77f543e963535f559aaa83cc4b5004dbef\"" May 27 03:27:05.473387 containerd[1895]: time="2025-05-27T03:27:05.473370704Z" level=info msg="StartContainer for \"436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187\"" May 27 03:27:05.474361 containerd[1895]: time="2025-05-27T03:27:05.474337753Z" level=info msg="connecting to shim 436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187" address="unix:///run/containerd/s/3e530b6cfa15efd8d23f031e713f5ea1a754b2facde1d92a0676ffa47cc590f5" protocol=ttrpc version=3 May 27 03:27:05.475034 containerd[1895]: time="2025-05-27T03:27:05.475018462Z" level=info msg="StartContainer for \"f3e02d8557ab7149c5e39c31c6d15c77f543e963535f559aaa83cc4b5004dbef\"" May 27 03:27:05.477334 containerd[1895]: time="2025-05-27T03:27:05.476987454Z" level=info msg="connecting to shim f3e02d8557ab7149c5e39c31c6d15c77f543e963535f559aaa83cc4b5004dbef" address="unix:///run/containerd/s/3060c7d25f758a3d46df2da0568d523fc64a615ad467933730c9b6b7e480d00a" protocol=ttrpc version=3 May 27 03:27:05.479376 containerd[1895]: time="2025-05-27T03:27:05.479348575Z" level=info msg="Container ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:05.494133 containerd[1895]: time="2025-05-27T03:27:05.493818414Z" level=info msg="CreateContainer within sandbox \"9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557\"" May 27 03:27:05.494945 containerd[1895]: time="2025-05-27T03:27:05.494908974Z" level=info msg="StartContainer for \"ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557\"" May 27 03:27:05.496602 containerd[1895]: time="2025-05-27T03:27:05.496545538Z" level=info msg="connecting to shim ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557" address="unix:///run/containerd/s/80f20d15877b2529fc9ca266ce1ef8e61815738865c79d923fc688cf4d70633e" protocol=ttrpc version=3 May 27 03:27:05.504520 systemd[1]: Started cri-containerd-436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187.scope - libcontainer container 436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187. May 27 03:27:05.505548 systemd[1]: Started cri-containerd-f3e02d8557ab7149c5e39c31c6d15c77f543e963535f559aaa83cc4b5004dbef.scope - libcontainer container f3e02d8557ab7149c5e39c31c6d15c77f543e963535f559aaa83cc4b5004dbef. May 27 03:27:05.522593 systemd[1]: Started cri-containerd-ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557.scope - libcontainer container ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557. May 27 03:27:05.627290 kubelet[2755]: E0527 03:27:05.627221 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.17.111:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" May 27 03:27:05.632453 kubelet[2755]: E0527 03:27:05.632386 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.17.111:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" May 27 03:27:05.648679 containerd[1895]: time="2025-05-27T03:27:05.648236932Z" level=info msg="StartContainer for \"436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187\" returns successfully" May 27 03:27:05.656616 containerd[1895]: time="2025-05-27T03:27:05.656517364Z" level=info msg="StartContainer for \"f3e02d8557ab7149c5e39c31c6d15c77f543e963535f559aaa83cc4b5004dbef\" returns successfully" May 27 03:27:05.659166 containerd[1895]: time="2025-05-27T03:27:05.659065191Z" level=info msg="StartContainer for \"ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557\" returns successfully" May 27 03:27:05.966973 kubelet[2755]: E0527 03:27:05.966838 2755 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.17.111:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-17-111?timeout=10s\": dial tcp 172.31.17.111:6443: connect: connection refused" interval="1.6s" May 27 03:27:05.988779 kubelet[2755]: E0527 03:27:05.988732 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.17.111:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-17-111&limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" May 27 03:27:06.021950 kubelet[2755]: E0527 03:27:06.021902 2755 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.17.111:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" May 27 03:27:06.160647 kubelet[2755]: I0527 03:27:06.159921 2755 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-17-111" May 27 03:27:06.160647 kubelet[2755]: E0527 03:27:06.160234 2755 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.17.111:6443/api/v1/nodes\": dial tcp 172.31.17.111:6443: connect: connection refused" node="ip-172-31-17-111" May 27 03:27:06.555340 kubelet[2755]: E0527 03:27:06.555284 2755 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.31.17.111:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.17.111:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" May 27 03:27:06.666346 kubelet[2755]: E0527 03:27:06.664716 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:06.666346 kubelet[2755]: E0527 03:27:06.665140 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:06.672338 kubelet[2755]: E0527 03:27:06.671248 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:07.674344 kubelet[2755]: E0527 03:27:07.673453 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:07.674344 kubelet[2755]: E0527 03:27:07.673912 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:07.674344 kubelet[2755]: E0527 03:27:07.674259 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:07.762389 kubelet[2755]: I0527 03:27:07.762364 2755 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-17-111" May 27 03:27:08.229725 kubelet[2755]: I0527 03:27:08.229692 2755 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-17-111" May 27 03:27:08.229725 kubelet[2755]: E0527 03:27:08.229727 2755 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ip-172-31-17-111\": node \"ip-172-31-17-111\" not found" May 27 03:27:08.253727 kubelet[2755]: E0527 03:27:08.253694 2755 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:08.353978 kubelet[2755]: E0527 03:27:08.353921 2755 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:08.454985 kubelet[2755]: E0527 03:27:08.454904 2755 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:08.555412 kubelet[2755]: E0527 03:27:08.555278 2755 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:08.616538 systemd[1]: systemd-hostnamed.service: Deactivated successfully. May 27 03:27:08.656032 kubelet[2755]: E0527 03:27:08.655988 2755 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:08.674336 kubelet[2755]: E0527 03:27:08.674292 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:08.674481 kubelet[2755]: E0527 03:27:08.674386 2755 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-17-111\" not found" node="ip-172-31-17-111" May 27 03:27:08.756852 kubelet[2755]: E0527 03:27:08.756812 2755 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:08.851884 kubelet[2755]: I0527 03:27:08.851771 2755 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:08.861434 kubelet[2755]: E0527 03:27:08.861393 2755 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-17-111\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:08.861434 kubelet[2755]: I0527 03:27:08.861422 2755 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:08.863343 kubelet[2755]: E0527 03:27:08.863297 2755 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-17-111\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:08.863470 kubelet[2755]: I0527 03:27:08.863358 2755 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:08.865573 kubelet[2755]: E0527 03:27:08.865541 2755 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-17-111\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:09.505387 kubelet[2755]: I0527 03:27:09.505346 2755 apiserver.go:52] "Watching apiserver" May 27 03:27:09.552341 kubelet[2755]: I0527 03:27:09.552267 2755 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" May 27 03:27:09.674419 kubelet[2755]: I0527 03:27:09.674392 2755 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:10.542350 systemd[1]: Reload requested from client PID 3038 ('systemctl') (unit session-7.scope)... May 27 03:27:10.542369 systemd[1]: Reloading... May 27 03:27:10.682359 zram_generator::config[3078]: No configuration found. May 27 03:27:10.807783 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 27 03:27:10.964854 systemd[1]: Reloading finished in 421 ms. May 27 03:27:11.000305 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:27:11.012344 systemd[1]: kubelet.service: Deactivated successfully. May 27 03:27:11.012684 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:27:11.012765 systemd[1]: kubelet.service: Consumed 1.331s CPU time, 128.3M memory peak. May 27 03:27:11.014991 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 27 03:27:11.281479 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 27 03:27:11.291923 (kubelet)[3142]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 27 03:27:11.352208 kubelet[3142]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:27:11.352208 kubelet[3142]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 27 03:27:11.352208 kubelet[3142]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 27 03:27:11.352614 kubelet[3142]: I0527 03:27:11.352269 3142 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 27 03:27:11.362476 kubelet[3142]: I0527 03:27:11.362047 3142 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" May 27 03:27:11.362476 kubelet[3142]: I0527 03:27:11.362076 3142 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 27 03:27:11.362476 kubelet[3142]: I0527 03:27:11.362297 3142 server.go:956] "Client rotation is on, will bootstrap in background" May 27 03:27:11.363509 kubelet[3142]: I0527 03:27:11.363483 3142 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" May 27 03:27:11.365786 kubelet[3142]: I0527 03:27:11.365741 3142 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 27 03:27:11.372853 kubelet[3142]: I0527 03:27:11.372825 3142 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 27 03:27:11.375871 kubelet[3142]: I0527 03:27:11.375837 3142 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 27 03:27:11.376111 kubelet[3142]: I0527 03:27:11.376072 3142 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 27 03:27:11.377332 kubelet[3142]: I0527 03:27:11.376121 3142 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-17-111","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 27 03:27:11.377332 kubelet[3142]: I0527 03:27:11.376652 3142 topology_manager.go:138] "Creating topology manager with none policy" May 27 03:27:11.377332 kubelet[3142]: I0527 03:27:11.376665 3142 container_manager_linux.go:303] "Creating device plugin manager" May 27 03:27:11.377332 kubelet[3142]: I0527 03:27:11.376782 3142 state_mem.go:36] "Initialized new in-memory state store" May 27 03:27:11.377332 kubelet[3142]: I0527 03:27:11.376934 3142 kubelet.go:480] "Attempting to sync node with API server" May 27 03:27:11.377561 kubelet[3142]: I0527 03:27:11.377030 3142 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" May 27 03:27:11.377561 kubelet[3142]: I0527 03:27:11.377057 3142 kubelet.go:386] "Adding apiserver pod source" May 27 03:27:11.377561 kubelet[3142]: I0527 03:27:11.377072 3142 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 27 03:27:11.382473 kubelet[3142]: I0527 03:27:11.382450 3142 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 27 03:27:11.383268 kubelet[3142]: I0527 03:27:11.383238 3142 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" May 27 03:27:11.392287 kubelet[3142]: I0527 03:27:11.392257 3142 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 27 03:27:11.392429 kubelet[3142]: I0527 03:27:11.392308 3142 server.go:1289] "Started kubelet" May 27 03:27:11.395337 kubelet[3142]: I0527 03:27:11.395298 3142 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 27 03:27:11.396342 kubelet[3142]: I0527 03:27:11.396289 3142 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 May 27 03:27:11.399158 kubelet[3142]: I0527 03:27:11.399130 3142 server.go:317] "Adding debug handlers to kubelet server" May 27 03:27:11.402491 kubelet[3142]: I0527 03:27:11.402429 3142 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 27 03:27:11.402714 kubelet[3142]: I0527 03:27:11.402628 3142 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 27 03:27:11.403726 kubelet[3142]: I0527 03:27:11.402842 3142 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 27 03:27:11.406638 kubelet[3142]: I0527 03:27:11.406616 3142 volume_manager.go:297] "Starting Kubelet Volume Manager" May 27 03:27:11.407115 kubelet[3142]: E0527 03:27:11.407005 3142 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-17-111\" not found" May 27 03:27:11.410082 kubelet[3142]: I0527 03:27:11.410058 3142 desired_state_of_world_populator.go:150] "Desired state populator starts to run" May 27 03:27:11.410182 kubelet[3142]: I0527 03:27:11.410171 3142 reconciler.go:26] "Reconciler: start to sync state" May 27 03:27:11.415142 kubelet[3142]: I0527 03:27:11.414880 3142 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 27 03:27:11.417113 kubelet[3142]: I0527 03:27:11.416965 3142 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" May 27 03:27:11.417558 kubelet[3142]: I0527 03:27:11.417386 3142 factory.go:223] Registration of the containerd container factory successfully May 27 03:27:11.417759 kubelet[3142]: I0527 03:27:11.417729 3142 factory.go:223] Registration of the systemd container factory successfully May 27 03:27:11.418507 kubelet[3142]: I0527 03:27:11.418188 3142 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" May 27 03:27:11.418507 kubelet[3142]: I0527 03:27:11.418212 3142 status_manager.go:230] "Starting to sync pod status with apiserver" May 27 03:27:11.418507 kubelet[3142]: I0527 03:27:11.418231 3142 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 27 03:27:11.418507 kubelet[3142]: I0527 03:27:11.418239 3142 kubelet.go:2436] "Starting kubelet main sync loop" May 27 03:27:11.418507 kubelet[3142]: E0527 03:27:11.418284 3142 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 27 03:27:11.473193 kubelet[3142]: I0527 03:27:11.473168 3142 cpu_manager.go:221] "Starting CPU manager" policy="none" May 27 03:27:11.473433 kubelet[3142]: I0527 03:27:11.473415 3142 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 27 03:27:11.473535 kubelet[3142]: I0527 03:27:11.473525 3142 state_mem.go:36] "Initialized new in-memory state store" May 27 03:27:11.474100 kubelet[3142]: I0527 03:27:11.473704 3142 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 27 03:27:11.474100 kubelet[3142]: I0527 03:27:11.473713 3142 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 27 03:27:11.474100 kubelet[3142]: I0527 03:27:11.473730 3142 policy_none.go:49] "None policy: Start" May 27 03:27:11.474100 kubelet[3142]: I0527 03:27:11.473740 3142 memory_manager.go:186] "Starting memorymanager" policy="None" May 27 03:27:11.474100 kubelet[3142]: I0527 03:27:11.473749 3142 state_mem.go:35] "Initializing new in-memory state store" May 27 03:27:11.474100 kubelet[3142]: I0527 03:27:11.473832 3142 state_mem.go:75] "Updated machine memory state" May 27 03:27:11.479427 kubelet[3142]: E0527 03:27:11.479240 3142 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" May 27 03:27:11.479784 kubelet[3142]: I0527 03:27:11.479769 3142 eviction_manager.go:189] "Eviction manager: starting control loop" May 27 03:27:11.479911 kubelet[3142]: I0527 03:27:11.479870 3142 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 27 03:27:11.480342 kubelet[3142]: I0527 03:27:11.480166 3142 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 27 03:27:11.484011 kubelet[3142]: E0527 03:27:11.483974 3142 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 27 03:27:11.519161 kubelet[3142]: I0527 03:27:11.519116 3142 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:11.520352 kubelet[3142]: I0527 03:27:11.519741 3142 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:11.520815 kubelet[3142]: I0527 03:27:11.519981 3142 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:11.532702 kubelet[3142]: E0527 03:27:11.532584 3142 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-17-111\" already exists" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:11.563508 sudo[3177]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 27 03:27:11.564416 sudo[3177]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) May 27 03:27:11.584347 kubelet[3142]: I0527 03:27:11.582644 3142 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-17-111" May 27 03:27:11.597965 kubelet[3142]: I0527 03:27:11.597575 3142 kubelet_node_status.go:124] "Node was previously registered" node="ip-172-31-17-111" May 27 03:27:11.597965 kubelet[3142]: I0527 03:27:11.597645 3142 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-17-111" May 27 03:27:11.711488 kubelet[3142]: I0527 03:27:11.711389 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0a9ff9cbfcbe58b31618c7e9f09efe42-ca-certs\") pod \"kube-apiserver-ip-172-31-17-111\" (UID: \"0a9ff9cbfcbe58b31618c7e9f09efe42\") " pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:11.711907 kubelet[3142]: I0527 03:27:11.711874 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-ca-certs\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:11.712151 kubelet[3142]: I0527 03:27:11.712111 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:11.712549 kubelet[3142]: I0527 03:27:11.712528 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-k8s-certs\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:11.713078 kubelet[3142]: I0527 03:27:11.712686 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0a9ff9cbfcbe58b31618c7e9f09efe42-k8s-certs\") pod \"kube-apiserver-ip-172-31-17-111\" (UID: \"0a9ff9cbfcbe58b31618c7e9f09efe42\") " pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:11.713078 kubelet[3142]: I0527 03:27:11.713044 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0a9ff9cbfcbe58b31618c7e9f09efe42-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-17-111\" (UID: \"0a9ff9cbfcbe58b31618c7e9f09efe42\") " pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:11.713455 kubelet[3142]: I0527 03:27:11.713341 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-kubeconfig\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:11.713805 kubelet[3142]: I0527 03:27:11.713785 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/46f2cad2ef124d0ed9bf2d351de54cf0-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-17-111\" (UID: \"46f2cad2ef124d0ed9bf2d351de54cf0\") " pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:11.714156 kubelet[3142]: I0527 03:27:11.714048 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d5ffa022a545bafe8ff473dc5bf49c84-kubeconfig\") pod \"kube-scheduler-ip-172-31-17-111\" (UID: \"d5ffa022a545bafe8ff473dc5bf49c84\") " pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:12.166030 sudo[3177]: pam_unix(sudo:session): session closed for user root May 27 03:27:12.379746 kubelet[3142]: I0527 03:27:12.379706 3142 apiserver.go:52] "Watching apiserver" May 27 03:27:12.411336 kubelet[3142]: I0527 03:27:12.410932 3142 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" May 27 03:27:12.448646 kubelet[3142]: I0527 03:27:12.448549 3142 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:12.460609 kubelet[3142]: E0527 03:27:12.460489 3142 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-17-111\" already exists" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:12.512337 kubelet[3142]: I0527 03:27:12.510348 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-17-111" podStartSLOduration=1.510327521 podStartE2EDuration="1.510327521s" podCreationTimestamp="2025-05-27 03:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:27:12.502532836 +0000 UTC m=+1.198972334" watchObservedRunningTime="2025-05-27 03:27:12.510327521 +0000 UTC m=+1.206767010" May 27 03:27:12.583664 kubelet[3142]: I0527 03:27:12.582616 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-17-111" podStartSLOduration=3.582572592 podStartE2EDuration="3.582572592s" podCreationTimestamp="2025-05-27 03:27:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:27:12.541475418 +0000 UTC m=+1.237914916" watchObservedRunningTime="2025-05-27 03:27:12.582572592 +0000 UTC m=+1.279012083" May 27 03:27:12.605759 kubelet[3142]: I0527 03:27:12.605602 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-17-111" podStartSLOduration=1.605580298 podStartE2EDuration="1.605580298s" podCreationTimestamp="2025-05-27 03:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:27:12.583067103 +0000 UTC m=+1.279506596" watchObservedRunningTime="2025-05-27 03:27:12.605580298 +0000 UTC m=+1.302019794" May 27 03:27:13.888347 sudo[2241]: pam_unix(sudo:session): session closed for user root May 27 03:27:13.910473 sshd[2240]: Connection closed by 139.178.68.195 port 58918 May 27 03:27:13.912043 sshd-session[2238]: pam_unix(sshd:session): session closed for user core May 27 03:27:13.916377 systemd[1]: sshd@6-172.31.17.111:22-139.178.68.195:58918.service: Deactivated successfully. May 27 03:27:13.918697 systemd[1]: session-7.scope: Deactivated successfully. May 27 03:27:13.918936 systemd[1]: session-7.scope: Consumed 6.638s CPU time, 211.4M memory peak. May 27 03:27:13.920878 systemd-logind[1875]: Session 7 logged out. Waiting for processes to exit. May 27 03:27:13.923182 systemd-logind[1875]: Removed session 7. May 27 03:27:17.134916 kubelet[3142]: I0527 03:27:17.134884 3142 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 27 03:27:17.135640 kubelet[3142]: I0527 03:27:17.135390 3142 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 27 03:27:17.135689 containerd[1895]: time="2025-05-27T03:27:17.135181526Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 27 03:27:17.856327 systemd[1]: Created slice kubepods-burstable-pod72b54b69_33bf_4546_96cd_e9cdc1ab2ae6.slice - libcontainer container kubepods-burstable-pod72b54b69_33bf_4546_96cd_e9cdc1ab2ae6.slice. May 27 03:27:17.860359 kubelet[3142]: I0527 03:27:17.860042 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-config-path\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860359 kubelet[3142]: I0527 03:27:17.860098 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-kernel\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860359 kubelet[3142]: I0527 03:27:17.860130 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-run\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860359 kubelet[3142]: I0527 03:27:17.860149 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hostproc\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860359 kubelet[3142]: I0527 03:27:17.860165 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cni-path\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860359 kubelet[3142]: I0527 03:27:17.860187 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-xtables-lock\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860600 kubelet[3142]: I0527 03:27:17.860212 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-clustermesh-secrets\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860600 kubelet[3142]: I0527 03:27:17.860232 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/d2a64e20-428c-4dd0-b8ff-727a3ef645a7-kube-proxy\") pod \"kube-proxy-k7s89\" (UID: \"d2a64e20-428c-4dd0-b8ff-727a3ef645a7\") " pod="kube-system/kube-proxy-k7s89" May 27 03:27:17.860600 kubelet[3142]: I0527 03:27:17.860250 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d2a64e20-428c-4dd0-b8ff-727a3ef645a7-xtables-lock\") pod \"kube-proxy-k7s89\" (UID: \"d2a64e20-428c-4dd0-b8ff-727a3ef645a7\") " pod="kube-system/kube-proxy-k7s89" May 27 03:27:17.860600 kubelet[3142]: I0527 03:27:17.860272 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxszl\" (UniqueName: \"kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-kube-api-access-hxszl\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.860600 kubelet[3142]: I0527 03:27:17.860304 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rncd5\" (UniqueName: \"kubernetes.io/projected/d2a64e20-428c-4dd0-b8ff-727a3ef645a7-kube-api-access-rncd5\") pod \"kube-proxy-k7s89\" (UID: \"d2a64e20-428c-4dd0-b8ff-727a3ef645a7\") " pod="kube-system/kube-proxy-k7s89" May 27 03:27:17.861357 kubelet[3142]: I0527 03:27:17.860746 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-cgroup\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.861357 kubelet[3142]: I0527 03:27:17.860768 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-net\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.861357 kubelet[3142]: I0527 03:27:17.860781 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hubble-tls\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.861357 kubelet[3142]: I0527 03:27:17.860794 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d2a64e20-428c-4dd0-b8ff-727a3ef645a7-lib-modules\") pod \"kube-proxy-k7s89\" (UID: \"d2a64e20-428c-4dd0-b8ff-727a3ef645a7\") " pod="kube-system/kube-proxy-k7s89" May 27 03:27:17.861357 kubelet[3142]: I0527 03:27:17.860825 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-bpf-maps\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.861357 kubelet[3142]: I0527 03:27:17.860837 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-etc-cni-netd\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.861580 kubelet[3142]: I0527 03:27:17.860853 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-lib-modules\") pod \"cilium-kw7mc\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " pod="kube-system/cilium-kw7mc" May 27 03:27:17.874267 systemd[1]: Created slice kubepods-besteffort-podd2a64e20_428c_4dd0_b8ff_727a3ef645a7.slice - libcontainer container kubepods-besteffort-podd2a64e20_428c_4dd0_b8ff_727a3ef645a7.slice. May 27 03:27:17.988377 kubelet[3142]: E0527 03:27:17.987676 3142 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found May 27 03:27:17.988377 kubelet[3142]: E0527 03:27:17.987707 3142 projected.go:194] Error preparing data for projected volume kube-api-access-hxszl for pod kube-system/cilium-kw7mc: configmap "kube-root-ca.crt" not found May 27 03:27:17.988377 kubelet[3142]: E0527 03:27:17.987770 3142 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-kube-api-access-hxszl podName:72b54b69-33bf-4546-96cd-e9cdc1ab2ae6 nodeName:}" failed. No retries permitted until 2025-05-27 03:27:18.487750878 +0000 UTC m=+7.184190368 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hxszl" (UniqueName: "kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-kube-api-access-hxszl") pod "cilium-kw7mc" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6") : configmap "kube-root-ca.crt" not found May 27 03:27:17.989813 kubelet[3142]: E0527 03:27:17.989758 3142 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found May 27 03:27:17.989813 kubelet[3142]: E0527 03:27:17.989797 3142 projected.go:194] Error preparing data for projected volume kube-api-access-rncd5 for pod kube-system/kube-proxy-k7s89: configmap "kube-root-ca.crt" not found May 27 03:27:17.989942 kubelet[3142]: E0527 03:27:17.989858 3142 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d2a64e20-428c-4dd0-b8ff-727a3ef645a7-kube-api-access-rncd5 podName:d2a64e20-428c-4dd0-b8ff-727a3ef645a7 nodeName:}" failed. No retries permitted until 2025-05-27 03:27:18.489841807 +0000 UTC m=+7.186281282 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-rncd5" (UniqueName: "kubernetes.io/projected/d2a64e20-428c-4dd0-b8ff-727a3ef645a7-kube-api-access-rncd5") pod "kube-proxy-k7s89" (UID: "d2a64e20-428c-4dd0-b8ff-727a3ef645a7") : configmap "kube-root-ca.crt" not found May 27 03:27:18.349808 systemd[1]: Created slice kubepods-besteffort-pod2f588127_76ec_4bcc_aae8_deb799bd22db.slice - libcontainer container kubepods-besteffort-pod2f588127_76ec_4bcc_aae8_deb799bd22db.slice. May 27 03:27:18.364240 kubelet[3142]: I0527 03:27:18.364199 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2f588127-76ec-4bcc-aae8-deb799bd22db-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-rw6nj\" (UID: \"2f588127-76ec-4bcc-aae8-deb799bd22db\") " pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:27:18.364713 kubelet[3142]: I0527 03:27:18.364252 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z8hr\" (UniqueName: \"kubernetes.io/projected/2f588127-76ec-4bcc-aae8-deb799bd22db-kube-api-access-4z8hr\") pod \"cilium-operator-6c4d7847fc-rw6nj\" (UID: \"2f588127-76ec-4bcc-aae8-deb799bd22db\") " pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:27:18.657154 containerd[1895]: time="2025-05-27T03:27:18.657026166Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-rw6nj,Uid:2f588127-76ec-4bcc-aae8-deb799bd22db,Namespace:kube-system,Attempt:0,}" May 27 03:27:18.692722 containerd[1895]: time="2025-05-27T03:27:18.692375788Z" level=info msg="connecting to shim 44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f" address="unix:///run/containerd/s/9ef580f2a5102259eee9da8a8db76493f4cf2ad837c3202ce48b049acea7b1f5" namespace=k8s.io protocol=ttrpc version=3 May 27 03:27:18.728622 systemd[1]: Started cri-containerd-44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f.scope - libcontainer container 44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f. May 27 03:27:18.774207 containerd[1895]: time="2025-05-27T03:27:18.774063155Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kw7mc,Uid:72b54b69-33bf-4546-96cd-e9cdc1ab2ae6,Namespace:kube-system,Attempt:0,}" May 27 03:27:18.786306 containerd[1895]: time="2025-05-27T03:27:18.786268589Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-k7s89,Uid:d2a64e20-428c-4dd0-b8ff-727a3ef645a7,Namespace:kube-system,Attempt:0,}" May 27 03:27:18.793058 containerd[1895]: time="2025-05-27T03:27:18.793018271Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-rw6nj,Uid:2f588127-76ec-4bcc-aae8-deb799bd22db,Namespace:kube-system,Attempt:0,} returns sandbox id \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\"" May 27 03:27:18.798878 containerd[1895]: time="2025-05-27T03:27:18.798697149Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 27 03:27:18.822662 containerd[1895]: time="2025-05-27T03:27:18.822616247Z" level=info msg="connecting to shim d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca" address="unix:///run/containerd/s/58318f2a8b9c943c3e249565de8f183355cb67ed1e530e211ffbf9508a984248" namespace=k8s.io protocol=ttrpc version=3 May 27 03:27:18.837197 containerd[1895]: time="2025-05-27T03:27:18.837092227Z" level=info msg="connecting to shim 6e941c230013fc5a5937397d0a36f0d82354bd9533ece7469ff59e6636119ab7" address="unix:///run/containerd/s/8f1179692e1ed812194b8e984dec73dc7c894152b3396289b0c70c590720888e" namespace=k8s.io protocol=ttrpc version=3 May 27 03:27:18.850743 systemd[1]: Started cri-containerd-d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca.scope - libcontainer container d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca. May 27 03:27:18.860735 systemd[1]: Started cri-containerd-6e941c230013fc5a5937397d0a36f0d82354bd9533ece7469ff59e6636119ab7.scope - libcontainer container 6e941c230013fc5a5937397d0a36f0d82354bd9533ece7469ff59e6636119ab7. May 27 03:27:18.891454 containerd[1895]: time="2025-05-27T03:27:18.891423963Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kw7mc,Uid:72b54b69-33bf-4546-96cd-e9cdc1ab2ae6,Namespace:kube-system,Attempt:0,} returns sandbox id \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\"" May 27 03:27:18.901347 containerd[1895]: time="2025-05-27T03:27:18.901259049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-k7s89,Uid:d2a64e20-428c-4dd0-b8ff-727a3ef645a7,Namespace:kube-system,Attempt:0,} returns sandbox id \"6e941c230013fc5a5937397d0a36f0d82354bd9533ece7469ff59e6636119ab7\"" May 27 03:27:18.908935 containerd[1895]: time="2025-05-27T03:27:18.908521845Z" level=info msg="CreateContainer within sandbox \"6e941c230013fc5a5937397d0a36f0d82354bd9533ece7469ff59e6636119ab7\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 27 03:27:18.927030 containerd[1895]: time="2025-05-27T03:27:18.926056649Z" level=info msg="Container 186d95020560a6c067ce2e8ed4654a480c8838d286bf8ec6ae45a6478257a3da: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:18.939735 containerd[1895]: time="2025-05-27T03:27:18.939698625Z" level=info msg="CreateContainer within sandbox \"6e941c230013fc5a5937397d0a36f0d82354bd9533ece7469ff59e6636119ab7\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"186d95020560a6c067ce2e8ed4654a480c8838d286bf8ec6ae45a6478257a3da\"" May 27 03:27:18.940487 containerd[1895]: time="2025-05-27T03:27:18.940465139Z" level=info msg="StartContainer for \"186d95020560a6c067ce2e8ed4654a480c8838d286bf8ec6ae45a6478257a3da\"" May 27 03:27:18.941917 containerd[1895]: time="2025-05-27T03:27:18.941891641Z" level=info msg="connecting to shim 186d95020560a6c067ce2e8ed4654a480c8838d286bf8ec6ae45a6478257a3da" address="unix:///run/containerd/s/8f1179692e1ed812194b8e984dec73dc7c894152b3396289b0c70c590720888e" protocol=ttrpc version=3 May 27 03:27:18.969569 systemd[1]: Started cri-containerd-186d95020560a6c067ce2e8ed4654a480c8838d286bf8ec6ae45a6478257a3da.scope - libcontainer container 186d95020560a6c067ce2e8ed4654a480c8838d286bf8ec6ae45a6478257a3da. May 27 03:27:19.020196 containerd[1895]: time="2025-05-27T03:27:19.020166302Z" level=info msg="StartContainer for \"186d95020560a6c067ce2e8ed4654a480c8838d286bf8ec6ae45a6478257a3da\" returns successfully" May 27 03:27:21.165717 kubelet[3142]: I0527 03:27:21.165630 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-k7s89" podStartSLOduration=4.165613488 podStartE2EDuration="4.165613488s" podCreationTimestamp="2025-05-27 03:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:27:19.486556588 +0000 UTC m=+8.182996083" watchObservedRunningTime="2025-05-27 03:27:21.165613488 +0000 UTC m=+9.862053017" May 27 03:27:21.242700 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount228220323.mount: Deactivated successfully. May 27 03:27:22.020439 containerd[1895]: time="2025-05-27T03:27:22.020391899Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:27:22.021163 containerd[1895]: time="2025-05-27T03:27:22.021132170Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" May 27 03:27:22.022659 containerd[1895]: time="2025-05-27T03:27:22.022445739Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:27:22.024036 containerd[1895]: time="2025-05-27T03:27:22.023998196Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 3.225249795s" May 27 03:27:22.024203 containerd[1895]: time="2025-05-27T03:27:22.024181221Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" May 27 03:27:22.026886 containerd[1895]: time="2025-05-27T03:27:22.026854390Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 27 03:27:22.031260 containerd[1895]: time="2025-05-27T03:27:22.030779592Z" level=info msg="CreateContainer within sandbox \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 27 03:27:22.048238 containerd[1895]: time="2025-05-27T03:27:22.048201905Z" level=info msg="Container ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:22.077897 containerd[1895]: time="2025-05-27T03:27:22.077850445Z" level=info msg="CreateContainer within sandbox \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\"" May 27 03:27:22.078632 containerd[1895]: time="2025-05-27T03:27:22.078531324Z" level=info msg="StartContainer for \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\"" May 27 03:27:22.080231 containerd[1895]: time="2025-05-27T03:27:22.080134173Z" level=info msg="connecting to shim ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18" address="unix:///run/containerd/s/9ef580f2a5102259eee9da8a8db76493f4cf2ad837c3202ce48b049acea7b1f5" protocol=ttrpc version=3 May 27 03:27:22.104161 systemd[1]: Started cri-containerd-ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18.scope - libcontainer container ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18. May 27 03:27:22.144761 containerd[1895]: time="2025-05-27T03:27:22.144720845Z" level=info msg="StartContainer for \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" returns successfully" May 27 03:27:22.331593 update_engine[1876]: I20250527 03:27:22.331370 1876 update_attempter.cc:509] Updating boot flags... May 27 03:27:27.864645 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1634189511.mount: Deactivated successfully. May 27 03:27:30.414084 containerd[1895]: time="2025-05-27T03:27:30.414034217Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:27:30.416048 containerd[1895]: time="2025-05-27T03:27:30.415985564Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" May 27 03:27:30.419119 containerd[1895]: time="2025-05-27T03:27:30.419032258Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 27 03:27:30.432682 containerd[1895]: time="2025-05-27T03:27:30.432617903Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 8.405725373s" May 27 03:27:30.432682 containerd[1895]: time="2025-05-27T03:27:30.432656765Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" May 27 03:27:30.443264 containerd[1895]: time="2025-05-27T03:27:30.443179119Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 27 03:27:30.476287 containerd[1895]: time="2025-05-27T03:27:30.476098604Z" level=info msg="Container 5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:30.479777 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2995831052.mount: Deactivated successfully. May 27 03:27:30.493530 containerd[1895]: time="2025-05-27T03:27:30.493488469Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\"" May 27 03:27:30.494075 containerd[1895]: time="2025-05-27T03:27:30.494044524Z" level=info msg="StartContainer for \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\"" May 27 03:27:30.497516 containerd[1895]: time="2025-05-27T03:27:30.497483019Z" level=info msg="connecting to shim 5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b" address="unix:///run/containerd/s/58318f2a8b9c943c3e249565de8f183355cb67ed1e530e211ffbf9508a984248" protocol=ttrpc version=3 May 27 03:27:30.562574 systemd[1]: Started cri-containerd-5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b.scope - libcontainer container 5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b. May 27 03:27:30.600183 containerd[1895]: time="2025-05-27T03:27:30.600137943Z" level=info msg="StartContainer for \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" returns successfully" May 27 03:27:30.617202 systemd[1]: cri-containerd-5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b.scope: Deactivated successfully. May 27 03:27:30.659598 containerd[1895]: time="2025-05-27T03:27:30.659268261Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" id:\"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" pid:3874 exited_at:{seconds:1748316450 nanos:620737001}" May 27 03:27:30.660335 containerd[1895]: time="2025-05-27T03:27:30.660272349Z" level=info msg="received exit event container_id:\"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" id:\"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" pid:3874 exited_at:{seconds:1748316450 nanos:620737001}" May 27 03:27:31.473779 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b-rootfs.mount: Deactivated successfully. May 27 03:27:31.564859 containerd[1895]: time="2025-05-27T03:27:31.564800641Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 27 03:27:31.581461 containerd[1895]: time="2025-05-27T03:27:31.581400627Z" level=info msg="Container 07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:31.594021 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2004062408.mount: Deactivated successfully. May 27 03:27:31.602967 kubelet[3142]: I0527 03:27:31.602921 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:27:31.604122 kubelet[3142]: I0527 03:27:31.602979 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:27:31.611341 kubelet[3142]: I0527 03:27:31.608788 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" podStartSLOduration=10.376861564 podStartE2EDuration="13.607593502s" podCreationTimestamp="2025-05-27 03:27:18 +0000 UTC" firstStartedPulling="2025-05-27 03:27:18.794671709 +0000 UTC m=+7.491111182" lastFinishedPulling="2025-05-27 03:27:22.025403628 +0000 UTC m=+10.721843120" observedRunningTime="2025-05-27 03:27:22.555652553 +0000 UTC m=+11.252092048" watchObservedRunningTime="2025-05-27 03:27:31.607593502 +0000 UTC m=+20.304032996" May 27 03:27:31.621157 kubelet[3142]: I0527 03:27:31.621130 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:27:31.623655 containerd[1895]: time="2025-05-27T03:27:31.623559170Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\"" May 27 03:27:31.625278 containerd[1895]: time="2025-05-27T03:27:31.625237787Z" level=info msg="StartContainer for \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\"" May 27 03:27:31.627931 containerd[1895]: time="2025-05-27T03:27:31.627896102Z" level=info msg="connecting to shim 07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d" address="unix:///run/containerd/s/58318f2a8b9c943c3e249565de8f183355cb67ed1e530e211ffbf9508a984248" protocol=ttrpc version=3 May 27 03:27:31.639677 kubelet[3142]: I0527 03:27:31.639367 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:27:31.639677 kubelet[3142]: I0527 03:27:31.639480 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-kw7mc","kube-system/cilium-operator-6c4d7847fc-rw6nj","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-proxy-k7s89","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:27:31.639677 kubelet[3142]: E0527 03:27:31.639538 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-kw7mc" May 27 03:27:31.639677 kubelet[3142]: E0527 03:27:31.639558 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:27:31.639677 kubelet[3142]: E0527 03:27:31.639603 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:31.639677 kubelet[3142]: E0527 03:27:31.639617 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:27:31.639677 kubelet[3142]: E0527 03:27:31.639629 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:31.639677 kubelet[3142]: E0527 03:27:31.639641 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:31.639677 kubelet[3142]: I0527 03:27:31.639657 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:27:31.668548 systemd[1]: Started cri-containerd-07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d.scope - libcontainer container 07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d. May 27 03:27:31.709474 containerd[1895]: time="2025-05-27T03:27:31.709434071Z" level=info msg="StartContainer for \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" returns successfully" May 27 03:27:31.725540 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 27 03:27:31.725903 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 27 03:27:31.727543 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... May 27 03:27:31.730353 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 27 03:27:31.736004 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 27 03:27:31.738946 containerd[1895]: time="2025-05-27T03:27:31.737021737Z" level=info msg="TaskExit event in podsandbox handler container_id:\"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" id:\"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" pid:3925 exited_at:{seconds:1748316451 nanos:736224298}" May 27 03:27:31.739843 containerd[1895]: time="2025-05-27T03:27:31.739727962Z" level=info msg="received exit event container_id:\"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" id:\"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" pid:3925 exited_at:{seconds:1748316451 nanos:736224298}" May 27 03:27:31.741099 systemd[1]: cri-containerd-07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d.scope: Deactivated successfully. May 27 03:27:31.788509 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 27 03:27:32.473966 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d-rootfs.mount: Deactivated successfully. May 27 03:27:32.570192 containerd[1895]: time="2025-05-27T03:27:32.570152204Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 27 03:27:32.596252 containerd[1895]: time="2025-05-27T03:27:32.594347599Z" level=info msg="Container 700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:32.606536 containerd[1895]: time="2025-05-27T03:27:32.606497008Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\"" May 27 03:27:32.607242 containerd[1895]: time="2025-05-27T03:27:32.607216521Z" level=info msg="StartContainer for \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\"" May 27 03:27:32.608749 containerd[1895]: time="2025-05-27T03:27:32.608721060Z" level=info msg="connecting to shim 700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e" address="unix:///run/containerd/s/58318f2a8b9c943c3e249565de8f183355cb67ed1e530e211ffbf9508a984248" protocol=ttrpc version=3 May 27 03:27:32.631594 systemd[1]: Started cri-containerd-700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e.scope - libcontainer container 700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e. May 27 03:27:32.677732 containerd[1895]: time="2025-05-27T03:27:32.677665239Z" level=info msg="StartContainer for \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" returns successfully" May 27 03:27:32.690769 systemd[1]: cri-containerd-700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e.scope: Deactivated successfully. May 27 03:27:32.691032 systemd[1]: cri-containerd-700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e.scope: Consumed 27ms CPU time, 4.2M memory peak, 1.2M read from disk. May 27 03:27:32.701573 containerd[1895]: time="2025-05-27T03:27:32.701178450Z" level=info msg="received exit event container_id:\"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" id:\"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" pid:3972 exited_at:{seconds:1748316452 nanos:700876662}" May 27 03:27:32.701876 containerd[1895]: time="2025-05-27T03:27:32.701821648Z" level=info msg="TaskExit event in podsandbox handler container_id:\"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" id:\"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" pid:3972 exited_at:{seconds:1748316452 nanos:700876662}" May 27 03:27:32.729349 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e-rootfs.mount: Deactivated successfully. May 27 03:27:33.573225 containerd[1895]: time="2025-05-27T03:27:33.573181580Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 27 03:27:33.586145 containerd[1895]: time="2025-05-27T03:27:33.586102989Z" level=info msg="Container cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:33.600144 containerd[1895]: time="2025-05-27T03:27:33.600098371Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\"" May 27 03:27:33.601056 containerd[1895]: time="2025-05-27T03:27:33.601027923Z" level=info msg="StartContainer for \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\"" May 27 03:27:33.602644 containerd[1895]: time="2025-05-27T03:27:33.602574443Z" level=info msg="connecting to shim cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d" address="unix:///run/containerd/s/58318f2a8b9c943c3e249565de8f183355cb67ed1e530e211ffbf9508a984248" protocol=ttrpc version=3 May 27 03:27:33.629603 systemd[1]: Started cri-containerd-cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d.scope - libcontainer container cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d. May 27 03:27:33.663096 systemd[1]: cri-containerd-cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d.scope: Deactivated successfully. May 27 03:27:33.665589 containerd[1895]: time="2025-05-27T03:27:33.665466408Z" level=info msg="received exit event container_id:\"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" id:\"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" pid:4013 exited_at:{seconds:1748316453 nanos:664169664}" May 27 03:27:33.668629 containerd[1895]: time="2025-05-27T03:27:33.668581954Z" level=info msg="TaskExit event in podsandbox handler container_id:\"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" id:\"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" pid:4013 exited_at:{seconds:1748316453 nanos:664169664}" May 27 03:27:33.668912 containerd[1895]: time="2025-05-27T03:27:33.668724114Z" level=info msg="StartContainer for \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" returns successfully" May 27 03:27:33.694936 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d-rootfs.mount: Deactivated successfully. May 27 03:27:34.584435 containerd[1895]: time="2025-05-27T03:27:34.584249787Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 27 03:27:34.606697 containerd[1895]: time="2025-05-27T03:27:34.605993382Z" level=info msg="Container 335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd: CDI devices from CRI Config.CDIDevices: []" May 27 03:27:34.610945 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount913399560.mount: Deactivated successfully. May 27 03:27:34.619847 containerd[1895]: time="2025-05-27T03:27:34.619794064Z" level=info msg="CreateContainer within sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\"" May 27 03:27:34.620562 containerd[1895]: time="2025-05-27T03:27:34.620377731Z" level=info msg="StartContainer for \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\"" May 27 03:27:34.621333 containerd[1895]: time="2025-05-27T03:27:34.621298444Z" level=info msg="connecting to shim 335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd" address="unix:///run/containerd/s/58318f2a8b9c943c3e249565de8f183355cb67ed1e530e211ffbf9508a984248" protocol=ttrpc version=3 May 27 03:27:34.641717 systemd[1]: Started cri-containerd-335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd.scope - libcontainer container 335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd. May 27 03:27:34.708267 containerd[1895]: time="2025-05-27T03:27:34.708223299Z" level=info msg="StartContainer for \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" returns successfully" May 27 03:27:34.828341 containerd[1895]: time="2025-05-27T03:27:34.828263138Z" level=info msg="TaskExit event in podsandbox handler container_id:\"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" id:\"d7b503ba58532c0a3192c7fc06066dfebd0be109ada9a3a0a96e10f50ebf875f\" pid:4080 exited_at:{seconds:1748316454 nanos:827025407}" May 27 03:27:34.903621 kubelet[3142]: I0527 03:27:34.903020 3142 kubelet_node_status.go:501] "Fast updating node status as it just became ready" May 27 03:27:37.212820 systemd-networkd[1825]: cilium_host: Link UP May 27 03:27:37.213729 systemd-networkd[1825]: cilium_net: Link UP May 27 03:27:37.213949 systemd-networkd[1825]: cilium_net: Gained carrier May 27 03:27:37.214150 systemd-networkd[1825]: cilium_host: Gained carrier May 27 03:27:37.215442 (udev-worker)[4117]: Network interface NamePolicy= disabled on kernel command line. May 27 03:27:37.215477 (udev-worker)[4150]: Network interface NamePolicy= disabled on kernel command line. May 27 03:27:37.351752 (udev-worker)[4164]: Network interface NamePolicy= disabled on kernel command line. May 27 03:27:37.357262 systemd-networkd[1825]: cilium_vxlan: Link UP May 27 03:27:37.357270 systemd-networkd[1825]: cilium_vxlan: Gained carrier May 27 03:27:37.680759 systemd-networkd[1825]: cilium_net: Gained IPv6LL May 27 03:27:37.881407 kernel: NET: Registered PF_ALG protocol family May 27 03:27:37.952539 systemd-networkd[1825]: cilium_host: Gained IPv6LL May 27 03:27:38.528661 systemd-networkd[1825]: cilium_vxlan: Gained IPv6LL May 27 03:27:38.650525 systemd-networkd[1825]: lxc_health: Link UP May 27 03:27:38.658271 systemd-networkd[1825]: lxc_health: Gained carrier May 27 03:27:38.814424 kubelet[3142]: I0527 03:27:38.814255 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-kw7mc" podStartSLOduration=10.270025981 podStartE2EDuration="21.814233432s" podCreationTimestamp="2025-05-27 03:27:17 +0000 UTC" firstStartedPulling="2025-05-27 03:27:18.893184881 +0000 UTC m=+7.589624354" lastFinishedPulling="2025-05-27 03:27:30.437392329 +0000 UTC m=+19.133831805" observedRunningTime="2025-05-27 03:27:35.617418183 +0000 UTC m=+24.313857672" watchObservedRunningTime="2025-05-27 03:27:38.814233432 +0000 UTC m=+27.510672930" May 27 03:27:40.640619 systemd-networkd[1825]: lxc_health: Gained IPv6LL May 27 03:27:41.679349 kubelet[3142]: I0527 03:27:41.673338 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:27:41.683146 kubelet[3142]: I0527 03:27:41.682969 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:27:41.687341 kubelet[3142]: I0527 03:27:41.687259 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:27:41.703986 kubelet[3142]: I0527 03:27:41.703666 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:27:41.703986 kubelet[3142]: I0527 03:27:41.703834 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-rw6nj","kube-system/cilium-kw7mc","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-proxy-k7s89","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:27:41.703986 kubelet[3142]: E0527 03:27:41.703881 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:27:41.703986 kubelet[3142]: E0527 03:27:41.703903 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-kw7mc" May 27 03:27:41.703986 kubelet[3142]: E0527 03:27:41.703916 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:41.703986 kubelet[3142]: E0527 03:27:41.703928 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:27:41.703986 kubelet[3142]: E0527 03:27:41.703939 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:41.703986 kubelet[3142]: E0527 03:27:41.703950 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:41.703986 kubelet[3142]: I0527 03:27:41.703965 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:27:42.644649 ntpd[1869]: Listen normally on 6 cilium_host 192.168.0.70:123 May 27 03:27:42.645361 ntpd[1869]: 27 May 03:27:42 ntpd[1869]: Listen normally on 6 cilium_host 192.168.0.70:123 May 27 03:27:42.645361 ntpd[1869]: 27 May 03:27:42 ntpd[1869]: Listen normally on 7 cilium_net [fe80::7861:efff:fe93:78b1%4]:123 May 27 03:27:42.645361 ntpd[1869]: 27 May 03:27:42 ntpd[1869]: Listen normally on 8 cilium_host [fe80::249f:faff:fedc:28af%5]:123 May 27 03:27:42.645361 ntpd[1869]: 27 May 03:27:42 ntpd[1869]: Listen normally on 9 cilium_vxlan [fe80::3cd9:5bff:feba:363e%6]:123 May 27 03:27:42.645361 ntpd[1869]: 27 May 03:27:42 ntpd[1869]: Listen normally on 10 lxc_health [fe80::d86f:bdff:fe2f:6110%8]:123 May 27 03:27:42.644753 ntpd[1869]: Listen normally on 7 cilium_net [fe80::7861:efff:fe93:78b1%4]:123 May 27 03:27:42.644811 ntpd[1869]: Listen normally on 8 cilium_host [fe80::249f:faff:fedc:28af%5]:123 May 27 03:27:42.644853 ntpd[1869]: Listen normally on 9 cilium_vxlan [fe80::3cd9:5bff:feba:363e%6]:123 May 27 03:27:42.644892 ntpd[1869]: Listen normally on 10 lxc_health [fe80::d86f:bdff:fe2f:6110%8]:123 May 27 03:27:46.591054 systemd[1]: Started sshd@7-172.31.17.111:22-139.178.68.195:37738.service - OpenSSH per-connection server daemon (139.178.68.195:37738). May 27 03:27:46.795832 sshd[4515]: Accepted publickey for core from 139.178.68.195 port 37738 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:27:46.798647 sshd-session[4515]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:27:46.806608 systemd-logind[1875]: New session 8 of user core. May 27 03:27:46.810543 systemd[1]: Started session-8.scope - Session 8 of User core. May 27 03:27:47.616669 sshd[4517]: Connection closed by 139.178.68.195 port 37738 May 27 03:27:47.617293 sshd-session[4515]: pam_unix(sshd:session): session closed for user core May 27 03:27:47.620799 systemd[1]: sshd@7-172.31.17.111:22-139.178.68.195:37738.service: Deactivated successfully. May 27 03:27:47.628296 systemd[1]: session-8.scope: Deactivated successfully. May 27 03:27:47.631214 systemd-logind[1875]: Session 8 logged out. Waiting for processes to exit. May 27 03:27:47.633521 systemd-logind[1875]: Removed session 8. May 27 03:27:51.720807 kubelet[3142]: I0527 03:27:51.720769 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:27:51.723102 kubelet[3142]: I0527 03:27:51.723065 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:27:51.725522 kubelet[3142]: I0527 03:27:51.725440 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:27:51.742128 kubelet[3142]: I0527 03:27:51.742084 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:27:51.742276 kubelet[3142]: I0527 03:27:51.742231 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-rw6nj","kube-system/cilium-kw7mc","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-proxy-k7s89","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:27:51.742754 kubelet[3142]: E0527 03:27:51.742278 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:27:51.742754 kubelet[3142]: E0527 03:27:51.742297 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-kw7mc" May 27 03:27:51.742754 kubelet[3142]: E0527 03:27:51.742309 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:27:51.742754 kubelet[3142]: E0527 03:27:51.742343 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:27:51.742754 kubelet[3142]: E0527 03:27:51.742356 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:27:51.742754 kubelet[3142]: E0527 03:27:51.742645 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:27:51.742754 kubelet[3142]: I0527 03:27:51.742660 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:27:52.651501 systemd[1]: Started sshd@8-172.31.17.111:22-139.178.68.195:37740.service - OpenSSH per-connection server daemon (139.178.68.195:37740). May 27 03:27:52.833215 sshd[4533]: Accepted publickey for core from 139.178.68.195 port 37740 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:27:52.834548 sshd-session[4533]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:27:52.840237 systemd-logind[1875]: New session 9 of user core. May 27 03:27:52.844523 systemd[1]: Started session-9.scope - Session 9 of User core. May 27 03:27:53.070395 sshd[4535]: Connection closed by 139.178.68.195 port 37740 May 27 03:27:53.070960 sshd-session[4533]: pam_unix(sshd:session): session closed for user core May 27 03:27:53.074893 systemd[1]: sshd@8-172.31.17.111:22-139.178.68.195:37740.service: Deactivated successfully. May 27 03:27:53.077025 systemd[1]: session-9.scope: Deactivated successfully. May 27 03:27:53.078427 systemd-logind[1875]: Session 9 logged out. Waiting for processes to exit. May 27 03:27:53.080100 systemd-logind[1875]: Removed session 9. May 27 03:27:58.104774 systemd[1]: Started sshd@9-172.31.17.111:22-139.178.68.195:50432.service - OpenSSH per-connection server daemon (139.178.68.195:50432). May 27 03:27:58.293033 sshd[4548]: Accepted publickey for core from 139.178.68.195 port 50432 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:27:58.295059 sshd-session[4548]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:27:58.300357 systemd-logind[1875]: New session 10 of user core. May 27 03:27:58.307682 systemd[1]: Started session-10.scope - Session 10 of User core. May 27 03:27:58.500992 sshd[4550]: Connection closed by 139.178.68.195 port 50432 May 27 03:27:58.501579 sshd-session[4548]: pam_unix(sshd:session): session closed for user core May 27 03:27:58.505634 systemd[1]: sshd@9-172.31.17.111:22-139.178.68.195:50432.service: Deactivated successfully. May 27 03:27:58.507451 systemd[1]: session-10.scope: Deactivated successfully. May 27 03:27:58.508744 systemd-logind[1875]: Session 10 logged out. Waiting for processes to exit. May 27 03:27:58.510467 systemd-logind[1875]: Removed session 10. May 27 03:28:01.780287 kubelet[3142]: I0527 03:28:01.779568 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:28:01.780287 kubelet[3142]: I0527 03:28:01.779633 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:28:01.834889 kubelet[3142]: I0527 03:28:01.834858 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:28:01.870592 kubelet[3142]: I0527 03:28:01.870556 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:28:01.870762 kubelet[3142]: I0527 03:28:01.870709 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-rw6nj","kube-system/cilium-kw7mc","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-proxy-k7s89","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:28:01.870815 kubelet[3142]: E0527 03:28:01.870763 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:28:01.870815 kubelet[3142]: E0527 03:28:01.870785 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-kw7mc" May 27 03:28:01.870815 kubelet[3142]: E0527 03:28:01.870799 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:28:01.870815 kubelet[3142]: E0527 03:28:01.870813 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:28:01.870958 kubelet[3142]: E0527 03:28:01.870825 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:28:01.870958 kubelet[3142]: E0527 03:28:01.870839 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:28:01.870958 kubelet[3142]: I0527 03:28:01.870856 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:28:03.535152 systemd[1]: Started sshd@10-172.31.17.111:22-139.178.68.195:52426.service - OpenSSH per-connection server daemon (139.178.68.195:52426). May 27 03:28:03.707000 sshd[4563]: Accepted publickey for core from 139.178.68.195 port 52426 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:03.708701 sshd-session[4563]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:03.715324 systemd-logind[1875]: New session 11 of user core. May 27 03:28:03.722588 systemd[1]: Started session-11.scope - Session 11 of User core. May 27 03:28:03.908494 sshd[4565]: Connection closed by 139.178.68.195 port 52426 May 27 03:28:03.909387 sshd-session[4563]: pam_unix(sshd:session): session closed for user core May 27 03:28:03.913546 systemd[1]: sshd@10-172.31.17.111:22-139.178.68.195:52426.service: Deactivated successfully. May 27 03:28:03.915932 systemd[1]: session-11.scope: Deactivated successfully. May 27 03:28:03.917163 systemd-logind[1875]: Session 11 logged out. Waiting for processes to exit. May 27 03:28:03.919488 systemd-logind[1875]: Removed session 11. May 27 03:28:03.947672 systemd[1]: Started sshd@11-172.31.17.111:22-139.178.68.195:52432.service - OpenSSH per-connection server daemon (139.178.68.195:52432). May 27 03:28:04.121928 sshd[4578]: Accepted publickey for core from 139.178.68.195 port 52432 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:04.123579 sshd-session[4578]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:04.129641 systemd-logind[1875]: New session 12 of user core. May 27 03:28:04.134536 systemd[1]: Started session-12.scope - Session 12 of User core. May 27 03:28:04.376336 sshd[4580]: Connection closed by 139.178.68.195 port 52432 May 27 03:28:04.376638 sshd-session[4578]: pam_unix(sshd:session): session closed for user core May 27 03:28:04.386179 systemd-logind[1875]: Session 12 logged out. Waiting for processes to exit. May 27 03:28:04.386908 systemd[1]: sshd@11-172.31.17.111:22-139.178.68.195:52432.service: Deactivated successfully. May 27 03:28:04.390695 systemd[1]: session-12.scope: Deactivated successfully. May 27 03:28:04.395271 systemd-logind[1875]: Removed session 12. May 27 03:28:04.409825 systemd[1]: Started sshd@12-172.31.17.111:22-139.178.68.195:52436.service - OpenSSH per-connection server daemon (139.178.68.195:52436). May 27 03:28:04.582648 sshd[4590]: Accepted publickey for core from 139.178.68.195 port 52436 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:04.584252 sshd-session[4590]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:04.589424 systemd-logind[1875]: New session 13 of user core. May 27 03:28:04.596548 systemd[1]: Started session-13.scope - Session 13 of User core. May 27 03:28:04.790086 sshd[4592]: Connection closed by 139.178.68.195 port 52436 May 27 03:28:04.790730 sshd-session[4590]: pam_unix(sshd:session): session closed for user core May 27 03:28:04.795386 systemd[1]: sshd@12-172.31.17.111:22-139.178.68.195:52436.service: Deactivated successfully. May 27 03:28:04.798714 systemd[1]: session-13.scope: Deactivated successfully. May 27 03:28:04.801383 systemd-logind[1875]: Session 13 logged out. Waiting for processes to exit. May 27 03:28:04.804336 systemd-logind[1875]: Removed session 13. May 27 03:28:09.825889 systemd[1]: Started sshd@13-172.31.17.111:22-139.178.68.195:52440.service - OpenSSH per-connection server daemon (139.178.68.195:52440). May 27 03:28:09.994930 sshd[4606]: Accepted publickey for core from 139.178.68.195 port 52440 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:09.996562 sshd-session[4606]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:10.003060 systemd-logind[1875]: New session 14 of user core. May 27 03:28:10.010563 systemd[1]: Started session-14.scope - Session 14 of User core. May 27 03:28:10.198408 sshd[4608]: Connection closed by 139.178.68.195 port 52440 May 27 03:28:10.198991 sshd-session[4606]: pam_unix(sshd:session): session closed for user core May 27 03:28:10.202951 systemd-logind[1875]: Session 14 logged out. Waiting for processes to exit. May 27 03:28:10.203842 systemd[1]: sshd@13-172.31.17.111:22-139.178.68.195:52440.service: Deactivated successfully. May 27 03:28:10.206040 systemd[1]: session-14.scope: Deactivated successfully. May 27 03:28:10.208061 systemd-logind[1875]: Removed session 14. May 27 03:28:11.889671 kubelet[3142]: I0527 03:28:11.889636 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:28:11.890516 kubelet[3142]: I0527 03:28:11.889687 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:28:11.893019 kubelet[3142]: I0527 03:28:11.893002 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:28:11.910723 kubelet[3142]: I0527 03:28:11.910689 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:28:11.910868 kubelet[3142]: I0527 03:28:11.910852 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-rw6nj","kube-system/cilium-kw7mc","kube-system/kube-proxy-k7s89","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:28:11.910918 kubelet[3142]: E0527 03:28:11.910900 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:28:11.910984 kubelet[3142]: E0527 03:28:11.910919 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-kw7mc" May 27 03:28:11.910984 kubelet[3142]: E0527 03:28:11.910931 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:28:11.910984 kubelet[3142]: E0527 03:28:11.910945 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:28:11.910984 kubelet[3142]: E0527 03:28:11.910956 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:28:11.910984 kubelet[3142]: E0527 03:28:11.910968 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:28:11.910984 kubelet[3142]: I0527 03:28:11.910985 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:28:15.231614 systemd[1]: Started sshd@14-172.31.17.111:22-139.178.68.195:60188.service - OpenSSH per-connection server daemon (139.178.68.195:60188). May 27 03:28:15.409571 sshd[4629]: Accepted publickey for core from 139.178.68.195 port 60188 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:15.411252 sshd-session[4629]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:15.417038 systemd-logind[1875]: New session 15 of user core. May 27 03:28:15.422539 systemd[1]: Started session-15.scope - Session 15 of User core. May 27 03:28:15.609140 sshd[4631]: Connection closed by 139.178.68.195 port 60188 May 27 03:28:15.609617 sshd-session[4629]: pam_unix(sshd:session): session closed for user core May 27 03:28:15.612856 systemd[1]: sshd@14-172.31.17.111:22-139.178.68.195:60188.service: Deactivated successfully. May 27 03:28:15.615598 systemd[1]: session-15.scope: Deactivated successfully. May 27 03:28:15.618042 systemd-logind[1875]: Session 15 logged out. Waiting for processes to exit. May 27 03:28:15.619964 systemd-logind[1875]: Removed session 15. May 27 03:28:15.645729 systemd[1]: Started sshd@15-172.31.17.111:22-139.178.68.195:60204.service - OpenSSH per-connection server daemon (139.178.68.195:60204). May 27 03:28:15.821799 sshd[4643]: Accepted publickey for core from 139.178.68.195 port 60204 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:15.823333 sshd-session[4643]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:15.828785 systemd-logind[1875]: New session 16 of user core. May 27 03:28:15.834553 systemd[1]: Started session-16.scope - Session 16 of User core. May 27 03:28:16.618670 sshd[4645]: Connection closed by 139.178.68.195 port 60204 May 27 03:28:16.619622 sshd-session[4643]: pam_unix(sshd:session): session closed for user core May 27 03:28:16.624479 systemd[1]: sshd@15-172.31.17.111:22-139.178.68.195:60204.service: Deactivated successfully. May 27 03:28:16.626503 systemd[1]: session-16.scope: Deactivated successfully. May 27 03:28:16.627769 systemd-logind[1875]: Session 16 logged out. Waiting for processes to exit. May 27 03:28:16.629919 systemd-logind[1875]: Removed session 16. May 27 03:28:16.650657 systemd[1]: Started sshd@16-172.31.17.111:22-139.178.68.195:60210.service - OpenSSH per-connection server daemon (139.178.68.195:60210). May 27 03:28:16.856075 sshd[4656]: Accepted publickey for core from 139.178.68.195 port 60210 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:16.857558 sshd-session[4656]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:16.863423 systemd-logind[1875]: New session 17 of user core. May 27 03:28:16.868529 systemd[1]: Started session-17.scope - Session 17 of User core. May 27 03:28:17.813676 sshd[4658]: Connection closed by 139.178.68.195 port 60210 May 27 03:28:17.814267 sshd-session[4656]: pam_unix(sshd:session): session closed for user core May 27 03:28:17.833775 systemd-logind[1875]: Session 17 logged out. Waiting for processes to exit. May 27 03:28:17.842748 systemd[1]: sshd@16-172.31.17.111:22-139.178.68.195:60210.service: Deactivated successfully. May 27 03:28:17.845346 systemd[1]: session-17.scope: Deactivated successfully. May 27 03:28:17.850121 systemd-logind[1875]: Removed session 17. May 27 03:28:17.853602 systemd[1]: Started sshd@17-172.31.17.111:22-139.178.68.195:60222.service - OpenSSH per-connection server daemon (139.178.68.195:60222). May 27 03:28:18.023076 sshd[4674]: Accepted publickey for core from 139.178.68.195 port 60222 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:18.024589 sshd-session[4674]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:18.030948 systemd-logind[1875]: New session 18 of user core. May 27 03:28:18.035583 systemd[1]: Started session-18.scope - Session 18 of User core. May 27 03:28:18.400793 sshd[4676]: Connection closed by 139.178.68.195 port 60222 May 27 03:28:18.401808 sshd-session[4674]: pam_unix(sshd:session): session closed for user core May 27 03:28:18.408466 systemd[1]: sshd@17-172.31.17.111:22-139.178.68.195:60222.service: Deactivated successfully. May 27 03:28:18.412632 systemd[1]: session-18.scope: Deactivated successfully. May 27 03:28:18.417789 systemd-logind[1875]: Session 18 logged out. Waiting for processes to exit. May 27 03:28:18.419693 systemd-logind[1875]: Removed session 18. May 27 03:28:18.434709 systemd[1]: Started sshd@18-172.31.17.111:22-139.178.68.195:60238.service - OpenSSH per-connection server daemon (139.178.68.195:60238). May 27 03:28:18.618383 sshd[4686]: Accepted publickey for core from 139.178.68.195 port 60238 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:18.620016 sshd-session[4686]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:18.632844 systemd-logind[1875]: New session 19 of user core. May 27 03:28:18.644577 systemd[1]: Started session-19.scope - Session 19 of User core. May 27 03:28:18.835043 sshd[4688]: Connection closed by 139.178.68.195 port 60238 May 27 03:28:18.835788 sshd-session[4686]: pam_unix(sshd:session): session closed for user core May 27 03:28:18.839346 systemd[1]: sshd@18-172.31.17.111:22-139.178.68.195:60238.service: Deactivated successfully. May 27 03:28:18.841824 systemd[1]: session-19.scope: Deactivated successfully. May 27 03:28:18.843765 systemd-logind[1875]: Session 19 logged out. Waiting for processes to exit. May 27 03:28:18.846218 systemd-logind[1875]: Removed session 19. May 27 03:28:21.926075 kubelet[3142]: I0527 03:28:21.926040 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:28:21.926986 kubelet[3142]: I0527 03:28:21.926092 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:28:21.930579 kubelet[3142]: I0527 03:28:21.930537 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:28:21.942188 kubelet[3142]: I0527 03:28:21.941940 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:28:21.942188 kubelet[3142]: I0527 03:28:21.942081 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-rw6nj","kube-system/cilium-kw7mc","kube-system/kube-proxy-k7s89","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:28:21.942188 kubelet[3142]: E0527 03:28:21.942114 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:28:21.942188 kubelet[3142]: E0527 03:28:21.942126 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-kw7mc" May 27 03:28:21.942188 kubelet[3142]: E0527 03:28:21.942134 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:28:21.942188 kubelet[3142]: E0527 03:28:21.942144 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:28:21.942188 kubelet[3142]: E0527 03:28:21.942152 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:28:21.942188 kubelet[3142]: E0527 03:28:21.942160 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:28:21.942188 kubelet[3142]: I0527 03:28:21.942171 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:28:23.873644 systemd[1]: Started sshd@19-172.31.17.111:22-139.178.68.195:51644.service - OpenSSH per-connection server daemon (139.178.68.195:51644). May 27 03:28:24.048476 sshd[4704]: Accepted publickey for core from 139.178.68.195 port 51644 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:24.050157 sshd-session[4704]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:24.055700 systemd-logind[1875]: New session 20 of user core. May 27 03:28:24.063568 systemd[1]: Started session-20.scope - Session 20 of User core. May 27 03:28:24.254067 sshd[4706]: Connection closed by 139.178.68.195 port 51644 May 27 03:28:24.254921 sshd-session[4704]: pam_unix(sshd:session): session closed for user core May 27 03:28:24.260601 systemd-logind[1875]: Session 20 logged out. Waiting for processes to exit. May 27 03:28:24.261168 systemd[1]: sshd@19-172.31.17.111:22-139.178.68.195:51644.service: Deactivated successfully. May 27 03:28:24.263261 systemd[1]: session-20.scope: Deactivated successfully. May 27 03:28:24.265182 systemd-logind[1875]: Removed session 20. May 27 03:28:29.288702 systemd[1]: Started sshd@20-172.31.17.111:22-139.178.68.195:51654.service - OpenSSH per-connection server daemon (139.178.68.195:51654). May 27 03:28:29.466092 sshd[4717]: Accepted publickey for core from 139.178.68.195 port 51654 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:29.467829 sshd-session[4717]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:29.472820 systemd-logind[1875]: New session 21 of user core. May 27 03:28:29.480542 systemd[1]: Started session-21.scope - Session 21 of User core. May 27 03:28:29.666651 sshd[4719]: Connection closed by 139.178.68.195 port 51654 May 27 03:28:29.667581 sshd-session[4717]: pam_unix(sshd:session): session closed for user core May 27 03:28:29.671389 systemd-logind[1875]: Session 21 logged out. Waiting for processes to exit. May 27 03:28:29.671860 systemd[1]: sshd@20-172.31.17.111:22-139.178.68.195:51654.service: Deactivated successfully. May 27 03:28:29.674215 systemd[1]: session-21.scope: Deactivated successfully. May 27 03:28:29.676152 systemd-logind[1875]: Removed session 21. May 27 03:28:31.955887 kubelet[3142]: I0527 03:28:31.955847 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:28:31.955887 kubelet[3142]: I0527 03:28:31.955894 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:28:31.957853 kubelet[3142]: I0527 03:28:31.957829 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:28:31.974531 kubelet[3142]: I0527 03:28:31.974494 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:28:31.974666 kubelet[3142]: I0527 03:28:31.974621 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-rw6nj","kube-system/cilium-kw7mc","kube-system/kube-proxy-k7s89","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:28:31.974666 kubelet[3142]: E0527 03:28:31.974653 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-rw6nj" May 27 03:28:31.974666 kubelet[3142]: E0527 03:28:31.974666 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-kw7mc" May 27 03:28:31.974746 kubelet[3142]: E0527 03:28:31.974673 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:28:31.974746 kubelet[3142]: E0527 03:28:31.974681 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:28:31.974746 kubelet[3142]: E0527 03:28:31.974690 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:28:31.974746 kubelet[3142]: E0527 03:28:31.974698 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:28:31.974746 kubelet[3142]: I0527 03:28:31.974708 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:28:34.699839 systemd[1]: Started sshd@21-172.31.17.111:22-139.178.68.195:58740.service - OpenSSH per-connection server daemon (139.178.68.195:58740). May 27 03:28:34.871733 sshd[4731]: Accepted publickey for core from 139.178.68.195 port 58740 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:34.873120 sshd-session[4731]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:34.878373 systemd-logind[1875]: New session 22 of user core. May 27 03:28:34.887572 systemd[1]: Started session-22.scope - Session 22 of User core. May 27 03:28:35.124236 sshd[4733]: Connection closed by 139.178.68.195 port 58740 May 27 03:28:35.124744 sshd-session[4731]: pam_unix(sshd:session): session closed for user core May 27 03:28:35.129867 systemd[1]: sshd@21-172.31.17.111:22-139.178.68.195:58740.service: Deactivated successfully. May 27 03:28:35.132418 systemd[1]: session-22.scope: Deactivated successfully. May 27 03:28:35.134399 systemd-logind[1875]: Session 22 logged out. Waiting for processes to exit. May 27 03:28:35.136201 systemd-logind[1875]: Removed session 22. May 27 03:28:35.160717 systemd[1]: Started sshd@22-172.31.17.111:22-139.178.68.195:58748.service - OpenSSH per-connection server daemon (139.178.68.195:58748). May 27 03:28:35.333990 sshd[4745]: Accepted publickey for core from 139.178.68.195 port 58748 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:35.335540 sshd-session[4745]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:35.341267 systemd-logind[1875]: New session 23 of user core. May 27 03:28:35.347549 systemd[1]: Started session-23.scope - Session 23 of User core. May 27 03:28:37.078416 containerd[1895]: time="2025-05-27T03:28:37.078350047Z" level=info msg="StopContainer for \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" with timeout 30 (s)" May 27 03:28:37.087773 containerd[1895]: time="2025-05-27T03:28:37.087466594Z" level=info msg="Stop container \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" with signal terminated" May 27 03:28:37.102396 containerd[1895]: time="2025-05-27T03:28:37.102142506Z" level=info msg="TaskExit event in podsandbox handler container_id:\"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" id:\"f166e90f3341a4c11a35094e856a9385b44e0444b4967761eba17b10b6d6b283\" pid:4768 exited_at:{seconds:1748316517 nanos:101602601}" May 27 03:28:37.102709 containerd[1895]: time="2025-05-27T03:28:37.102580098Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 27 03:28:37.106953 containerd[1895]: time="2025-05-27T03:28:37.106900438Z" level=info msg="StopContainer for \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" with timeout 2 (s)" May 27 03:28:37.107443 containerd[1895]: time="2025-05-27T03:28:37.107399187Z" level=info msg="Stop container \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" with signal terminated" May 27 03:28:37.119134 systemd[1]: cri-containerd-ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18.scope: Deactivated successfully. May 27 03:28:37.119997 systemd[1]: cri-containerd-ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18.scope: Consumed 418ms CPU time, 36.4M memory peak, 15.7M read from disk, 4K written to disk. May 27 03:28:37.123743 containerd[1895]: time="2025-05-27T03:28:37.123578818Z" level=info msg="received exit event container_id:\"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" id:\"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" pid:3546 exited_at:{seconds:1748316517 nanos:122795164}" May 27 03:28:37.124273 containerd[1895]: time="2025-05-27T03:28:37.124240249Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" id:\"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" pid:3546 exited_at:{seconds:1748316517 nanos:122795164}" May 27 03:28:37.127533 systemd-networkd[1825]: lxc_health: Link DOWN May 27 03:28:37.127541 systemd-networkd[1825]: lxc_health: Lost carrier May 27 03:28:37.152121 systemd[1]: cri-containerd-335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd.scope: Deactivated successfully. May 27 03:28:37.152524 systemd[1]: cri-containerd-335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd.scope: Consumed 6.876s CPU time, 188.1M memory peak, 74.8M read from disk, 13.3M written to disk. May 27 03:28:37.154913 containerd[1895]: time="2025-05-27T03:28:37.154665712Z" level=info msg="TaskExit event in podsandbox handler container_id:\"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" id:\"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" pid:4052 exited_at:{seconds:1748316517 nanos:154253434}" May 27 03:28:37.154913 containerd[1895]: time="2025-05-27T03:28:37.154802259Z" level=info msg="received exit event container_id:\"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" id:\"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" pid:4052 exited_at:{seconds:1748316517 nanos:154253434}" May 27 03:28:37.186329 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18-rootfs.mount: Deactivated successfully. May 27 03:28:37.202228 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd-rootfs.mount: Deactivated successfully. May 27 03:28:37.231414 containerd[1895]: time="2025-05-27T03:28:37.231358246Z" level=info msg="StopContainer for \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" returns successfully" May 27 03:28:37.231976 containerd[1895]: time="2025-05-27T03:28:37.231945546Z" level=info msg="StopPodSandbox for \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\"" May 27 03:28:37.235577 containerd[1895]: time="2025-05-27T03:28:37.235532516Z" level=info msg="StopContainer for \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" returns successfully" May 27 03:28:37.236502 containerd[1895]: time="2025-05-27T03:28:37.236454215Z" level=info msg="StopPodSandbox for \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\"" May 27 03:28:37.239457 containerd[1895]: time="2025-05-27T03:28:37.239306830Z" level=info msg="Container to stop \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 27 03:28:37.239457 containerd[1895]: time="2025-05-27T03:28:37.239419958Z" level=info msg="Container to stop \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 27 03:28:37.239457 containerd[1895]: time="2025-05-27T03:28:37.239430205Z" level=info msg="Container to stop \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 27 03:28:37.239457 containerd[1895]: time="2025-05-27T03:28:37.239438744Z" level=info msg="Container to stop \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 27 03:28:37.239457 containerd[1895]: time="2025-05-27T03:28:37.239447139Z" level=info msg="Container to stop \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 27 03:28:37.242429 containerd[1895]: time="2025-05-27T03:28:37.242390781Z" level=info msg="Container to stop \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 27 03:28:37.250142 systemd[1]: cri-containerd-d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca.scope: Deactivated successfully. May 27 03:28:37.258299 containerd[1895]: time="2025-05-27T03:28:37.258259046Z" level=info msg="TaskExit event in podsandbox handler container_id:\"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" id:\"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" pid:3327 exit_status:137 exited_at:{seconds:1748316517 nanos:257902619}" May 27 03:28:37.261086 systemd[1]: cri-containerd-44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f.scope: Deactivated successfully. May 27 03:28:37.306457 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f-rootfs.mount: Deactivated successfully. May 27 03:28:37.311156 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca-rootfs.mount: Deactivated successfully. May 27 03:28:37.325510 containerd[1895]: time="2025-05-27T03:28:37.325410053Z" level=info msg="shim disconnected" id=d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca namespace=k8s.io May 27 03:28:37.325510 containerd[1895]: time="2025-05-27T03:28:37.325438148Z" level=warning msg="cleaning up after shim disconnected" id=d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca namespace=k8s.io May 27 03:28:37.331582 containerd[1895]: time="2025-05-27T03:28:37.325445937Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 27 03:28:37.331687 containerd[1895]: time="2025-05-27T03:28:37.326028566Z" level=info msg="shim disconnected" id=44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f namespace=k8s.io May 27 03:28:37.331687 containerd[1895]: time="2025-05-27T03:28:37.331640760Z" level=warning msg="cleaning up after shim disconnected" id=44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f namespace=k8s.io May 27 03:28:37.331687 containerd[1895]: time="2025-05-27T03:28:37.331647706Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 27 03:28:37.388344 containerd[1895]: time="2025-05-27T03:28:37.386684850Z" level=info msg="TaskExit event in podsandbox handler container_id:\"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" id:\"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" pid:3250 exit_status:137 exited_at:{seconds:1748316517 nanos:267745220}" May 27 03:28:37.391574 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f-shm.mount: Deactivated successfully. May 27 03:28:37.401558 containerd[1895]: time="2025-05-27T03:28:37.401508409Z" level=info msg="TearDown network for sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" successfully" May 27 03:28:37.402591 containerd[1895]: time="2025-05-27T03:28:37.402554189Z" level=info msg="StopPodSandbox for \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" returns successfully" May 27 03:28:37.403850 containerd[1895]: time="2025-05-27T03:28:37.403804615Z" level=info msg="received exit event sandbox_id:\"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" exit_status:137 exited_at:{seconds:1748316517 nanos:267745220}" May 27 03:28:37.404349 containerd[1895]: time="2025-05-27T03:28:37.404282639Z" level=info msg="received exit event sandbox_id:\"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" exit_status:137 exited_at:{seconds:1748316517 nanos:257902619}" May 27 03:28:37.408330 containerd[1895]: time="2025-05-27T03:28:37.407666222Z" level=info msg="TearDown network for sandbox \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" successfully" May 27 03:28:37.408330 containerd[1895]: time="2025-05-27T03:28:37.407733373Z" level=info msg="StopPodSandbox for \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" returns successfully" May 27 03:28:37.542273 kubelet[3142]: I0527 03:28:37.542223 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z8hr\" (UniqueName: \"kubernetes.io/projected/2f588127-76ec-4bcc-aae8-deb799bd22db-kube-api-access-4z8hr\") pod \"2f588127-76ec-4bcc-aae8-deb799bd22db\" (UID: \"2f588127-76ec-4bcc-aae8-deb799bd22db\") " May 27 03:28:37.542273 kubelet[3142]: I0527 03:28:37.542265 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-config-path\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.542885 kubelet[3142]: I0527 03:28:37.542289 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-run\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.542885 kubelet[3142]: I0527 03:28:37.542306 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hostproc\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.542885 kubelet[3142]: I0527 03:28:37.542334 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cni-path\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.542885 kubelet[3142]: I0527 03:28:37.542349 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-cgroup\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.542885 kubelet[3142]: I0527 03:28:37.542365 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hubble-tls\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.542885 kubelet[3142]: I0527 03:28:37.542381 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-xtables-lock\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543234 kubelet[3142]: I0527 03:28:37.542396 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-clustermesh-secrets\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543234 kubelet[3142]: I0527 03:28:37.542414 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxszl\" (UniqueName: \"kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-kube-api-access-hxszl\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543234 kubelet[3142]: I0527 03:28:37.542432 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-etc-cni-netd\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543234 kubelet[3142]: I0527 03:28:37.542447 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-lib-modules\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543234 kubelet[3142]: I0527 03:28:37.542463 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2f588127-76ec-4bcc-aae8-deb799bd22db-cilium-config-path\") pod \"2f588127-76ec-4bcc-aae8-deb799bd22db\" (UID: \"2f588127-76ec-4bcc-aae8-deb799bd22db\") " May 27 03:28:37.543234 kubelet[3142]: I0527 03:28:37.542479 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-net\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543430 kubelet[3142]: I0527 03:28:37.542493 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-bpf-maps\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543430 kubelet[3142]: I0527 03:28:37.542506 3142 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-kernel\") pod \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\" (UID: \"72b54b69-33bf-4546-96cd-e9cdc1ab2ae6\") " May 27 03:28:37.543430 kubelet[3142]: I0527 03:28:37.542573 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.543430 kubelet[3142]: I0527 03:28:37.542809 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.548414 kubelet[3142]: I0527 03:28:37.548373 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 27 03:28:37.548882 kubelet[3142]: I0527 03:28:37.548596 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.548882 kubelet[3142]: I0527 03:28:37.548620 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hostproc" (OuterVolumeSpecName: "hostproc") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.548882 kubelet[3142]: I0527 03:28:37.548634 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cni-path" (OuterVolumeSpecName: "cni-path") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.548882 kubelet[3142]: I0527 03:28:37.548648 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.550339 kubelet[3142]: I0527 03:28:37.550187 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f588127-76ec-4bcc-aae8-deb799bd22db-kube-api-access-4z8hr" (OuterVolumeSpecName: "kube-api-access-4z8hr") pod "2f588127-76ec-4bcc-aae8-deb799bd22db" (UID: "2f588127-76ec-4bcc-aae8-deb799bd22db"). InnerVolumeSpecName "kube-api-access-4z8hr". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 27 03:28:37.551520 kubelet[3142]: I0527 03:28:37.551490 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 27 03:28:37.551597 kubelet[3142]: I0527 03:28:37.551539 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.554108 kubelet[3142]: I0527 03:28:37.554070 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 27 03:28:37.555438 kubelet[3142]: I0527 03:28:37.555396 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-kube-api-access-hxszl" (OuterVolumeSpecName: "kube-api-access-hxszl") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "kube-api-access-hxszl". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 27 03:28:37.555514 kubelet[3142]: I0527 03:28:37.555459 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.555514 kubelet[3142]: I0527 03:28:37.555480 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.555514 kubelet[3142]: I0527 03:28:37.555498 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" (UID: "72b54b69-33bf-4546-96cd-e9cdc1ab2ae6"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 27 03:28:37.556541 kubelet[3142]: I0527 03:28:37.556495 3142 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f588127-76ec-4bcc-aae8-deb799bd22db-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "2f588127-76ec-4bcc-aae8-deb799bd22db" (UID: "2f588127-76ec-4bcc-aae8-deb799bd22db"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643178 3142 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hostproc\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643217 3142 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cni-path\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643226 3142 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-cgroup\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643239 3142 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-hubble-tls\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643249 3142 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-xtables-lock\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643258 3142 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-clustermesh-secrets\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643266 3142 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hxszl\" (UniqueName: \"kubernetes.io/projected/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-kube-api-access-hxszl\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643307 kubelet[3142]: I0527 03:28:37.643274 3142 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-etc-cni-netd\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643284 3142 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-lib-modules\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643294 3142 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2f588127-76ec-4bcc-aae8-deb799bd22db-cilium-config-path\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643302 3142 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-net\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643310 3142 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-bpf-maps\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643370 3142 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-host-proc-sys-kernel\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643379 3142 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4z8hr\" (UniqueName: \"kubernetes.io/projected/2f588127-76ec-4bcc-aae8-deb799bd22db-kube-api-access-4z8hr\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643387 3142 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-config-path\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.643674 kubelet[3142]: I0527 03:28:37.643395 3142 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6-cilium-run\") on node \"ip-172-31-17-111\" DevicePath \"\"" May 27 03:28:37.714522 kubelet[3142]: I0527 03:28:37.714430 3142 scope.go:117] "RemoveContainer" containerID="ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18" May 27 03:28:37.719259 containerd[1895]: time="2025-05-27T03:28:37.719208046Z" level=info msg="RemoveContainer for \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\"" May 27 03:28:37.722808 systemd[1]: Removed slice kubepods-besteffort-pod2f588127_76ec_4bcc_aae8_deb799bd22db.slice - libcontainer container kubepods-besteffort-pod2f588127_76ec_4bcc_aae8_deb799bd22db.slice. May 27 03:28:37.723461 systemd[1]: kubepods-besteffort-pod2f588127_76ec_4bcc_aae8_deb799bd22db.slice: Consumed 451ms CPU time, 36.6M memory peak, 15.7M read from disk, 4K written to disk. May 27 03:28:37.731176 systemd[1]: Removed slice kubepods-burstable-pod72b54b69_33bf_4546_96cd_e9cdc1ab2ae6.slice - libcontainer container kubepods-burstable-pod72b54b69_33bf_4546_96cd_e9cdc1ab2ae6.slice. May 27 03:28:37.731635 systemd[1]: kubepods-burstable-pod72b54b69_33bf_4546_96cd_e9cdc1ab2ae6.slice: Consumed 6.976s CPU time, 188.5M memory peak, 76.6M read from disk, 13.3M written to disk. May 27 03:28:37.733762 containerd[1895]: time="2025-05-27T03:28:37.733729018Z" level=info msg="RemoveContainer for \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" returns successfully" May 27 03:28:37.734367 kubelet[3142]: I0527 03:28:37.734297 3142 scope.go:117] "RemoveContainer" containerID="ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18" May 27 03:28:37.745879 containerd[1895]: time="2025-05-27T03:28:37.734892819Z" level=error msg="ContainerStatus for \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\": not found" May 27 03:28:37.746508 kubelet[3142]: E0527 03:28:37.746299 3142 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\": not found" containerID="ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18" May 27 03:28:37.746749 kubelet[3142]: I0527 03:28:37.746476 3142 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18"} err="failed to get container status \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\": rpc error: code = NotFound desc = an error occurred when try to find container \"ed55081720558001f572875770585224711fba9b48ece583b01bb3806a60be18\": not found" May 27 03:28:37.746821 kubelet[3142]: I0527 03:28:37.746751 3142 scope.go:117] "RemoveContainer" containerID="335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd" May 27 03:28:37.749492 containerd[1895]: time="2025-05-27T03:28:37.749453905Z" level=info msg="RemoveContainer for \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\"" May 27 03:28:37.756685 containerd[1895]: time="2025-05-27T03:28:37.756638175Z" level=info msg="RemoveContainer for \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" returns successfully" May 27 03:28:37.757037 kubelet[3142]: I0527 03:28:37.757008 3142 scope.go:117] "RemoveContainer" containerID="cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d" May 27 03:28:37.759391 containerd[1895]: time="2025-05-27T03:28:37.759160709Z" level=info msg="RemoveContainer for \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\"" May 27 03:28:37.767702 containerd[1895]: time="2025-05-27T03:28:37.767652050Z" level=info msg="RemoveContainer for \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" returns successfully" May 27 03:28:37.767914 kubelet[3142]: I0527 03:28:37.767847 3142 scope.go:117] "RemoveContainer" containerID="700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e" May 27 03:28:37.770072 containerd[1895]: time="2025-05-27T03:28:37.770037099Z" level=info msg="RemoveContainer for \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\"" May 27 03:28:37.776174 containerd[1895]: time="2025-05-27T03:28:37.776134467Z" level=info msg="RemoveContainer for \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" returns successfully" May 27 03:28:37.776472 kubelet[3142]: I0527 03:28:37.776385 3142 scope.go:117] "RemoveContainer" containerID="07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d" May 27 03:28:37.777980 containerd[1895]: time="2025-05-27T03:28:37.777949619Z" level=info msg="RemoveContainer for \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\"" May 27 03:28:37.783594 containerd[1895]: time="2025-05-27T03:28:37.783554768Z" level=info msg="RemoveContainer for \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" returns successfully" May 27 03:28:37.784011 kubelet[3142]: I0527 03:28:37.783978 3142 scope.go:117] "RemoveContainer" containerID="5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b" May 27 03:28:37.785685 containerd[1895]: time="2025-05-27T03:28:37.785642186Z" level=info msg="RemoveContainer for \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\"" May 27 03:28:37.790915 containerd[1895]: time="2025-05-27T03:28:37.790876976Z" level=info msg="RemoveContainer for \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" returns successfully" May 27 03:28:37.791446 kubelet[3142]: I0527 03:28:37.791307 3142 scope.go:117] "RemoveContainer" containerID="335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd" May 27 03:28:37.791727 containerd[1895]: time="2025-05-27T03:28:37.791695274Z" level=error msg="ContainerStatus for \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\": not found" May 27 03:28:37.791970 kubelet[3142]: E0527 03:28:37.791941 3142 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\": not found" containerID="335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd" May 27 03:28:37.792044 kubelet[3142]: I0527 03:28:37.791973 3142 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd"} err="failed to get container status \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\": rpc error: code = NotFound desc = an error occurred when try to find container \"335264dd9ee4207f1cedba83a7d0368fa079f4e807ba50e0683a147597d26bdd\": not found" May 27 03:28:37.792044 kubelet[3142]: I0527 03:28:37.792000 3142 scope.go:117] "RemoveContainer" containerID="cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d" May 27 03:28:37.792234 containerd[1895]: time="2025-05-27T03:28:37.792205438Z" level=error msg="ContainerStatus for \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\": not found" May 27 03:28:37.792404 kubelet[3142]: E0527 03:28:37.792381 3142 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\": not found" containerID="cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d" May 27 03:28:37.792484 kubelet[3142]: I0527 03:28:37.792406 3142 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d"} err="failed to get container status \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\": rpc error: code = NotFound desc = an error occurred when try to find container \"cac85f938977db93cdeb6af0f642ab04c164cb1bb316e5a45583075130ffc12d\": not found" May 27 03:28:37.792484 kubelet[3142]: I0527 03:28:37.792427 3142 scope.go:117] "RemoveContainer" containerID="700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e" May 27 03:28:37.792625 containerd[1895]: time="2025-05-27T03:28:37.792600161Z" level=error msg="ContainerStatus for \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\": not found" May 27 03:28:37.792766 kubelet[3142]: E0527 03:28:37.792716 3142 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\": not found" containerID="700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e" May 27 03:28:37.792766 kubelet[3142]: I0527 03:28:37.792741 3142 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e"} err="failed to get container status \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\": rpc error: code = NotFound desc = an error occurred when try to find container \"700b98cc7e2f1dcce4ed220909f1e7e1a853e8d4d4376fd45bc3535e7540d42e\": not found" May 27 03:28:37.792766 kubelet[3142]: I0527 03:28:37.792763 3142 scope.go:117] "RemoveContainer" containerID="07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d" May 27 03:28:37.793003 containerd[1895]: time="2025-05-27T03:28:37.792968485Z" level=error msg="ContainerStatus for \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\": not found" May 27 03:28:37.793119 kubelet[3142]: E0527 03:28:37.793092 3142 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\": not found" containerID="07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d" May 27 03:28:37.793197 kubelet[3142]: I0527 03:28:37.793122 3142 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d"} err="failed to get container status \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\": rpc error: code = NotFound desc = an error occurred when try to find container \"07c69ca0f7fff41c5205ef25e4e615d97a13b7e9c14f87937874a3acaeba218d\": not found" May 27 03:28:37.793197 kubelet[3142]: I0527 03:28:37.793142 3142 scope.go:117] "RemoveContainer" containerID="5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b" May 27 03:28:37.793445 containerd[1895]: time="2025-05-27T03:28:37.793393682Z" level=error msg="ContainerStatus for \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\": not found" May 27 03:28:37.793591 kubelet[3142]: E0527 03:28:37.793565 3142 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\": not found" containerID="5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b" May 27 03:28:37.793657 kubelet[3142]: I0527 03:28:37.793617 3142 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b"} err="failed to get container status \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\": rpc error: code = NotFound desc = an error occurred when try to find container \"5c39a248a62349eda67125480f88ffacadc37f4cdf66a74446c853c7a9e0619b\": not found" May 27 03:28:38.184463 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca-shm.mount: Deactivated successfully. May 27 03:28:38.184871 systemd[1]: var-lib-kubelet-pods-72b54b69\x2d33bf\x2d4546\x2d96cd\x2de9cdc1ab2ae6-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dhxszl.mount: Deactivated successfully. May 27 03:28:38.184952 systemd[1]: var-lib-kubelet-pods-2f588127\x2d76ec\x2d4bcc\x2daae8\x2ddeb799bd22db-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4z8hr.mount: Deactivated successfully. May 27 03:28:38.185009 systemd[1]: var-lib-kubelet-pods-72b54b69\x2d33bf\x2d4546\x2d96cd\x2de9cdc1ab2ae6-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 27 03:28:38.185069 systemd[1]: var-lib-kubelet-pods-72b54b69\x2d33bf\x2d4546\x2d96cd\x2de9cdc1ab2ae6-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 27 03:28:39.013918 sshd[4748]: Connection closed by 139.178.68.195 port 58748 May 27 03:28:39.016179 sshd-session[4745]: pam_unix(sshd:session): session closed for user core May 27 03:28:39.020553 systemd-logind[1875]: Session 23 logged out. Waiting for processes to exit. May 27 03:28:39.021227 systemd[1]: sshd@22-172.31.17.111:22-139.178.68.195:58748.service: Deactivated successfully. May 27 03:28:39.024109 systemd[1]: session-23.scope: Deactivated successfully. May 27 03:28:39.026510 systemd-logind[1875]: Removed session 23. May 27 03:28:39.046746 systemd[1]: Started sshd@23-172.31.17.111:22-139.178.68.195:58756.service - OpenSSH per-connection server daemon (139.178.68.195:58756). May 27 03:28:39.237304 sshd[4898]: Accepted publickey for core from 139.178.68.195 port 58756 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:39.239059 sshd-session[4898]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:39.245373 systemd-logind[1875]: New session 24 of user core. May 27 03:28:39.253608 systemd[1]: Started session-24.scope - Session 24 of User core. May 27 03:28:39.423430 kubelet[3142]: I0527 03:28:39.422334 3142 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f588127-76ec-4bcc-aae8-deb799bd22db" path="/var/lib/kubelet/pods/2f588127-76ec-4bcc-aae8-deb799bd22db/volumes" May 27 03:28:39.423430 kubelet[3142]: I0527 03:28:39.422725 3142 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b54b69-33bf-4546-96cd-e9cdc1ab2ae6" path="/var/lib/kubelet/pods/72b54b69-33bf-4546-96cd-e9cdc1ab2ae6/volumes" May 27 03:28:39.644591 ntpd[1869]: Deleting interface #10 lxc_health, fe80::d86f:bdff:fe2f:6110%8#123, interface stats: received=0, sent=0, dropped=0, active_time=57 secs May 27 03:28:39.645021 ntpd[1869]: 27 May 03:28:39 ntpd[1869]: Deleting interface #10 lxc_health, fe80::d86f:bdff:fe2f:6110%8#123, interface stats: received=0, sent=0, dropped=0, active_time=57 secs May 27 03:28:39.979251 sshd[4900]: Connection closed by 139.178.68.195 port 58756 May 27 03:28:39.979639 sshd-session[4898]: pam_unix(sshd:session): session closed for user core May 27 03:28:39.984530 systemd[1]: sshd@23-172.31.17.111:22-139.178.68.195:58756.service: Deactivated successfully. May 27 03:28:39.989719 systemd[1]: session-24.scope: Deactivated successfully. May 27 03:28:39.994584 systemd-logind[1875]: Session 24 logged out. Waiting for processes to exit. May 27 03:28:39.996982 systemd-logind[1875]: Removed session 24. May 27 03:28:40.022638 systemd[1]: Started sshd@24-172.31.17.111:22-139.178.68.195:58772.service - OpenSSH per-connection server daemon (139.178.68.195:58772). May 27 03:28:40.049326 systemd[1]: Created slice kubepods-burstable-podf7d14abb_670f_4189_9f9e_67ef12df7fe8.slice - libcontainer container kubepods-burstable-podf7d14abb_670f_4189_9f9e_67ef12df7fe8.slice. May 27 03:28:40.059711 kubelet[3142]: I0527 03:28:40.059682 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-cni-path\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.060082 kubelet[3142]: I0527 03:28:40.059921 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-xtables-lock\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.061923 systemd[1]: Created slice kubepods-besteffort-podf4b71616_842a_4973_b261_fadaeac96cb8.slice - libcontainer container kubepods-besteffort-podf4b71616_842a_4973_b261_fadaeac96cb8.slice. May 27 03:28:40.062473 kubelet[3142]: I0527 03:28:40.062002 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-host-proc-sys-net\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.062473 kubelet[3142]: I0527 03:28:40.062203 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-cilium-run\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.062681 kubelet[3142]: I0527 03:28:40.062628 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-bpf-maps\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.065882 kubelet[3142]: I0527 03:28:40.062666 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f7d14abb-670f-4189-9f9e-67ef12df7fe8-hubble-tls\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069581 kubelet[3142]: I0527 03:28:40.067507 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f7d14abb-670f-4189-9f9e-67ef12df7fe8-clustermesh-secrets\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069581 kubelet[3142]: I0527 03:28:40.067660 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhmdz\" (UniqueName: \"kubernetes.io/projected/f7d14abb-670f-4189-9f9e-67ef12df7fe8-kube-api-access-hhmdz\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069581 kubelet[3142]: I0527 03:28:40.067697 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f7d14abb-670f-4189-9f9e-67ef12df7fe8-cilium-config-path\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069581 kubelet[3142]: I0527 03:28:40.067720 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f4b71616-842a-4973-b261-fadaeac96cb8-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-m6qb7\" (UID: \"f4b71616-842a-4973-b261-fadaeac96cb8\") " pod="kube-system/cilium-operator-6c4d7847fc-m6qb7" May 27 03:28:40.069581 kubelet[3142]: I0527 03:28:40.067843 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-lib-modules\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069882 kubelet[3142]: I0527 03:28:40.067866 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-host-proc-sys-kernel\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069882 kubelet[3142]: I0527 03:28:40.069384 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-etc-cni-netd\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069882 kubelet[3142]: I0527 03:28:40.069518 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/f7d14abb-670f-4189-9f9e-67ef12df7fe8-cilium-ipsec-secrets\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.069882 kubelet[3142]: I0527 03:28:40.069552 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qgsp\" (UniqueName: \"kubernetes.io/projected/f4b71616-842a-4973-b261-fadaeac96cb8-kube-api-access-2qgsp\") pod \"cilium-operator-6c4d7847fc-m6qb7\" (UID: \"f4b71616-842a-4973-b261-fadaeac96cb8\") " pod="kube-system/cilium-operator-6c4d7847fc-m6qb7" May 27 03:28:40.074202 kubelet[3142]: I0527 03:28:40.072843 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-hostproc\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.074202 kubelet[3142]: I0527 03:28:40.072888 3142 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f7d14abb-670f-4189-9f9e-67ef12df7fe8-cilium-cgroup\") pod \"cilium-c2n78\" (UID: \"f7d14abb-670f-4189-9f9e-67ef12df7fe8\") " pod="kube-system/cilium-c2n78" May 27 03:28:40.251519 sshd[4911]: Accepted publickey for core from 139.178.68.195 port 58772 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:40.254085 sshd-session[4911]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:40.262905 systemd-logind[1875]: New session 25 of user core. May 27 03:28:40.271557 systemd[1]: Started session-25.scope - Session 25 of User core. May 27 03:28:40.358624 containerd[1895]: time="2025-05-27T03:28:40.358576427Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c2n78,Uid:f7d14abb-670f-4189-9f9e-67ef12df7fe8,Namespace:kube-system,Attempt:0,}" May 27 03:28:40.371827 containerd[1895]: time="2025-05-27T03:28:40.371783283Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-m6qb7,Uid:f4b71616-842a-4973-b261-fadaeac96cb8,Namespace:kube-system,Attempt:0,}" May 27 03:28:40.385256 sshd[4918]: Connection closed by 139.178.68.195 port 58772 May 27 03:28:40.385855 sshd-session[4911]: pam_unix(sshd:session): session closed for user core May 27 03:28:40.392719 systemd[1]: sshd@24-172.31.17.111:22-139.178.68.195:58772.service: Deactivated successfully. May 27 03:28:40.395579 systemd[1]: session-25.scope: Deactivated successfully. May 27 03:28:40.398098 containerd[1895]: time="2025-05-27T03:28:40.397555174Z" level=info msg="connecting to shim 7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f" address="unix:///run/containerd/s/b8449c8471db36701d96b303a38dd5efe970695ffa4d4b29c7d588d5a355f521" namespace=k8s.io protocol=ttrpc version=3 May 27 03:28:40.397738 systemd-logind[1875]: Session 25 logged out. Waiting for processes to exit. May 27 03:28:40.399601 systemd-logind[1875]: Removed session 25. May 27 03:28:40.413873 containerd[1895]: time="2025-05-27T03:28:40.413836463Z" level=info msg="connecting to shim 9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2" address="unix:///run/containerd/s/97d798f72db9026dc387eef7ff1de28a97b7e1221a77782d94b6b74b1cb231b1" namespace=k8s.io protocol=ttrpc version=3 May 27 03:28:40.418856 systemd[1]: Started sshd@25-172.31.17.111:22-139.178.68.195:58782.service - OpenSSH per-connection server daemon (139.178.68.195:58782). May 27 03:28:40.431442 systemd[1]: Started cri-containerd-7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f.scope - libcontainer container 7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f. May 27 03:28:40.459429 systemd[1]: Started cri-containerd-9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2.scope - libcontainer container 9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2. May 27 03:28:40.521655 containerd[1895]: time="2025-05-27T03:28:40.521544325Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-c2n78,Uid:f7d14abb-670f-4189-9f9e-67ef12df7fe8,Namespace:kube-system,Attempt:0,} returns sandbox id \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\"" May 27 03:28:40.537441 containerd[1895]: time="2025-05-27T03:28:40.536485037Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 27 03:28:40.560134 containerd[1895]: time="2025-05-27T03:28:40.560095350Z" level=info msg="Container 3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073: CDI devices from CRI Config.CDIDevices: []" May 27 03:28:40.574732 containerd[1895]: time="2025-05-27T03:28:40.574585727Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073\"" May 27 03:28:40.576596 containerd[1895]: time="2025-05-27T03:28:40.576557704Z" level=info msg="StartContainer for \"3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073\"" May 27 03:28:40.576959 containerd[1895]: time="2025-05-27T03:28:40.576783756Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-m6qb7,Uid:f4b71616-842a-4973-b261-fadaeac96cb8,Namespace:kube-system,Attempt:0,} returns sandbox id \"9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2\"" May 27 03:28:40.579034 containerd[1895]: time="2025-05-27T03:28:40.578998509Z" level=info msg="connecting to shim 3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073" address="unix:///run/containerd/s/b8449c8471db36701d96b303a38dd5efe970695ffa4d4b29c7d588d5a355f521" protocol=ttrpc version=3 May 27 03:28:40.586978 containerd[1895]: time="2025-05-27T03:28:40.586939954Z" level=info msg="CreateContainer within sandbox \"9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 27 03:28:40.604253 containerd[1895]: time="2025-05-27T03:28:40.603758202Z" level=info msg="Container 8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d: CDI devices from CRI Config.CDIDevices: []" May 27 03:28:40.608937 sshd[4968]: Accepted publickey for core from 139.178.68.195 port 58782 ssh2: RSA SHA256:Uw58Bn7G+SJd5XoMf+3ukvYab1VfQ8PtnN9pHyXmnUI May 27 03:28:40.609755 systemd[1]: Started cri-containerd-3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073.scope - libcontainer container 3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073. May 27 03:28:40.610883 sshd-session[4968]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 27 03:28:40.621135 containerd[1895]: time="2025-05-27T03:28:40.619439452Z" level=info msg="CreateContainer within sandbox \"9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d\"" May 27 03:28:40.621135 containerd[1895]: time="2025-05-27T03:28:40.620752480Z" level=info msg="StartContainer for \"8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d\"" May 27 03:28:40.628513 containerd[1895]: time="2025-05-27T03:28:40.625083527Z" level=info msg="connecting to shim 8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d" address="unix:///run/containerd/s/97d798f72db9026dc387eef7ff1de28a97b7e1221a77782d94b6b74b1cb231b1" protocol=ttrpc version=3 May 27 03:28:40.628693 systemd-logind[1875]: New session 26 of user core. May 27 03:28:40.633573 systemd[1]: Started session-26.scope - Session 26 of User core. May 27 03:28:40.655527 systemd[1]: Started cri-containerd-8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d.scope - libcontainer container 8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d. May 27 03:28:40.689349 containerd[1895]: time="2025-05-27T03:28:40.688657216Z" level=info msg="StartContainer for \"3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073\" returns successfully" May 27 03:28:40.707284 systemd[1]: cri-containerd-3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073.scope: Deactivated successfully. May 27 03:28:40.707651 systemd[1]: cri-containerd-3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073.scope: Consumed 28ms CPU time, 9.7M memory peak, 3.3M read from disk. May 27 03:28:40.713551 containerd[1895]: time="2025-05-27T03:28:40.713516083Z" level=info msg="StartContainer for \"8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d\" returns successfully" May 27 03:28:40.715587 containerd[1895]: time="2025-05-27T03:28:40.715551456Z" level=info msg="received exit event container_id:\"3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073\" id:\"3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073\" pid:5030 exited_at:{seconds:1748316520 nanos:713530692}" May 27 03:28:40.715881 containerd[1895]: time="2025-05-27T03:28:40.715820879Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073\" id:\"3252067bdc6b107b086d4cc16e61219cd1b432f39c33bc2d2ed7d76896ffb073\" pid:5030 exited_at:{seconds:1748316520 nanos:713530692}" May 27 03:28:40.825566 kubelet[3142]: I0527 03:28:40.823404 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-m6qb7" podStartSLOduration=1.823382157 podStartE2EDuration="1.823382157s" podCreationTimestamp="2025-05-27 03:28:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:28:40.780939834 +0000 UTC m=+89.477379329" watchObservedRunningTime="2025-05-27 03:28:40.823382157 +0000 UTC m=+89.519821650" May 27 03:28:41.522391 kubelet[3142]: E0527 03:28:41.522342 3142 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 27 03:28:41.775444 containerd[1895]: time="2025-05-27T03:28:41.775201182Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 27 03:28:41.790189 containerd[1895]: time="2025-05-27T03:28:41.790134484Z" level=info msg="Container 499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d: CDI devices from CRI Config.CDIDevices: []" May 27 03:28:41.805041 containerd[1895]: time="2025-05-27T03:28:41.804965358Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d\"" May 27 03:28:41.807455 containerd[1895]: time="2025-05-27T03:28:41.806409833Z" level=info msg="StartContainer for \"499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d\"" May 27 03:28:41.808263 containerd[1895]: time="2025-05-27T03:28:41.808209326Z" level=info msg="connecting to shim 499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d" address="unix:///run/containerd/s/b8449c8471db36701d96b303a38dd5efe970695ffa4d4b29c7d588d5a355f521" protocol=ttrpc version=3 May 27 03:28:41.846522 systemd[1]: Started cri-containerd-499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d.scope - libcontainer container 499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d. May 27 03:28:41.887580 containerd[1895]: time="2025-05-27T03:28:41.887514315Z" level=info msg="StartContainer for \"499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d\" returns successfully" May 27 03:28:41.920243 systemd[1]: cri-containerd-499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d.scope: Deactivated successfully. May 27 03:28:41.920962 systemd[1]: cri-containerd-499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d.scope: Consumed 22ms CPU time, 7.4M memory peak, 2.2M read from disk. May 27 03:28:41.927118 containerd[1895]: time="2025-05-27T03:28:41.927069040Z" level=info msg="TaskExit event in podsandbox handler container_id:\"499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d\" id:\"499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d\" pid:5114 exited_at:{seconds:1748316521 nanos:925762560}" May 27 03:28:41.927510 containerd[1895]: time="2025-05-27T03:28:41.927439161Z" level=info msg="received exit event container_id:\"499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d\" id:\"499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d\" pid:5114 exited_at:{seconds:1748316521 nanos:925762560}" May 27 03:28:41.950587 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-499ded030d1f93935b10641e81abba1774fc186dc01271147c75d1b0680c738d-rootfs.mount: Deactivated successfully. May 27 03:28:41.987855 kubelet[3142]: I0527 03:28:41.987811 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:28:41.987855 kubelet[3142]: I0527 03:28:41.987847 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:28:41.990680 containerd[1895]: time="2025-05-27T03:28:41.990649336Z" level=info msg="StopPodSandbox for \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\"" May 27 03:28:41.990814 containerd[1895]: time="2025-05-27T03:28:41.990772857Z" level=info msg="TearDown network for sandbox \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" successfully" May 27 03:28:41.990814 containerd[1895]: time="2025-05-27T03:28:41.990783458Z" level=info msg="StopPodSandbox for \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" returns successfully" May 27 03:28:41.991372 containerd[1895]: time="2025-05-27T03:28:41.991240921Z" level=info msg="RemovePodSandbox for \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\"" May 27 03:28:41.991372 containerd[1895]: time="2025-05-27T03:28:41.991280078Z" level=info msg="Forcibly stopping sandbox \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\"" May 27 03:28:41.991473 containerd[1895]: time="2025-05-27T03:28:41.991448155Z" level=info msg="TearDown network for sandbox \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" successfully" May 27 03:28:41.994666 containerd[1895]: time="2025-05-27T03:28:41.994631321Z" level=info msg="Ensure that sandbox 44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f in task-service has been cleanup successfully" May 27 03:28:42.002201 containerd[1895]: time="2025-05-27T03:28:42.001195440Z" level=info msg="RemovePodSandbox \"44320f1fcb88d9a48fede0350d55b710fe97106819958bb75c2d3c91a871499f\" returns successfully" May 27 03:28:42.019467 containerd[1895]: time="2025-05-27T03:28:42.019417306Z" level=info msg="StopPodSandbox for \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\"" May 27 03:28:42.019638 containerd[1895]: time="2025-05-27T03:28:42.019574345Z" level=info msg="TearDown network for sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" successfully" May 27 03:28:42.019638 containerd[1895]: time="2025-05-27T03:28:42.019592600Z" level=info msg="StopPodSandbox for \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" returns successfully" May 27 03:28:42.020044 containerd[1895]: time="2025-05-27T03:28:42.020014410Z" level=info msg="RemovePodSandbox for \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\"" May 27 03:28:42.020144 containerd[1895]: time="2025-05-27T03:28:42.020046547Z" level=info msg="Forcibly stopping sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\"" May 27 03:28:42.020184 containerd[1895]: time="2025-05-27T03:28:42.020148220Z" level=info msg="TearDown network for sandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" successfully" May 27 03:28:42.021495 containerd[1895]: time="2025-05-27T03:28:42.021457834Z" level=info msg="Ensure that sandbox d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca in task-service has been cleanup successfully" May 27 03:28:42.028499 containerd[1895]: time="2025-05-27T03:28:42.027581142Z" level=info msg="RemovePodSandbox \"d07d54a3ceebd58e302610fc7178713f3354544ccd0a32b0ac81e44d1d7651ca\" returns successfully" May 27 03:28:42.028599 kubelet[3142]: I0527 03:28:42.028459 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:28:42.038463 kubelet[3142]: I0527 03:28:42.038422 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:28:42.038615 kubelet[3142]: I0527 03:28:42.038497 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-c2n78","kube-system/cilium-operator-6c4d7847fc-m6qb7","kube-system/kube-proxy-k7s89","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:28:42.038615 kubelet[3142]: E0527 03:28:42.038527 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-c2n78" May 27 03:28:42.038615 kubelet[3142]: E0527 03:28:42.038540 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-m6qb7" May 27 03:28:42.038615 kubelet[3142]: E0527 03:28:42.038548 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:28:42.038615 kubelet[3142]: E0527 03:28:42.038559 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:28:42.038615 kubelet[3142]: E0527 03:28:42.038566 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:28:42.038615 kubelet[3142]: E0527 03:28:42.038574 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:28:42.038615 kubelet[3142]: I0527 03:28:42.038584 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:28:42.779200 containerd[1895]: time="2025-05-27T03:28:42.779144143Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 27 03:28:42.798590 containerd[1895]: time="2025-05-27T03:28:42.798544833Z" level=info msg="Container da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef: CDI devices from CRI Config.CDIDevices: []" May 27 03:28:42.817200 containerd[1895]: time="2025-05-27T03:28:42.817161012Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef\"" May 27 03:28:42.820549 containerd[1895]: time="2025-05-27T03:28:42.818023340Z" level=info msg="StartContainer for \"da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef\"" May 27 03:28:42.821753 containerd[1895]: time="2025-05-27T03:28:42.821723472Z" level=info msg="connecting to shim da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef" address="unix:///run/containerd/s/b8449c8471db36701d96b303a38dd5efe970695ffa4d4b29c7d588d5a355f521" protocol=ttrpc version=3 May 27 03:28:42.869239 systemd[1]: Started cri-containerd-da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef.scope - libcontainer container da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef. May 27 03:28:42.956685 containerd[1895]: time="2025-05-27T03:28:42.956605309Z" level=info msg="StartContainer for \"da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef\" returns successfully" May 27 03:28:42.964081 systemd[1]: cri-containerd-da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef.scope: Deactivated successfully. May 27 03:28:42.964486 systemd[1]: cri-containerd-da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef.scope: Consumed 35ms CPU time, 5.9M memory peak, 1.1M read from disk. May 27 03:28:42.965674 containerd[1895]: time="2025-05-27T03:28:42.965647300Z" level=info msg="received exit event container_id:\"da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef\" id:\"da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef\" pid:5156 exited_at:{seconds:1748316522 nanos:965082601}" May 27 03:28:42.965820 containerd[1895]: time="2025-05-27T03:28:42.965802731Z" level=info msg="TaskExit event in podsandbox handler container_id:\"da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef\" id:\"da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef\" pid:5156 exited_at:{seconds:1748316522 nanos:965082601}" May 27 03:28:42.991837 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-da839f7962882674233741d24bac9f0bba398814da5bd6f772c2a24da4296bef-rootfs.mount: Deactivated successfully. May 27 03:28:43.784643 containerd[1895]: time="2025-05-27T03:28:43.784597700Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 27 03:28:43.803472 containerd[1895]: time="2025-05-27T03:28:43.800589278Z" level=info msg="Container 0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54: CDI devices from CRI Config.CDIDevices: []" May 27 03:28:43.803585 kubelet[3142]: I0527 03:28:43.800719 3142 setters.go:618] "Node became not ready" node="ip-172-31-17-111" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-27T03:28:43Z","lastTransitionTime":"2025-05-27T03:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 27 03:28:43.820968 containerd[1895]: time="2025-05-27T03:28:43.820919675Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54\"" May 27 03:28:43.823212 containerd[1895]: time="2025-05-27T03:28:43.822883443Z" level=info msg="StartContainer for \"0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54\"" May 27 03:28:43.824084 containerd[1895]: time="2025-05-27T03:28:43.824020149Z" level=info msg="connecting to shim 0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54" address="unix:///run/containerd/s/b8449c8471db36701d96b303a38dd5efe970695ffa4d4b29c7d588d5a355f521" protocol=ttrpc version=3 May 27 03:28:43.855527 systemd[1]: Started cri-containerd-0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54.scope - libcontainer container 0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54. May 27 03:28:43.885759 systemd[1]: cri-containerd-0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54.scope: Deactivated successfully. May 27 03:28:43.887588 containerd[1895]: time="2025-05-27T03:28:43.887539883Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54\" id:\"0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54\" pid:5196 exited_at:{seconds:1748316523 nanos:886836092}" May 27 03:28:43.890810 containerd[1895]: time="2025-05-27T03:28:43.890722063Z" level=info msg="received exit event container_id:\"0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54\" id:\"0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54\" pid:5196 exited_at:{seconds:1748316523 nanos:886836092}" May 27 03:28:43.894457 containerd[1895]: time="2025-05-27T03:28:43.894378273Z" level=info msg="StartContainer for \"0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54\" returns successfully" May 27 03:28:43.938641 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0415ed3e91fa64606407e14be573e43f8405640bee1a78481ab9b9826d4c0e54-rootfs.mount: Deactivated successfully. May 27 03:28:44.797777 containerd[1895]: time="2025-05-27T03:28:44.796129222Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 27 03:28:44.821118 containerd[1895]: time="2025-05-27T03:28:44.821082868Z" level=info msg="Container dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14: CDI devices from CRI Config.CDIDevices: []" May 27 03:28:44.836470 containerd[1895]: time="2025-05-27T03:28:44.836423028Z" level=info msg="CreateContainer within sandbox \"7f78ba94cb3c1e11ea68cf49239b888d3e3a082bba0d83a0199f3b46d6c2839f\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\"" May 27 03:28:44.837337 containerd[1895]: time="2025-05-27T03:28:44.837287576Z" level=info msg="StartContainer for \"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\"" May 27 03:28:44.838794 containerd[1895]: time="2025-05-27T03:28:44.838710902Z" level=info msg="connecting to shim dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14" address="unix:///run/containerd/s/b8449c8471db36701d96b303a38dd5efe970695ffa4d4b29c7d588d5a355f521" protocol=ttrpc version=3 May 27 03:28:44.873573 systemd[1]: Started cri-containerd-dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14.scope - libcontainer container dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14. May 27 03:28:44.920330 containerd[1895]: time="2025-05-27T03:28:44.920267525Z" level=info msg="StartContainer for \"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\" returns successfully" May 27 03:28:45.083598 containerd[1895]: time="2025-05-27T03:28:45.081410483Z" level=info msg="TaskExit event in podsandbox handler container_id:\"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\" id:\"55fbbf520d3db8a52d27da33d7d2752b092212048d89a8f14b3cd6957946fe86\" pid:5263 exited_at:{seconds:1748316525 nanos:81092963}" May 27 03:28:45.608348 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni-avx)) May 27 03:28:47.519675 containerd[1895]: time="2025-05-27T03:28:47.519532210Z" level=info msg="TaskExit event in podsandbox handler container_id:\"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\" id:\"da365cf3e7ec87f428582cce495e35d483366b372eaf69f659e99c89b55dc718\" pid:5423 exit_status:1 exited_at:{seconds:1748316527 nanos:518857585}" May 27 03:28:48.571922 (udev-worker)[5297]: Network interface NamePolicy= disabled on kernel command line. May 27 03:28:48.573303 systemd-networkd[1825]: lxc_health: Link UP May 27 03:28:48.578005 (udev-worker)[5729]: Network interface NamePolicy= disabled on kernel command line. May 27 03:28:48.599293 systemd-networkd[1825]: lxc_health: Gained carrier May 27 03:28:49.786594 containerd[1895]: time="2025-05-27T03:28:49.786457875Z" level=info msg="TaskExit event in podsandbox handler container_id:\"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\" id:\"8f08c8cf2e6b2eb6008660352615176178a89b140ab387ac2a419ed0c8284911\" pid:5772 exited_at:{seconds:1748316529 nanos:785615523}" May 27 03:28:50.338367 systemd-networkd[1825]: lxc_health: Gained IPv6LL May 27 03:28:50.389341 kubelet[3142]: I0527 03:28:50.388689 3142 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-c2n78" podStartSLOduration=11.388669636 podStartE2EDuration="11.388669636s" podCreationTimestamp="2025-05-27 03:28:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-27 03:28:45.814398661 +0000 UTC m=+94.510838156" watchObservedRunningTime="2025-05-27 03:28:50.388669636 +0000 UTC m=+99.085109135" May 27 03:28:51.966064 containerd[1895]: time="2025-05-27T03:28:51.966019117Z" level=info msg="TaskExit event in podsandbox handler container_id:\"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\" id:\"ca86b5b74b5d5376355db9048b2236182c53cc6b7d48ef740500d3a21e0776e3\" pid:5796 exited_at:{seconds:1748316531 nanos:965711183}" May 27 03:28:52.062044 kubelet[3142]: I0527 03:28:52.062007 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:28:52.062500 kubelet[3142]: I0527 03:28:52.062086 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:28:52.067060 kubelet[3142]: I0527 03:28:52.066994 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:28:52.087628 kubelet[3142]: I0527 03:28:52.087594 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:28:52.088354 kubelet[3142]: I0527 03:28:52.087816 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-m6qb7","kube-system/cilium-c2n78","kube-system/kube-proxy-k7s89","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:28:52.088354 kubelet[3142]: E0527 03:28:52.087884 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-m6qb7" May 27 03:28:52.088354 kubelet[3142]: E0527 03:28:52.087904 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-c2n78" May 27 03:28:52.088354 kubelet[3142]: E0527 03:28:52.087920 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:28:52.088354 kubelet[3142]: E0527 03:28:52.087950 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:28:52.088354 kubelet[3142]: E0527 03:28:52.087967 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:28:52.088354 kubelet[3142]: E0527 03:28:52.087978 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:28:52.088354 kubelet[3142]: I0527 03:28:52.087995 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:28:52.644629 ntpd[1869]: Listen normally on 11 lxc_health [fe80::80d0:3cff:fe48:525a%10]:123 May 27 03:28:52.646422 ntpd[1869]: 27 May 03:28:52 ntpd[1869]: Listen normally on 11 lxc_health [fe80::80d0:3cff:fe48:525a%10]:123 May 27 03:28:54.155385 containerd[1895]: time="2025-05-27T03:28:54.155336560Z" level=info msg="TaskExit event in podsandbox handler container_id:\"dd6b7282faf86e47ede508233a07ce3e6b9bb3ab93ef365ab0e47aeb64c12b14\" id:\"8dc79841cfb5ec27061ca19f0b55812c312336cf0012b14ad8c6151f6f13ea63\" pid:5825 exited_at:{seconds:1748316534 nanos:154658763}" May 27 03:28:54.263752 sshd[5045]: Connection closed by 139.178.68.195 port 58782 May 27 03:28:54.265835 sshd-session[4968]: pam_unix(sshd:session): session closed for user core May 27 03:28:54.270256 systemd-logind[1875]: Session 26 logged out. Waiting for processes to exit. May 27 03:28:54.270415 systemd[1]: sshd@25-172.31.17.111:22-139.178.68.195:58782.service: Deactivated successfully. May 27 03:28:54.272389 systemd[1]: session-26.scope: Deactivated successfully. May 27 03:28:54.275038 systemd-logind[1875]: Removed session 26. May 27 03:29:02.114106 kubelet[3142]: I0527 03:29:02.107865 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:29:02.120183 kubelet[3142]: I0527 03:29:02.114148 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:29:02.123288 kubelet[3142]: I0527 03:29:02.123194 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:29:02.139825 kubelet[3142]: I0527 03:29:02.139788 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:29:02.139995 kubelet[3142]: I0527 03:29:02.139956 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-m6qb7","kube-system/cilium-c2n78","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-proxy-k7s89","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:29:02.140048 kubelet[3142]: E0527 03:29:02.140004 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-m6qb7" May 27 03:29:02.140048 kubelet[3142]: E0527 03:29:02.140024 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-c2n78" May 27 03:29:02.140048 kubelet[3142]: E0527 03:29:02.140037 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:29:02.140154 kubelet[3142]: E0527 03:29:02.140051 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:29:02.140154 kubelet[3142]: E0527 03:29:02.140064 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:29:02.140154 kubelet[3142]: E0527 03:29:02.140076 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:29:02.140154 kubelet[3142]: I0527 03:29:02.140093 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:29:12.153925 kubelet[3142]: I0527 03:29:12.153891 3142 eviction_manager.go:376] "Eviction manager: attempting to reclaim" resourceName="ephemeral-storage" May 27 03:29:12.153925 kubelet[3142]: I0527 03:29:12.153943 3142 container_gc.go:86] "Attempting to delete unused containers" May 27 03:29:12.156585 kubelet[3142]: I0527 03:29:12.156532 3142 image_gc_manager.go:447] "Attempting to delete unused images" May 27 03:29:12.158506 kubelet[3142]: I0527 03:29:12.158472 3142 image_gc_manager.go:514] "Removing image to free bytes" imageID="sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b" size=20939036 runtimeHandler="" May 27 03:29:12.159049 containerd[1895]: time="2025-05-27T03:29:12.159005389Z" level=info msg="RemoveImage \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" May 27 03:29:12.167160 containerd[1895]: time="2025-05-27T03:29:12.167109846Z" level=info msg="ImageDelete event name:\"registry.k8s.io/coredns/coredns:v1.12.0\"" May 27 03:29:12.169336 containerd[1895]: time="2025-05-27T03:29:12.169267387Z" level=info msg="ImageDelete event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\"" May 27 03:29:12.171493 containerd[1895]: time="2025-05-27T03:29:12.171441292Z" level=info msg="RemoveImage \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" returns successfully" May 27 03:29:12.171619 containerd[1895]: time="2025-05-27T03:29:12.171584385Z" level=info msg="ImageDelete event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" May 27 03:29:12.171896 kubelet[3142]: I0527 03:29:12.171863 3142 image_gc_manager.go:514] "Removing image to free bytes" imageID="sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1" size=58938593 runtimeHandler="" May 27 03:29:12.172238 containerd[1895]: time="2025-05-27T03:29:12.172101517Z" level=info msg="RemoveImage \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" May 27 03:29:12.174499 containerd[1895]: time="2025-05-27T03:29:12.174466013Z" level=info msg="ImageDelete event name:\"registry.k8s.io/etcd:3.5.21-0\"" May 27 03:29:12.177020 containerd[1895]: time="2025-05-27T03:29:12.176980404Z" level=info msg="ImageDelete event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\"" May 27 03:29:12.179130 containerd[1895]: time="2025-05-27T03:29:12.178954910Z" level=info msg="RemoveImage \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" returns successfully" May 27 03:29:12.179130 containerd[1895]: time="2025-05-27T03:29:12.179082382Z" level=info msg="ImageDelete event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" May 27 03:29:12.188442 kubelet[3142]: I0527 03:29:12.188398 3142 eviction_manager.go:387] "Eviction manager: must evict pod(s) to reclaim" resourceName="ephemeral-storage" May 27 03:29:12.188568 kubelet[3142]: I0527 03:29:12.188531 3142 eviction_manager.go:405] "Eviction manager: pods ranked for eviction" pods=["kube-system/cilium-operator-6c4d7847fc-m6qb7","kube-system/cilium-c2n78","kube-system/kube-controller-manager-ip-172-31-17-111","kube-system/kube-proxy-k7s89","kube-system/kube-apiserver-ip-172-31-17-111","kube-system/kube-scheduler-ip-172-31-17-111"] May 27 03:29:12.188609 kubelet[3142]: E0527 03:29:12.188568 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-operator-6c4d7847fc-m6qb7" May 27 03:29:12.188609 kubelet[3142]: E0527 03:29:12.188581 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/cilium-c2n78" May 27 03:29:12.188609 kubelet[3142]: E0527 03:29:12.188589 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-controller-manager-ip-172-31-17-111" May 27 03:29:12.188609 kubelet[3142]: E0527 03:29:12.188598 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-proxy-k7s89" May 27 03:29:12.188609 kubelet[3142]: E0527 03:29:12.188608 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-apiserver-ip-172-31-17-111" May 27 03:29:12.188725 kubelet[3142]: E0527 03:29:12.188616 3142 eviction_manager.go:610] "Eviction manager: cannot evict a critical pod" pod="kube-system/kube-scheduler-ip-172-31-17-111" May 27 03:29:12.188725 kubelet[3142]: I0527 03:29:12.188626 3142 eviction_manager.go:439] "Eviction manager: unable to evict any pods from the node" May 27 03:29:30.715725 systemd[1]: cri-containerd-ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557.scope: Deactivated successfully. May 27 03:29:30.716699 systemd[1]: cri-containerd-ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557.scope: Consumed 3.262s CPU time, 79.7M memory peak, 33.8M read from disk. May 27 03:29:30.718563 containerd[1895]: time="2025-05-27T03:29:30.718536386Z" level=info msg="received exit event container_id:\"ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557\" id:\"ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557\" pid:2981 exit_status:1 exited_at:{seconds:1748316570 nanos:718236701}" May 27 03:29:30.719875 containerd[1895]: time="2025-05-27T03:29:30.719846516Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557\" id:\"ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557\" pid:2981 exit_status:1 exited_at:{seconds:1748316570 nanos:718236701}" May 27 03:29:30.746455 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557-rootfs.mount: Deactivated successfully. May 27 03:29:30.888591 kubelet[3142]: I0527 03:29:30.887793 3142 scope.go:117] "RemoveContainer" containerID="ec21de669f53fd8a1e8e29b30a581f9b2fcc42ff57eb5c9012159cb4f24dc557" May 27 03:29:30.891388 containerd[1895]: time="2025-05-27T03:29:30.891337957Z" level=info msg="CreateContainer within sandbox \"9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" May 27 03:29:30.907894 containerd[1895]: time="2025-05-27T03:29:30.906976882Z" level=info msg="Container da5f39b8992f7713ba6869f494486654f9d1edd9f60860a585fdce9a37e917c1: CDI devices from CRI Config.CDIDevices: []" May 27 03:29:30.909933 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2002481007.mount: Deactivated successfully. May 27 03:29:30.921285 containerd[1895]: time="2025-05-27T03:29:30.921227722Z" level=info msg="CreateContainer within sandbox \"9d1e6bb8516657c8cc34ad6a30302b144d17c214e222796f7b53b6f8b385d805\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"da5f39b8992f7713ba6869f494486654f9d1edd9f60860a585fdce9a37e917c1\"" May 27 03:29:30.921835 containerd[1895]: time="2025-05-27T03:29:30.921812670Z" level=info msg="StartContainer for \"da5f39b8992f7713ba6869f494486654f9d1edd9f60860a585fdce9a37e917c1\"" May 27 03:29:30.923042 containerd[1895]: time="2025-05-27T03:29:30.922992865Z" level=info msg="connecting to shim da5f39b8992f7713ba6869f494486654f9d1edd9f60860a585fdce9a37e917c1" address="unix:///run/containerd/s/80f20d15877b2529fc9ca266ce1ef8e61815738865c79d923fc688cf4d70633e" protocol=ttrpc version=3 May 27 03:29:30.949544 systemd[1]: Started cri-containerd-da5f39b8992f7713ba6869f494486654f9d1edd9f60860a585fdce9a37e917c1.scope - libcontainer container da5f39b8992f7713ba6869f494486654f9d1edd9f60860a585fdce9a37e917c1. May 27 03:29:31.010210 containerd[1895]: time="2025-05-27T03:29:31.009484504Z" level=info msg="StartContainer for \"da5f39b8992f7713ba6869f494486654f9d1edd9f60860a585fdce9a37e917c1\" returns successfully" May 27 03:29:31.074986 systemd[1]: cri-containerd-8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d.scope: Deactivated successfully. May 27 03:29:31.076432 systemd[1]: cri-containerd-8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d.scope: Consumed 299ms CPU time, 32.6M memory peak, 8.3M read from disk, 4K written to disk. May 27 03:29:31.077099 containerd[1895]: time="2025-05-27T03:29:31.076693955Z" level=info msg="received exit event container_id:\"8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d\" id:\"8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d\" pid:5049 exit_status:1 exited_at:{seconds:1748316571 nanos:76162492}" May 27 03:29:31.077895 containerd[1895]: time="2025-05-27T03:29:31.077797691Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d\" id:\"8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d\" pid:5049 exit_status:1 exited_at:{seconds:1748316571 nanos:76162492}" May 27 03:29:31.105705 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d-rootfs.mount: Deactivated successfully. May 27 03:29:31.892784 kubelet[3142]: I0527 03:29:31.892452 3142 scope.go:117] "RemoveContainer" containerID="8e360355187eb1a8370021f7210ccee3c3cbb37babfb69fd96c8ec6bc90e092d" May 27 03:29:31.896237 containerd[1895]: time="2025-05-27T03:29:31.896133997Z" level=info msg="CreateContainer within sandbox \"9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2\" for container &ContainerMetadata{Name:cilium-operator,Attempt:1,}" May 27 03:29:31.913332 containerd[1895]: time="2025-05-27T03:29:31.911426903Z" level=info msg="Container 12a105594886aa8a769fa5853907f5424e6903fe5b06795f9f81f3e4c32f2492: CDI devices from CRI Config.CDIDevices: []" May 27 03:29:31.924225 containerd[1895]: time="2025-05-27T03:29:31.924182535Z" level=info msg="CreateContainer within sandbox \"9c13442c39db62f309c70986f16fda24f8d4d1b519120f9f4f1485de7c61c5a2\" for &ContainerMetadata{Name:cilium-operator,Attempt:1,} returns container id \"12a105594886aa8a769fa5853907f5424e6903fe5b06795f9f81f3e4c32f2492\"" May 27 03:29:31.925161 containerd[1895]: time="2025-05-27T03:29:31.925133939Z" level=info msg="StartContainer for \"12a105594886aa8a769fa5853907f5424e6903fe5b06795f9f81f3e4c32f2492\"" May 27 03:29:31.926640 containerd[1895]: time="2025-05-27T03:29:31.926599706Z" level=info msg="connecting to shim 12a105594886aa8a769fa5853907f5424e6903fe5b06795f9f81f3e4c32f2492" address="unix:///run/containerd/s/97d798f72db9026dc387eef7ff1de28a97b7e1221a77782d94b6b74b1cb231b1" protocol=ttrpc version=3 May 27 03:29:31.960577 systemd[1]: Started cri-containerd-12a105594886aa8a769fa5853907f5424e6903fe5b06795f9f81f3e4c32f2492.scope - libcontainer container 12a105594886aa8a769fa5853907f5424e6903fe5b06795f9f81f3e4c32f2492. May 27 03:29:32.037618 containerd[1895]: time="2025-05-27T03:29:32.037555105Z" level=info msg="StartContainer for \"12a105594886aa8a769fa5853907f5424e6903fe5b06795f9f81f3e4c32f2492\" returns successfully" May 27 03:29:34.380652 kubelet[3142]: E0527 03:29:34.380594 3142 controller.go:195] "Failed to update lease" err="Put \"https://172.31.17.111:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-17-111?timeout=10s\": context deadline exceeded" May 27 03:29:35.359779 systemd[1]: cri-containerd-436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187.scope: Deactivated successfully. May 27 03:29:35.360046 systemd[1]: cri-containerd-436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187.scope: Consumed 2.353s CPU time, 31.7M memory peak, 13M read from disk. May 27 03:29:35.364047 containerd[1895]: time="2025-05-27T03:29:35.364012622Z" level=info msg="TaskExit event in podsandbox handler container_id:\"436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187\" id:\"436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187\" pid:2970 exit_status:1 exited_at:{seconds:1748316575 nanos:363104480}" May 27 03:29:35.364589 containerd[1895]: time="2025-05-27T03:29:35.364081625Z" level=info msg="received exit event container_id:\"436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187\" id:\"436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187\" pid:2970 exit_status:1 exited_at:{seconds:1748316575 nanos:363104480}" May 27 03:29:35.391399 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187-rootfs.mount: Deactivated successfully. May 27 03:29:35.907229 kubelet[3142]: I0527 03:29:35.907200 3142 scope.go:117] "RemoveContainer" containerID="436ce5b9b4843263cc5b0ee6cd68937815890fa7b7c9ce48355f86b2ecb6d187" May 27 03:29:35.909726 containerd[1895]: time="2025-05-27T03:29:35.909384291Z" level=info msg="CreateContainer within sandbox \"26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" May 27 03:29:35.924012 containerd[1895]: time="2025-05-27T03:29:35.923979168Z" level=info msg="Container 7a63dada0776f3bef771d336e5bb1cdd04f95db41367900e99711f8fe5c74f5e: CDI devices from CRI Config.CDIDevices: []" May 27 03:29:35.928199 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1769391932.mount: Deactivated successfully. May 27 03:29:35.936678 containerd[1895]: time="2025-05-27T03:29:35.936627184Z" level=info msg="CreateContainer within sandbox \"26d5425c3ff1182c7a8f1ed6ed335004fb013f221e9c215f7db32b223daee9ba\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"7a63dada0776f3bef771d336e5bb1cdd04f95db41367900e99711f8fe5c74f5e\"" May 27 03:29:35.938347 containerd[1895]: time="2025-05-27T03:29:35.937269817Z" level=info msg="StartContainer for \"7a63dada0776f3bef771d336e5bb1cdd04f95db41367900e99711f8fe5c74f5e\"" May 27 03:29:35.939631 containerd[1895]: time="2025-05-27T03:29:35.939596786Z" level=info msg="connecting to shim 7a63dada0776f3bef771d336e5bb1cdd04f95db41367900e99711f8fe5c74f5e" address="unix:///run/containerd/s/3e530b6cfa15efd8d23f031e713f5ea1a754b2facde1d92a0676ffa47cc590f5" protocol=ttrpc version=3 May 27 03:29:35.969543 systemd[1]: Started cri-containerd-7a63dada0776f3bef771d336e5bb1cdd04f95db41367900e99711f8fe5c74f5e.scope - libcontainer container 7a63dada0776f3bef771d336e5bb1cdd04f95db41367900e99711f8fe5c74f5e. May 27 03:29:36.024786 containerd[1895]: time="2025-05-27T03:29:36.024712271Z" level=info msg="StartContainer for \"7a63dada0776f3bef771d336e5bb1cdd04f95db41367900e99711f8fe5c74f5e\" returns successfully" May 27 03:29:44.381922 kubelet[3142]: E0527 03:29:44.381814 3142 controller.go:195] "Failed to update lease" err="Put \"https://172.31.17.111:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-17-111?timeout=10s\": context deadline exceeded"