Jul 10 00:21:44.922661 kernel: Linux version 6.12.36-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Wed Jul 9 22:15:30 -00 2025 Jul 10 00:21:44.922703 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=844005237fb9709f65a093d5533c4229fb6c54e8e257736d9c3d041b6d3080ea Jul 10 00:21:44.922719 kernel: BIOS-provided physical RAM map: Jul 10 00:21:44.922731 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Jul 10 00:21:44.922742 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable Jul 10 00:21:44.922754 kernel: BIOS-e820: [mem 0x00000000786ce000-0x000000007894dfff] reserved Jul 10 00:21:44.922768 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Jul 10 00:21:44.922780 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Jul 10 00:21:44.922795 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable Jul 10 00:21:44.922807 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Jul 10 00:21:44.922819 kernel: NX (Execute Disable) protection: active Jul 10 00:21:44.922831 kernel: APIC: Static calls initialized Jul 10 00:21:44.922843 kernel: e820: update [mem 0x768c0018-0x768c8e57] usable ==> usable Jul 10 00:21:44.922856 kernel: extended physical RAM map: Jul 10 00:21:44.923925 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Jul 10 00:21:44.923941 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000768c0017] usable Jul 10 00:21:44.923955 kernel: reserve setup_data: [mem 0x00000000768c0018-0x00000000768c8e57] usable Jul 10 00:21:44.923969 kernel: reserve setup_data: [mem 0x00000000768c8e58-0x00000000786cdfff] usable Jul 10 00:21:44.923983 kernel: reserve setup_data: [mem 0x00000000786ce000-0x000000007894dfff] reserved Jul 10 00:21:44.923997 kernel: reserve setup_data: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Jul 10 00:21:44.924011 kernel: reserve setup_data: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Jul 10 00:21:44.924026 kernel: reserve setup_data: [mem 0x00000000789de000-0x000000007c97bfff] usable Jul 10 00:21:44.924039 kernel: reserve setup_data: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Jul 10 00:21:44.924053 kernel: efi: EFI v2.7 by EDK II Jul 10 00:21:44.924070 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77003518 Jul 10 00:21:44.924084 kernel: secureboot: Secure boot disabled Jul 10 00:21:44.924098 kernel: SMBIOS 2.7 present. Jul 10 00:21:44.924112 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Jul 10 00:21:44.924126 kernel: DMI: Memory slots populated: 1/1 Jul 10 00:21:44.924139 kernel: Hypervisor detected: KVM Jul 10 00:21:44.924151 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jul 10 00:21:44.924165 kernel: kvm-clock: using sched offset of 7280935940 cycles Jul 10 00:21:44.924180 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jul 10 00:21:44.924194 kernel: tsc: Detected 2499.996 MHz processor Jul 10 00:21:44.924209 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jul 10 00:21:44.924226 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jul 10 00:21:44.924240 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 Jul 10 00:21:44.924254 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Jul 10 00:21:44.924268 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jul 10 00:21:44.924282 kernel: Using GB pages for direct mapping Jul 10 00:21:44.924301 kernel: ACPI: Early table checksum verification disabled Jul 10 00:21:44.924318 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) Jul 10 00:21:44.924333 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) Jul 10 00:21:44.924347 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Jul 10 00:21:44.924363 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Jul 10 00:21:44.924377 kernel: ACPI: FACS 0x00000000789D0000 000040 Jul 10 00:21:44.924392 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Jul 10 00:21:44.924407 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Jul 10 00:21:44.924422 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Jul 10 00:21:44.924440 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Jul 10 00:21:44.924454 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Jul 10 00:21:44.924469 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Jul 10 00:21:44.924484 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Jul 10 00:21:44.924499 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) Jul 10 00:21:44.924514 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] Jul 10 00:21:44.924529 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] Jul 10 00:21:44.924543 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] Jul 10 00:21:44.924561 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] Jul 10 00:21:44.924576 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] Jul 10 00:21:44.924590 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] Jul 10 00:21:44.924605 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] Jul 10 00:21:44.924620 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] Jul 10 00:21:44.924635 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] Jul 10 00:21:44.924649 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] Jul 10 00:21:44.924664 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] Jul 10 00:21:44.924678 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Jul 10 00:21:44.924693 kernel: NUMA: Initialized distance table, cnt=1 Jul 10 00:21:44.924711 kernel: NODE_DATA(0) allocated [mem 0x7a8eddc0-0x7a8f4fff] Jul 10 00:21:44.924726 kernel: Zone ranges: Jul 10 00:21:44.924740 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jul 10 00:21:44.924755 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] Jul 10 00:21:44.924770 kernel: Normal empty Jul 10 00:21:44.924785 kernel: Device empty Jul 10 00:21:44.924799 kernel: Movable zone start for each node Jul 10 00:21:44.924814 kernel: Early memory node ranges Jul 10 00:21:44.924828 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Jul 10 00:21:44.924846 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] Jul 10 00:21:44.924878 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] Jul 10 00:21:44.925598 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] Jul 10 00:21:44.925622 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jul 10 00:21:44.925635 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Jul 10 00:21:44.925650 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Jul 10 00:21:44.925665 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges Jul 10 00:21:44.925678 kernel: ACPI: PM-Timer IO Port: 0xb008 Jul 10 00:21:44.925693 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jul 10 00:21:44.925713 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Jul 10 00:21:44.925728 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jul 10 00:21:44.925742 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jul 10 00:21:44.925758 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jul 10 00:21:44.925772 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jul 10 00:21:44.925788 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jul 10 00:21:44.925803 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jul 10 00:21:44.925818 kernel: TSC deadline timer available Jul 10 00:21:44.925832 kernel: CPU topo: Max. logical packages: 1 Jul 10 00:21:44.925847 kernel: CPU topo: Max. logical dies: 1 Jul 10 00:21:44.926169 kernel: CPU topo: Max. dies per package: 1 Jul 10 00:21:44.926185 kernel: CPU topo: Max. threads per core: 2 Jul 10 00:21:44.926199 kernel: CPU topo: Num. cores per package: 1 Jul 10 00:21:44.926213 kernel: CPU topo: Num. threads per package: 2 Jul 10 00:21:44.926227 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Jul 10 00:21:44.926241 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jul 10 00:21:44.926255 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices Jul 10 00:21:44.926269 kernel: Booting paravirtualized kernel on KVM Jul 10 00:21:44.926283 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jul 10 00:21:44.926302 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Jul 10 00:21:44.926316 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Jul 10 00:21:44.926330 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Jul 10 00:21:44.926343 kernel: pcpu-alloc: [0] 0 1 Jul 10 00:21:44.926357 kernel: kvm-guest: PV spinlocks enabled Jul 10 00:21:44.926371 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jul 10 00:21:44.926388 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=844005237fb9709f65a093d5533c4229fb6c54e8e257736d9c3d041b6d3080ea Jul 10 00:21:44.926403 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 10 00:21:44.926419 kernel: random: crng init done Jul 10 00:21:44.926432 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jul 10 00:21:44.926446 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jul 10 00:21:44.926459 kernel: Fallback order for Node 0: 0 Jul 10 00:21:44.926473 kernel: Built 1 zonelists, mobility grouping on. Total pages: 509451 Jul 10 00:21:44.926487 kernel: Policy zone: DMA32 Jul 10 00:21:44.926515 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 10 00:21:44.926529 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jul 10 00:21:44.926544 kernel: Kernel/User page tables isolation: enabled Jul 10 00:21:44.926559 kernel: ftrace: allocating 40095 entries in 157 pages Jul 10 00:21:44.926575 kernel: ftrace: allocated 157 pages with 5 groups Jul 10 00:21:44.926592 kernel: Dynamic Preempt: voluntary Jul 10 00:21:44.926606 kernel: rcu: Preemptible hierarchical RCU implementation. Jul 10 00:21:44.926623 kernel: rcu: RCU event tracing is enabled. Jul 10 00:21:44.926640 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jul 10 00:21:44.926655 kernel: Trampoline variant of Tasks RCU enabled. Jul 10 00:21:44.926670 kernel: Rude variant of Tasks RCU enabled. Jul 10 00:21:44.926687 kernel: Tracing variant of Tasks RCU enabled. Jul 10 00:21:44.926700 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 10 00:21:44.926713 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jul 10 00:21:44.926727 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 10 00:21:44.926742 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 10 00:21:44.926755 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 10 00:21:44.926770 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Jul 10 00:21:44.926785 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jul 10 00:21:44.926803 kernel: Console: colour dummy device 80x25 Jul 10 00:21:44.926816 kernel: printk: legacy console [tty0] enabled Jul 10 00:21:44.927894 kernel: printk: legacy console [ttyS0] enabled Jul 10 00:21:44.927922 kernel: ACPI: Core revision 20240827 Jul 10 00:21:44.927939 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Jul 10 00:21:44.927956 kernel: APIC: Switch to symmetric I/O mode setup Jul 10 00:21:44.927972 kernel: x2apic enabled Jul 10 00:21:44.927988 kernel: APIC: Switched APIC routing to: physical x2apic Jul 10 00:21:44.928004 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Jul 10 00:21:44.928025 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499996) Jul 10 00:21:44.928041 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Jul 10 00:21:44.928057 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Jul 10 00:21:44.928073 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jul 10 00:21:44.928089 kernel: Spectre V2 : Mitigation: Retpolines Jul 10 00:21:44.928105 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Jul 10 00:21:44.928121 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Jul 10 00:21:44.928137 kernel: RETBleed: Vulnerable Jul 10 00:21:44.928152 kernel: Speculative Store Bypass: Vulnerable Jul 10 00:21:44.928168 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Jul 10 00:21:44.928184 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jul 10 00:21:44.928202 kernel: GDS: Unknown: Dependent on hypervisor status Jul 10 00:21:44.928218 kernel: ITS: Mitigation: Aligned branch/return thunks Jul 10 00:21:44.928233 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jul 10 00:21:44.928249 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jul 10 00:21:44.928264 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jul 10 00:21:44.928280 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Jul 10 00:21:44.928295 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Jul 10 00:21:44.928312 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Jul 10 00:21:44.928327 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Jul 10 00:21:44.928343 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Jul 10 00:21:44.928358 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Jul 10 00:21:44.928377 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jul 10 00:21:44.928393 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Jul 10 00:21:44.928409 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Jul 10 00:21:44.928424 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Jul 10 00:21:44.928440 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Jul 10 00:21:44.928456 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Jul 10 00:21:44.928472 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Jul 10 00:21:44.928488 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Jul 10 00:21:44.928503 kernel: Freeing SMP alternatives memory: 32K Jul 10 00:21:44.928518 kernel: pid_max: default: 32768 minimum: 301 Jul 10 00:21:44.928534 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jul 10 00:21:44.928552 kernel: landlock: Up and running. Jul 10 00:21:44.928568 kernel: SELinux: Initializing. Jul 10 00:21:44.928584 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jul 10 00:21:44.928600 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jul 10 00:21:44.928616 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) Jul 10 00:21:44.928632 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Jul 10 00:21:44.928648 kernel: signal: max sigframe size: 3632 Jul 10 00:21:44.928664 kernel: rcu: Hierarchical SRCU implementation. Jul 10 00:21:44.928681 kernel: rcu: Max phase no-delay instances is 400. Jul 10 00:21:44.928697 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Jul 10 00:21:44.928716 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jul 10 00:21:44.928732 kernel: smp: Bringing up secondary CPUs ... Jul 10 00:21:44.928747 kernel: smpboot: x86: Booting SMP configuration: Jul 10 00:21:44.928763 kernel: .... node #0, CPUs: #1 Jul 10 00:21:44.928780 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Jul 10 00:21:44.928796 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Jul 10 00:21:44.928812 kernel: smp: Brought up 1 node, 2 CPUs Jul 10 00:21:44.928828 kernel: smpboot: Total of 2 processors activated (9999.98 BogoMIPS) Jul 10 00:21:44.928844 kernel: Memory: 1908052K/2037804K available (14336K kernel code, 2430K rwdata, 9956K rodata, 54420K init, 2548K bss, 125188K reserved, 0K cma-reserved) Jul 10 00:21:44.928883 kernel: devtmpfs: initialized Jul 10 00:21:44.928899 kernel: x86/mm: Memory block size: 128MB Jul 10 00:21:44.928916 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) Jul 10 00:21:44.928932 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 10 00:21:44.928948 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jul 10 00:21:44.928964 kernel: pinctrl core: initialized pinctrl subsystem Jul 10 00:21:44.928981 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 10 00:21:44.928996 kernel: audit: initializing netlink subsys (disabled) Jul 10 00:21:44.929017 kernel: audit: type=2000 audit(1752106902.887:1): state=initialized audit_enabled=0 res=1 Jul 10 00:21:44.929033 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 10 00:21:44.929049 kernel: thermal_sys: Registered thermal governor 'user_space' Jul 10 00:21:44.929064 kernel: cpuidle: using governor menu Jul 10 00:21:44.929182 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 10 00:21:44.929199 kernel: dca service started, version 1.12.1 Jul 10 00:21:44.929215 kernel: PCI: Using configuration type 1 for base access Jul 10 00:21:44.929232 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jul 10 00:21:44.929248 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jul 10 00:21:44.929268 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jul 10 00:21:44.929285 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jul 10 00:21:44.929301 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jul 10 00:21:44.929317 kernel: ACPI: Added _OSI(Module Device) Jul 10 00:21:44.929333 kernel: ACPI: Added _OSI(Processor Device) Jul 10 00:21:44.929349 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 10 00:21:44.929365 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Jul 10 00:21:44.929381 kernel: ACPI: Interpreter enabled Jul 10 00:21:44.929396 kernel: ACPI: PM: (supports S0 S5) Jul 10 00:21:44.929416 kernel: ACPI: Using IOAPIC for interrupt routing Jul 10 00:21:44.929432 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jul 10 00:21:44.929448 kernel: PCI: Using E820 reservations for host bridge windows Jul 10 00:21:44.929464 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Jul 10 00:21:44.929479 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jul 10 00:21:44.929754 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Jul 10 00:21:44.932992 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Jul 10 00:21:44.933167 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Jul 10 00:21:44.933193 kernel: acpiphp: Slot [3] registered Jul 10 00:21:44.933208 kernel: acpiphp: Slot [4] registered Jul 10 00:21:44.933223 kernel: acpiphp: Slot [5] registered Jul 10 00:21:44.933238 kernel: acpiphp: Slot [6] registered Jul 10 00:21:44.933253 kernel: acpiphp: Slot [7] registered Jul 10 00:21:44.933268 kernel: acpiphp: Slot [8] registered Jul 10 00:21:44.933283 kernel: acpiphp: Slot [9] registered Jul 10 00:21:44.933298 kernel: acpiphp: Slot [10] registered Jul 10 00:21:44.933313 kernel: acpiphp: Slot [11] registered Jul 10 00:21:44.933331 kernel: acpiphp: Slot [12] registered Jul 10 00:21:44.933346 kernel: acpiphp: Slot [13] registered Jul 10 00:21:44.933360 kernel: acpiphp: Slot [14] registered Jul 10 00:21:44.933375 kernel: acpiphp: Slot [15] registered Jul 10 00:21:44.933390 kernel: acpiphp: Slot [16] registered Jul 10 00:21:44.933405 kernel: acpiphp: Slot [17] registered Jul 10 00:21:44.933419 kernel: acpiphp: Slot [18] registered Jul 10 00:21:44.933434 kernel: acpiphp: Slot [19] registered Jul 10 00:21:44.933449 kernel: acpiphp: Slot [20] registered Jul 10 00:21:44.933466 kernel: acpiphp: Slot [21] registered Jul 10 00:21:44.933481 kernel: acpiphp: Slot [22] registered Jul 10 00:21:44.933496 kernel: acpiphp: Slot [23] registered Jul 10 00:21:44.933511 kernel: acpiphp: Slot [24] registered Jul 10 00:21:44.933525 kernel: acpiphp: Slot [25] registered Jul 10 00:21:44.933540 kernel: acpiphp: Slot [26] registered Jul 10 00:21:44.933554 kernel: acpiphp: Slot [27] registered Jul 10 00:21:44.933569 kernel: acpiphp: Slot [28] registered Jul 10 00:21:44.933583 kernel: acpiphp: Slot [29] registered Jul 10 00:21:44.933598 kernel: acpiphp: Slot [30] registered Jul 10 00:21:44.933614 kernel: acpiphp: Slot [31] registered Jul 10 00:21:44.933636 kernel: PCI host bridge to bus 0000:00 Jul 10 00:21:44.933775 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jul 10 00:21:44.933992 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jul 10 00:21:44.934122 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jul 10 00:21:44.934242 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Jul 10 00:21:44.934363 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] Jul 10 00:21:44.934499 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jul 10 00:21:44.934662 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint Jul 10 00:21:44.934810 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint Jul 10 00:21:44.937048 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 conventional PCI endpoint Jul 10 00:21:44.937217 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Jul 10 00:21:44.937350 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Jul 10 00:21:44.937494 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Jul 10 00:21:44.937628 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Jul 10 00:21:44.937775 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Jul 10 00:21:44.937955 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Jul 10 00:21:44.938097 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Jul 10 00:21:44.938253 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 conventional PCI endpoint Jul 10 00:21:44.938398 kernel: pci 0000:00:03.0: BAR 0 [mem 0x80000000-0x803fffff pref] Jul 10 00:21:44.938546 kernel: pci 0000:00:03.0: ROM [mem 0xffff0000-0xffffffff pref] Jul 10 00:21:44.938686 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jul 10 00:21:44.938836 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 PCIe Endpoint Jul 10 00:21:44.938996 kernel: pci 0000:00:04.0: BAR 0 [mem 0x80404000-0x80407fff] Jul 10 00:21:44.939134 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 PCIe Endpoint Jul 10 00:21:44.939261 kernel: pci 0000:00:05.0: BAR 0 [mem 0x80400000-0x80403fff] Jul 10 00:21:44.939280 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jul 10 00:21:44.939300 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jul 10 00:21:44.939315 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jul 10 00:21:44.939330 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jul 10 00:21:44.939345 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Jul 10 00:21:44.939360 kernel: iommu: Default domain type: Translated Jul 10 00:21:44.939375 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jul 10 00:21:44.939391 kernel: efivars: Registered efivars operations Jul 10 00:21:44.939405 kernel: PCI: Using ACPI for IRQ routing Jul 10 00:21:44.939420 kernel: PCI: pci_cache_line_size set to 64 bytes Jul 10 00:21:44.939438 kernel: e820: reserve RAM buffer [mem 0x768c0018-0x77ffffff] Jul 10 00:21:44.939453 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] Jul 10 00:21:44.939467 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] Jul 10 00:21:44.939590 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Jul 10 00:21:44.939714 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Jul 10 00:21:44.939852 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jul 10 00:21:44.939976 kernel: vgaarb: loaded Jul 10 00:21:44.939991 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Jul 10 00:21:44.940010 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Jul 10 00:21:44.940024 kernel: clocksource: Switched to clocksource kvm-clock Jul 10 00:21:44.940038 kernel: VFS: Disk quotas dquot_6.6.0 Jul 10 00:21:44.940053 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 10 00:21:44.940067 kernel: pnp: PnP ACPI init Jul 10 00:21:44.940080 kernel: pnp: PnP ACPI: found 5 devices Jul 10 00:21:44.940095 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jul 10 00:21:44.940110 kernel: NET: Registered PF_INET protocol family Jul 10 00:21:44.940125 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Jul 10 00:21:44.940143 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Jul 10 00:21:44.940158 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 10 00:21:44.940174 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Jul 10 00:21:44.940189 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Jul 10 00:21:44.940203 kernel: TCP: Hash tables configured (established 16384 bind 16384) Jul 10 00:21:44.940218 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Jul 10 00:21:44.940233 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Jul 10 00:21:44.940248 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 10 00:21:44.940263 kernel: NET: Registered PF_XDP protocol family Jul 10 00:21:44.940414 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jul 10 00:21:44.940533 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jul 10 00:21:44.940649 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jul 10 00:21:44.940758 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Jul 10 00:21:44.941915 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] Jul 10 00:21:44.942075 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jul 10 00:21:44.942094 kernel: PCI: CLS 0 bytes, default 64 Jul 10 00:21:44.942108 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jul 10 00:21:44.942128 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Jul 10 00:21:44.942142 kernel: clocksource: Switched to clocksource tsc Jul 10 00:21:44.942156 kernel: Initialise system trusted keyrings Jul 10 00:21:44.942170 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Jul 10 00:21:44.942185 kernel: Key type asymmetric registered Jul 10 00:21:44.942200 kernel: Asymmetric key parser 'x509' registered Jul 10 00:21:44.942216 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jul 10 00:21:44.942232 kernel: io scheduler mq-deadline registered Jul 10 00:21:44.942248 kernel: io scheduler kyber registered Jul 10 00:21:44.942266 kernel: io scheduler bfq registered Jul 10 00:21:44.942282 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jul 10 00:21:44.942297 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 10 00:21:44.942311 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jul 10 00:21:44.942325 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jul 10 00:21:44.942340 kernel: i8042: Warning: Keylock active Jul 10 00:21:44.942356 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jul 10 00:21:44.942370 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jul 10 00:21:44.942535 kernel: rtc_cmos 00:00: RTC can wake from S4 Jul 10 00:21:44.942680 kernel: rtc_cmos 00:00: registered as rtc0 Jul 10 00:21:44.942811 kernel: rtc_cmos 00:00: setting system clock to 2025-07-10T00:21:44 UTC (1752106904) Jul 10 00:21:44.945000 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Jul 10 00:21:44.945031 kernel: intel_pstate: CPU model not supported Jul 10 00:21:44.945071 kernel: efifb: probing for efifb Jul 10 00:21:44.945101 kernel: efifb: framebuffer at 0x80000000, using 1876k, total 1875k Jul 10 00:21:44.945118 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Jul 10 00:21:44.945137 kernel: efifb: scrolling: redraw Jul 10 00:21:44.945153 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Jul 10 00:21:44.945169 kernel: Console: switching to colour frame buffer device 100x37 Jul 10 00:21:44.945185 kernel: fb0: EFI VGA frame buffer device Jul 10 00:21:44.945201 kernel: pstore: Using crash dump compression: deflate Jul 10 00:21:44.945217 kernel: pstore: Registered efi_pstore as persistent store backend Jul 10 00:21:44.945232 kernel: NET: Registered PF_INET6 protocol family Jul 10 00:21:44.945248 kernel: Segment Routing with IPv6 Jul 10 00:21:44.945264 kernel: In-situ OAM (IOAM) with IPv6 Jul 10 00:21:44.945280 kernel: NET: Registered PF_PACKET protocol family Jul 10 00:21:44.945298 kernel: Key type dns_resolver registered Jul 10 00:21:44.945314 kernel: IPI shorthand broadcast: enabled Jul 10 00:21:44.945329 kernel: sched_clock: Marking stable (2737002586, 162733777)->(2994166110, -94429747) Jul 10 00:21:44.945345 kernel: registered taskstats version 1 Jul 10 00:21:44.945361 kernel: Loading compiled-in X.509 certificates Jul 10 00:21:44.945377 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.36-flatcar: f515550de55d4e43b2ea11ae212aa0cb3a4e55cf' Jul 10 00:21:44.945392 kernel: Demotion targets for Node 0: null Jul 10 00:21:44.945408 kernel: Key type .fscrypt registered Jul 10 00:21:44.945424 kernel: Key type fscrypt-provisioning registered Jul 10 00:21:44.945442 kernel: ima: No TPM chip found, activating TPM-bypass! Jul 10 00:21:44.945457 kernel: ima: Allocated hash algorithm: sha1 Jul 10 00:21:44.945473 kernel: ima: No architecture policies found Jul 10 00:21:44.945488 kernel: clk: Disabling unused clocks Jul 10 00:21:44.945505 kernel: Warning: unable to open an initial console. Jul 10 00:21:44.945520 kernel: Freeing unused kernel image (initmem) memory: 54420K Jul 10 00:21:44.945536 kernel: Write protecting the kernel read-only data: 24576k Jul 10 00:21:44.945552 kernel: Freeing unused kernel image (rodata/data gap) memory: 284K Jul 10 00:21:44.945571 kernel: Run /init as init process Jul 10 00:21:44.945590 kernel: with arguments: Jul 10 00:21:44.945608 kernel: /init Jul 10 00:21:44.945624 kernel: with environment: Jul 10 00:21:44.945640 kernel: HOME=/ Jul 10 00:21:44.945656 kernel: TERM=linux Jul 10 00:21:44.945673 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 10 00:21:44.945692 systemd[1]: Successfully made /usr/ read-only. Jul 10 00:21:44.945712 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 10 00:21:44.945726 systemd[1]: Detected virtualization amazon. Jul 10 00:21:44.945740 systemd[1]: Detected architecture x86-64. Jul 10 00:21:44.945753 systemd[1]: Running in initrd. Jul 10 00:21:44.945767 systemd[1]: No hostname configured, using default hostname. Jul 10 00:21:44.945786 systemd[1]: Hostname set to . Jul 10 00:21:44.945801 systemd[1]: Initializing machine ID from VM UUID. Jul 10 00:21:44.945817 systemd[1]: Queued start job for default target initrd.target. Jul 10 00:21:44.945832 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 10 00:21:44.947970 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 10 00:21:44.948004 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jul 10 00:21:44.948021 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 10 00:21:44.948035 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jul 10 00:21:44.948057 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jul 10 00:21:44.948075 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jul 10 00:21:44.948093 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jul 10 00:21:44.948110 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 10 00:21:44.948128 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 10 00:21:44.948145 systemd[1]: Reached target paths.target - Path Units. Jul 10 00:21:44.948163 systemd[1]: Reached target slices.target - Slice Units. Jul 10 00:21:44.948184 systemd[1]: Reached target swap.target - Swaps. Jul 10 00:21:44.948201 systemd[1]: Reached target timers.target - Timer Units. Jul 10 00:21:44.948218 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jul 10 00:21:44.948236 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 10 00:21:44.948253 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 10 00:21:44.948271 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jul 10 00:21:44.948289 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 10 00:21:44.948305 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 10 00:21:44.948326 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 10 00:21:44.948342 systemd[1]: Reached target sockets.target - Socket Units. Jul 10 00:21:44.948359 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jul 10 00:21:44.948376 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 10 00:21:44.948393 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jul 10 00:21:44.948410 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jul 10 00:21:44.948427 systemd[1]: Starting systemd-fsck-usr.service... Jul 10 00:21:44.948444 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 10 00:21:44.948460 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 10 00:21:44.948480 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:21:44.948497 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jul 10 00:21:44.948515 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 10 00:21:44.948569 systemd-journald[207]: Collecting audit messages is disabled. Jul 10 00:21:44.948609 systemd[1]: Finished systemd-fsck-usr.service. Jul 10 00:21:44.948627 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 10 00:21:44.948645 systemd-journald[207]: Journal started Jul 10 00:21:44.948683 systemd-journald[207]: Runtime Journal (/run/log/journal/ec207223f90241a2490db0b13b4ab104) is 4.8M, max 38.4M, 33.6M free. Jul 10 00:21:44.946209 systemd-modules-load[208]: Inserted module 'overlay' Jul 10 00:21:44.959336 systemd[1]: Started systemd-journald.service - Journal Service. Jul 10 00:21:44.961423 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:21:44.967060 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 10 00:21:44.973067 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 10 00:21:44.978168 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 10 00:21:44.988042 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 10 00:21:44.990893 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 10 00:21:44.996755 systemd-modules-load[208]: Inserted module 'br_netfilter' Jul 10 00:21:45.001673 kernel: Bridge firewalling registered Jul 10 00:21:45.001621 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 10 00:21:45.004537 systemd-tmpfiles[224]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jul 10 00:21:45.008712 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jul 10 00:21:45.014090 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 10 00:21:45.018934 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 10 00:21:45.026784 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 10 00:21:45.031300 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 10 00:21:45.033551 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jul 10 00:21:45.043009 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:21:45.045841 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 10 00:21:45.061928 dracut-cmdline[243]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=844005237fb9709f65a093d5533c4229fb6c54e8e257736d9c3d041b6d3080ea Jul 10 00:21:45.106757 systemd-resolved[247]: Positive Trust Anchors: Jul 10 00:21:45.106966 systemd-resolved[247]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 10 00:21:45.107032 systemd-resolved[247]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 10 00:21:45.116961 systemd-resolved[247]: Defaulting to hostname 'linux'. Jul 10 00:21:45.118993 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 10 00:21:45.119958 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 10 00:21:45.162905 kernel: SCSI subsystem initialized Jul 10 00:21:45.172892 kernel: Loading iSCSI transport class v2.0-870. Jul 10 00:21:45.184891 kernel: iscsi: registered transport (tcp) Jul 10 00:21:45.207446 kernel: iscsi: registered transport (qla4xxx) Jul 10 00:21:45.207536 kernel: QLogic iSCSI HBA Driver Jul 10 00:21:45.226377 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 10 00:21:45.249831 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 10 00:21:45.252083 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 10 00:21:45.297483 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jul 10 00:21:45.299540 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jul 10 00:21:45.356908 kernel: raid6: avx512x4 gen() 17816 MB/s Jul 10 00:21:45.374898 kernel: raid6: avx512x2 gen() 17808 MB/s Jul 10 00:21:45.392893 kernel: raid6: avx512x1 gen() 17708 MB/s Jul 10 00:21:45.410895 kernel: raid6: avx2x4 gen() 17707 MB/s Jul 10 00:21:45.428895 kernel: raid6: avx2x2 gen() 16645 MB/s Jul 10 00:21:45.447387 kernel: raid6: avx2x1 gen() 13582 MB/s Jul 10 00:21:45.447475 kernel: raid6: using algorithm avx512x4 gen() 17816 MB/s Jul 10 00:21:45.466160 kernel: raid6: .... xor() 7462 MB/s, rmw enabled Jul 10 00:21:45.466236 kernel: raid6: using avx512x2 recovery algorithm Jul 10 00:21:45.487901 kernel: xor: automatically using best checksumming function avx Jul 10 00:21:45.656896 kernel: Btrfs loaded, zoned=no, fsverity=no Jul 10 00:21:45.663845 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jul 10 00:21:45.665959 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 10 00:21:45.696452 systemd-udevd[456]: Using default interface naming scheme 'v255'. Jul 10 00:21:45.703274 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 10 00:21:45.707912 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jul 10 00:21:45.735542 dracut-pre-trigger[463]: rd.md=0: removing MD RAID activation Jul 10 00:21:45.762334 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jul 10 00:21:45.764290 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 10 00:21:45.831422 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 10 00:21:45.834999 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jul 10 00:21:45.921889 kernel: cryptd: max_cpu_qlen set to 1000 Jul 10 00:21:45.936276 kernel: ena 0000:00:05.0: ENA device version: 0.10 Jul 10 00:21:45.936543 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Jul 10 00:21:45.943898 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Jul 10 00:21:45.956889 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:ea:18:b3:42:7d Jul 10 00:21:45.957172 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input2 Jul 10 00:21:45.970814 kernel: AES CTR mode by8 optimization enabled Jul 10 00:21:45.972271 (udev-worker)[502]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:21:45.977885 kernel: nvme nvme0: pci function 0000:00:04.0 Jul 10 00:21:45.978128 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Jul 10 00:21:45.990028 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 10 00:21:45.991029 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:21:45.992626 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:21:45.998185 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:21:46.005771 kernel: nvme nvme0: 2/0/0 default/read/poll queues Jul 10 00:21:46.005363 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Jul 10 00:21:46.019790 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 10 00:21:46.019952 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:21:46.021965 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:21:46.038947 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jul 10 00:21:46.038973 kernel: GPT:9289727 != 16777215 Jul 10 00:21:46.038984 kernel: GPT:Alternate GPT header not at the end of the disk. Jul 10 00:21:46.038996 kernel: GPT:9289727 != 16777215 Jul 10 00:21:46.039006 kernel: GPT: Use GNU Parted to correct GPT errors. Jul 10 00:21:46.039023 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jul 10 00:21:46.061890 kernel: nvme nvme0: using unchecked data buffer Jul 10 00:21:46.072513 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:21:46.137324 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Jul 10 00:21:46.168929 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Jul 10 00:21:46.169668 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Jul 10 00:21:46.173030 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jul 10 00:21:46.197301 disk-uuid[686]: Primary Header is updated. Jul 10 00:21:46.197301 disk-uuid[686]: Secondary Entries is updated. Jul 10 00:21:46.197301 disk-uuid[686]: Secondary Header is updated. Jul 10 00:21:46.218452 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Jul 10 00:21:46.237223 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Jul 10 00:21:46.354988 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jul 10 00:21:46.356701 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jul 10 00:21:46.357339 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 10 00:21:46.358534 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 10 00:21:46.360261 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jul 10 00:21:46.381589 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jul 10 00:21:47.222924 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jul 10 00:21:47.223524 disk-uuid[690]: The operation has completed successfully. Jul 10 00:21:47.372128 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 10 00:21:47.372255 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jul 10 00:21:47.400937 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jul 10 00:21:47.419597 sh[878]: Success Jul 10 00:21:47.447108 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 10 00:21:47.447187 kernel: device-mapper: uevent: version 1.0.3 Jul 10 00:21:47.451891 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jul 10 00:21:47.460885 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" Jul 10 00:21:47.607117 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jul 10 00:21:47.609857 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jul 10 00:21:47.618673 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jul 10 00:21:47.635933 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' Jul 10 00:21:47.636009 kernel: BTRFS: device fsid c4cb30b0-bb74-4f98-aab6-7a1c6f47edee devid 1 transid 36 /dev/mapper/usr (254:0) scanned by mount (901) Jul 10 00:21:47.643770 kernel: BTRFS info (device dm-0): first mount of filesystem c4cb30b0-bb74-4f98-aab6-7a1c6f47edee Jul 10 00:21:47.643837 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:21:47.643851 kernel: BTRFS info (device dm-0): using free-space-tree Jul 10 00:21:47.694914 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jul 10 00:21:47.695942 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jul 10 00:21:47.696557 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jul 10 00:21:47.697459 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jul 10 00:21:47.700002 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jul 10 00:21:47.733893 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (934) Jul 10 00:21:47.738974 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:21:47.739041 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:21:47.739054 kernel: BTRFS info (device nvme0n1p6): using free-space-tree Jul 10 00:21:47.761961 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:21:47.763636 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jul 10 00:21:47.766987 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jul 10 00:21:47.800535 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 10 00:21:47.803151 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 10 00:21:47.839967 systemd-networkd[1070]: lo: Link UP Jul 10 00:21:47.839977 systemd-networkd[1070]: lo: Gained carrier Jul 10 00:21:47.841672 systemd-networkd[1070]: Enumeration completed Jul 10 00:21:47.842214 systemd-networkd[1070]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 10 00:21:47.842219 systemd-networkd[1070]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 10 00:21:47.843597 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 10 00:21:47.844370 systemd[1]: Reached target network.target - Network. Jul 10 00:21:47.846744 systemd-networkd[1070]: eth0: Link UP Jul 10 00:21:47.846750 systemd-networkd[1070]: eth0: Gained carrier Jul 10 00:21:47.846769 systemd-networkd[1070]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 10 00:21:47.856982 systemd-networkd[1070]: eth0: DHCPv4 address 172.31.20.136/20, gateway 172.31.16.1 acquired from 172.31.16.1 Jul 10 00:21:48.126375 ignition[1027]: Ignition 2.21.0 Jul 10 00:21:48.126393 ignition[1027]: Stage: fetch-offline Jul 10 00:21:48.126599 ignition[1027]: no configs at "/usr/lib/ignition/base.d" Jul 10 00:21:48.126607 ignition[1027]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 10 00:21:48.128084 ignition[1027]: Ignition finished successfully Jul 10 00:21:48.129970 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jul 10 00:21:48.131508 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jul 10 00:21:48.161280 ignition[1080]: Ignition 2.21.0 Jul 10 00:21:48.161296 ignition[1080]: Stage: fetch Jul 10 00:21:48.161670 ignition[1080]: no configs at "/usr/lib/ignition/base.d" Jul 10 00:21:48.161683 ignition[1080]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 10 00:21:48.161800 ignition[1080]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 10 00:21:48.220413 ignition[1080]: PUT result: OK Jul 10 00:21:48.231131 ignition[1080]: parsed url from cmdline: "" Jul 10 00:21:48.231144 ignition[1080]: no config URL provided Jul 10 00:21:48.231155 ignition[1080]: reading system config file "/usr/lib/ignition/user.ign" Jul 10 00:21:48.231172 ignition[1080]: no config at "/usr/lib/ignition/user.ign" Jul 10 00:21:48.231199 ignition[1080]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 10 00:21:48.238659 ignition[1080]: PUT result: OK Jul 10 00:21:48.238773 ignition[1080]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Jul 10 00:21:48.241888 ignition[1080]: GET result: OK Jul 10 00:21:48.241979 ignition[1080]: parsing config with SHA512: e7e23a1f0fc4ea242b464db494c1f098ccbeee2ac517cd0cf87b530d7d8ed9d71e705c68286f1e39fd4a7cbd3a66c2971e779c9d487b5d2def9a00bbd925daff Jul 10 00:21:48.245810 unknown[1080]: fetched base config from "system" Jul 10 00:21:48.245857 unknown[1080]: fetched base config from "system" Jul 10 00:21:48.246221 ignition[1080]: fetch: fetch complete Jul 10 00:21:48.245893 unknown[1080]: fetched user config from "aws" Jul 10 00:21:48.246226 ignition[1080]: fetch: fetch passed Jul 10 00:21:48.246277 ignition[1080]: Ignition finished successfully Jul 10 00:21:48.248276 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jul 10 00:21:48.250694 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jul 10 00:21:48.286600 ignition[1087]: Ignition 2.21.0 Jul 10 00:21:48.286618 ignition[1087]: Stage: kargs Jul 10 00:21:48.287007 ignition[1087]: no configs at "/usr/lib/ignition/base.d" Jul 10 00:21:48.287021 ignition[1087]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 10 00:21:48.287138 ignition[1087]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 10 00:21:48.288594 ignition[1087]: PUT result: OK Jul 10 00:21:48.292089 ignition[1087]: kargs: kargs passed Jul 10 00:21:48.292648 ignition[1087]: Ignition finished successfully Jul 10 00:21:48.294604 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jul 10 00:21:48.296041 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jul 10 00:21:48.318110 ignition[1094]: Ignition 2.21.0 Jul 10 00:21:48.318126 ignition[1094]: Stage: disks Jul 10 00:21:48.318520 ignition[1094]: no configs at "/usr/lib/ignition/base.d" Jul 10 00:21:48.318535 ignition[1094]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 10 00:21:48.320425 ignition[1094]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 10 00:21:48.322447 ignition[1094]: PUT result: OK Jul 10 00:21:48.325823 ignition[1094]: disks: disks passed Jul 10 00:21:48.325908 ignition[1094]: Ignition finished successfully Jul 10 00:21:48.327044 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jul 10 00:21:48.327857 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jul 10 00:21:48.328488 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 10 00:21:48.328815 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 10 00:21:48.329568 systemd[1]: Reached target sysinit.target - System Initialization. Jul 10 00:21:48.330126 systemd[1]: Reached target basic.target - Basic System. Jul 10 00:21:48.331667 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jul 10 00:21:48.377557 systemd-fsck[1102]: ROOT: clean, 15/553520 files, 52789/553472 blocks Jul 10 00:21:48.380163 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jul 10 00:21:48.382696 systemd[1]: Mounting sysroot.mount - /sysroot... Jul 10 00:21:48.539892 kernel: EXT4-fs (nvme0n1p9): mounted filesystem a310c019-7915-47f5-9fce-db4a09ac26c2 r/w with ordered data mode. Quota mode: none. Jul 10 00:21:48.540787 systemd[1]: Mounted sysroot.mount - /sysroot. Jul 10 00:21:48.541937 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jul 10 00:21:48.544099 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 10 00:21:48.546539 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jul 10 00:21:48.549433 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jul 10 00:21:48.550602 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 10 00:21:48.550640 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jul 10 00:21:48.558886 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jul 10 00:21:48.561006 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jul 10 00:21:48.575901 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1121) Jul 10 00:21:48.580921 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:21:48.580981 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:21:48.581003 kernel: BTRFS info (device nvme0n1p6): using free-space-tree Jul 10 00:21:48.589325 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 10 00:21:48.780684 initrd-setup-root[1145]: cut: /sysroot/etc/passwd: No such file or directory Jul 10 00:21:48.785711 initrd-setup-root[1152]: cut: /sysroot/etc/group: No such file or directory Jul 10 00:21:48.791156 initrd-setup-root[1159]: cut: /sysroot/etc/shadow: No such file or directory Jul 10 00:21:48.796495 initrd-setup-root[1166]: cut: /sysroot/etc/gshadow: No such file or directory Jul 10 00:21:48.937759 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jul 10 00:21:48.939769 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jul 10 00:21:48.943081 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jul 10 00:21:48.957495 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jul 10 00:21:48.959892 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:21:48.985695 ignition[1233]: INFO : Ignition 2.21.0 Jul 10 00:21:48.987120 ignition[1233]: INFO : Stage: mount Jul 10 00:21:48.987120 ignition[1233]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 10 00:21:48.987120 ignition[1233]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 10 00:21:48.987120 ignition[1233]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 10 00:21:48.987120 ignition[1233]: INFO : PUT result: OK Jul 10 00:21:48.990158 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jul 10 00:21:48.990891 ignition[1233]: INFO : mount: mount passed Jul 10 00:21:48.991730 ignition[1233]: INFO : Ignition finished successfully Jul 10 00:21:48.992137 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jul 10 00:21:48.994605 systemd[1]: Starting ignition-files.service - Ignition (files)... Jul 10 00:21:49.016846 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 10 00:21:49.055888 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1246) Jul 10 00:21:49.058977 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:21:49.059050 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:21:49.061577 kernel: BTRFS info (device nvme0n1p6): using free-space-tree Jul 10 00:21:49.069243 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 10 00:21:49.103477 ignition[1262]: INFO : Ignition 2.21.0 Jul 10 00:21:49.103477 ignition[1262]: INFO : Stage: files Jul 10 00:21:49.105115 ignition[1262]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 10 00:21:49.105115 ignition[1262]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 10 00:21:49.105115 ignition[1262]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 10 00:21:49.106748 ignition[1262]: INFO : PUT result: OK Jul 10 00:21:49.110050 ignition[1262]: DEBUG : files: compiled without relabeling support, skipping Jul 10 00:21:49.111085 ignition[1262]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 10 00:21:49.111085 ignition[1262]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 10 00:21:49.115916 ignition[1262]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 10 00:21:49.116661 ignition[1262]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 10 00:21:49.116661 ignition[1262]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 10 00:21:49.116345 unknown[1262]: wrote ssh authorized keys file for user: core Jul 10 00:21:49.120900 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jul 10 00:21:49.122037 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Jul 10 00:21:49.205826 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jul 10 00:21:49.396459 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jul 10 00:21:49.396459 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 10 00:21:49.398585 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Jul 10 00:21:49.651145 systemd-networkd[1070]: eth0: Gained IPv6LL Jul 10 00:21:49.848509 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Jul 10 00:21:49.974390 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 10 00:21:49.974390 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Jul 10 00:21:49.976185 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Jul 10 00:21:49.976185 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 10 00:21:49.976185 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 10 00:21:49.976185 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 10 00:21:49.976185 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 10 00:21:49.976185 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 10 00:21:49.976185 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 10 00:21:49.981662 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 10 00:21:49.981662 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 10 00:21:49.981662 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 10 00:21:49.983995 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 10 00:21:49.983995 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 10 00:21:49.983995 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Jul 10 00:21:50.710808 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Jul 10 00:21:53.302904 ignition[1262]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 10 00:21:53.304120 ignition[1262]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Jul 10 00:21:53.305951 ignition[1262]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 10 00:21:53.310375 ignition[1262]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 10 00:21:53.310375 ignition[1262]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Jul 10 00:21:53.310375 ignition[1262]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" Jul 10 00:21:53.312837 ignition[1262]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" Jul 10 00:21:53.312837 ignition[1262]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 10 00:21:53.312837 ignition[1262]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 10 00:21:53.312837 ignition[1262]: INFO : files: files passed Jul 10 00:21:53.312837 ignition[1262]: INFO : Ignition finished successfully Jul 10 00:21:53.312116 systemd[1]: Finished ignition-files.service - Ignition (files). Jul 10 00:21:53.314467 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jul 10 00:21:53.317985 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jul 10 00:21:53.328453 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 10 00:21:53.328561 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jul 10 00:21:53.336890 initrd-setup-root-after-ignition[1293]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 10 00:21:53.336890 initrd-setup-root-after-ignition[1293]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jul 10 00:21:53.339386 initrd-setup-root-after-ignition[1297]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 10 00:21:53.339422 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 10 00:21:53.340030 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jul 10 00:21:53.341709 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jul 10 00:21:53.383010 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 10 00:21:53.383152 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jul 10 00:21:53.384638 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jul 10 00:21:53.385751 systemd[1]: Reached target initrd.target - Initrd Default Target. Jul 10 00:21:53.386644 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jul 10 00:21:53.387835 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jul 10 00:21:53.409894 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 10 00:21:53.411926 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jul 10 00:21:53.429180 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jul 10 00:21:53.430069 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 10 00:21:53.430791 systemd[1]: Stopped target timers.target - Timer Units. Jul 10 00:21:53.431555 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 10 00:21:53.431717 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 10 00:21:53.432743 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jul 10 00:21:53.434011 systemd[1]: Stopped target basic.target - Basic System. Jul 10 00:21:53.434640 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jul 10 00:21:53.435307 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jul 10 00:21:53.435939 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jul 10 00:21:53.436662 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jul 10 00:21:53.437491 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jul 10 00:21:53.438258 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jul 10 00:21:53.438949 systemd[1]: Stopped target sysinit.target - System Initialization. Jul 10 00:21:53.439989 systemd[1]: Stopped target local-fs.target - Local File Systems. Jul 10 00:21:53.440703 systemd[1]: Stopped target swap.target - Swaps. Jul 10 00:21:53.441569 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 10 00:21:53.441721 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jul 10 00:21:53.442675 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jul 10 00:21:53.443486 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 10 00:21:53.444070 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jul 10 00:21:53.444170 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 10 00:21:53.444785 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 10 00:21:53.444949 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jul 10 00:21:53.446528 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 10 00:21:53.446650 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 10 00:21:53.447249 systemd[1]: ignition-files.service: Deactivated successfully. Jul 10 00:21:53.447374 systemd[1]: Stopped ignition-files.service - Ignition (files). Jul 10 00:21:53.449948 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jul 10 00:21:53.451976 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 10 00:21:53.452117 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jul 10 00:21:53.465372 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jul 10 00:21:53.467996 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 10 00:21:53.468295 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jul 10 00:21:53.469510 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 10 00:21:53.469715 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jul 10 00:21:53.479175 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 10 00:21:53.480919 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jul 10 00:21:53.493302 ignition[1317]: INFO : Ignition 2.21.0 Jul 10 00:21:53.494973 ignition[1317]: INFO : Stage: umount Jul 10 00:21:53.494973 ignition[1317]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 10 00:21:53.494973 ignition[1317]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 10 00:21:53.494973 ignition[1317]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 10 00:21:53.497776 ignition[1317]: INFO : PUT result: OK Jul 10 00:21:53.500679 ignition[1317]: INFO : umount: umount passed Jul 10 00:21:53.500679 ignition[1317]: INFO : Ignition finished successfully Jul 10 00:21:53.503875 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 10 00:21:53.504675 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 10 00:21:53.504815 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jul 10 00:21:53.506282 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 10 00:21:53.506354 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jul 10 00:21:53.507696 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 10 00:21:53.507768 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jul 10 00:21:53.508917 systemd[1]: ignition-fetch.service: Deactivated successfully. Jul 10 00:21:53.509052 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jul 10 00:21:53.509639 systemd[1]: Stopped target network.target - Network. Jul 10 00:21:53.510289 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 10 00:21:53.510360 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jul 10 00:21:53.510990 systemd[1]: Stopped target paths.target - Path Units. Jul 10 00:21:53.511566 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 10 00:21:53.514965 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 10 00:21:53.515350 systemd[1]: Stopped target slices.target - Slice Units. Jul 10 00:21:53.516263 systemd[1]: Stopped target sockets.target - Socket Units. Jul 10 00:21:53.516918 systemd[1]: iscsid.socket: Deactivated successfully. Jul 10 00:21:53.516972 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jul 10 00:21:53.517626 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 10 00:21:53.517675 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 10 00:21:53.518232 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 10 00:21:53.518310 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jul 10 00:21:53.518879 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jul 10 00:21:53.518937 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jul 10 00:21:53.519635 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jul 10 00:21:53.520242 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jul 10 00:21:53.521412 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 10 00:21:53.521537 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jul 10 00:21:53.522708 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 10 00:21:53.522816 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jul 10 00:21:53.527189 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 10 00:21:53.527345 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jul 10 00:21:53.530817 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Jul 10 00:21:53.531140 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 10 00:21:53.531272 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jul 10 00:21:53.533607 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Jul 10 00:21:53.534386 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jul 10 00:21:53.535276 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 10 00:21:53.535345 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jul 10 00:21:53.537187 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jul 10 00:21:53.537710 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 10 00:21:53.537777 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 10 00:21:53.538410 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 10 00:21:53.538460 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:21:53.542288 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 10 00:21:53.542350 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jul 10 00:21:53.543355 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jul 10 00:21:53.543414 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 10 00:21:53.544160 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 10 00:21:53.546313 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Jul 10 00:21:53.546397 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Jul 10 00:21:53.553337 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 10 00:21:53.553548 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 10 00:21:53.556651 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 10 00:21:53.556728 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jul 10 00:21:53.557349 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 10 00:21:53.557394 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jul 10 00:21:53.558076 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 10 00:21:53.558139 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jul 10 00:21:53.559248 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 10 00:21:53.559305 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jul 10 00:21:53.560070 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 10 00:21:53.560116 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 10 00:21:53.563669 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jul 10 00:21:53.565424 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jul 10 00:21:53.565505 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jul 10 00:21:53.568693 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jul 10 00:21:53.568768 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 10 00:21:53.570156 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 10 00:21:53.570220 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:21:53.573139 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Jul 10 00:21:53.573225 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Jul 10 00:21:53.573280 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Jul 10 00:21:53.573790 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 10 00:21:53.573996 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jul 10 00:21:53.582579 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 10 00:21:53.582733 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jul 10 00:21:53.583823 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jul 10 00:21:53.585387 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jul 10 00:21:53.598787 systemd[1]: Switching root. Jul 10 00:21:53.647678 systemd-journald[207]: Journal stopped Jul 10 00:21:55.089251 systemd-journald[207]: Received SIGTERM from PID 1 (systemd). Jul 10 00:21:55.089325 kernel: SELinux: policy capability network_peer_controls=1 Jul 10 00:21:55.089346 kernel: SELinux: policy capability open_perms=1 Jul 10 00:21:55.089358 kernel: SELinux: policy capability extended_socket_class=1 Jul 10 00:21:55.089369 kernel: SELinux: policy capability always_check_network=0 Jul 10 00:21:55.089385 kernel: SELinux: policy capability cgroup_seclabel=1 Jul 10 00:21:55.089908 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jul 10 00:21:55.089922 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jul 10 00:21:55.089934 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jul 10 00:21:55.089945 kernel: SELinux: policy capability userspace_initial_context=0 Jul 10 00:21:55.089964 kernel: audit: type=1403 audit(1752106913.965:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jul 10 00:21:55.089979 systemd[1]: Successfully loaded SELinux policy in 43.327ms. Jul 10 00:21:55.089994 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 10.654ms. Jul 10 00:21:55.090008 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 10 00:21:55.090021 systemd[1]: Detected virtualization amazon. Jul 10 00:21:55.090033 systemd[1]: Detected architecture x86-64. Jul 10 00:21:55.090045 systemd[1]: Detected first boot. Jul 10 00:21:55.090057 systemd[1]: Initializing machine ID from VM UUID. Jul 10 00:21:55.090070 zram_generator::config[1361]: No configuration found. Jul 10 00:21:55.090086 kernel: Guest personality initialized and is inactive Jul 10 00:21:55.090098 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Jul 10 00:21:55.090109 kernel: Initialized host personality Jul 10 00:21:55.090120 kernel: NET: Registered PF_VSOCK protocol family Jul 10 00:21:55.090132 systemd[1]: Populated /etc with preset unit settings. Jul 10 00:21:55.090146 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Jul 10 00:21:55.090163 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jul 10 00:21:55.090176 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jul 10 00:21:55.090188 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jul 10 00:21:55.090204 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jul 10 00:21:55.090216 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jul 10 00:21:55.090229 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jul 10 00:21:55.090241 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jul 10 00:21:55.090254 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jul 10 00:21:55.090266 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jul 10 00:21:55.090278 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jul 10 00:21:55.090290 systemd[1]: Created slice user.slice - User and Session Slice. Jul 10 00:21:55.090306 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 10 00:21:55.090318 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 10 00:21:55.090334 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jul 10 00:21:55.090347 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jul 10 00:21:55.090360 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jul 10 00:21:55.090372 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 10 00:21:55.090384 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jul 10 00:21:55.090399 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 10 00:21:55.090412 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 10 00:21:55.090423 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jul 10 00:21:55.090435 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jul 10 00:21:55.090448 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jul 10 00:21:55.090460 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jul 10 00:21:55.090472 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 10 00:21:55.090484 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 10 00:21:55.090498 systemd[1]: Reached target slices.target - Slice Units. Jul 10 00:21:55.090510 systemd[1]: Reached target swap.target - Swaps. Jul 10 00:21:55.090525 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jul 10 00:21:55.090538 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jul 10 00:21:55.090550 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jul 10 00:21:55.090562 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 10 00:21:55.090575 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 10 00:21:55.090587 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 10 00:21:55.090600 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jul 10 00:21:55.090612 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jul 10 00:21:55.090624 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jul 10 00:21:55.090639 systemd[1]: Mounting media.mount - External Media Directory... Jul 10 00:21:55.090651 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:55.090663 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jul 10 00:21:55.090676 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jul 10 00:21:55.090688 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jul 10 00:21:55.090700 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jul 10 00:21:55.090712 systemd[1]: Reached target machines.target - Containers. Jul 10 00:21:55.090724 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jul 10 00:21:55.090739 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 10 00:21:55.090752 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 10 00:21:55.090764 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jul 10 00:21:55.090776 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 10 00:21:55.090788 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 10 00:21:55.090801 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 10 00:21:55.090814 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jul 10 00:21:55.090826 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 10 00:21:55.090838 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jul 10 00:21:55.090853 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jul 10 00:21:55.091900 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jul 10 00:21:55.091922 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jul 10 00:21:55.091935 systemd[1]: Stopped systemd-fsck-usr.service. Jul 10 00:21:55.091949 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 10 00:21:55.091961 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 10 00:21:55.091974 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 10 00:21:55.091986 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 10 00:21:55.092002 kernel: loop: module loaded Jul 10 00:21:55.092015 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jul 10 00:21:55.092028 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jul 10 00:21:55.092040 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 10 00:21:55.092058 systemd[1]: verity-setup.service: Deactivated successfully. Jul 10 00:21:55.092070 systemd[1]: Stopped verity-setup.service. Jul 10 00:21:55.092083 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:55.092096 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jul 10 00:21:55.092108 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jul 10 00:21:55.092120 systemd[1]: Mounted media.mount - External Media Directory. Jul 10 00:21:55.092136 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jul 10 00:21:55.092149 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jul 10 00:21:55.092161 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jul 10 00:21:55.092174 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 10 00:21:55.092186 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jul 10 00:21:55.092198 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jul 10 00:21:55.092210 kernel: ACPI: bus type drm_connector registered Jul 10 00:21:55.092221 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 10 00:21:55.092234 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 10 00:21:55.092249 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 10 00:21:55.092261 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 10 00:21:55.092274 kernel: fuse: init (API version 7.41) Jul 10 00:21:55.092286 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 10 00:21:55.092298 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 10 00:21:55.092311 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jul 10 00:21:55.092323 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jul 10 00:21:55.092336 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 10 00:21:55.092348 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 10 00:21:55.092394 systemd-journald[1447]: Collecting audit messages is disabled. Jul 10 00:21:55.092422 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 10 00:21:55.092434 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 10 00:21:55.092447 systemd-journald[1447]: Journal started Jul 10 00:21:55.092474 systemd-journald[1447]: Runtime Journal (/run/log/journal/ec207223f90241a2490db0b13b4ab104) is 4.8M, max 38.4M, 33.6M free. Jul 10 00:21:55.092517 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jul 10 00:21:54.789453 systemd[1]: Queued start job for default target multi-user.target. Jul 10 00:21:54.802639 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Jul 10 00:21:54.803135 systemd[1]: systemd-journald.service: Deactivated successfully. Jul 10 00:21:55.094912 systemd[1]: Started systemd-journald.service - Journal Service. Jul 10 00:21:55.097962 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jul 10 00:21:55.112724 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 10 00:21:55.117023 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jul 10 00:21:55.121231 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jul 10 00:21:55.121803 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jul 10 00:21:55.121841 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 10 00:21:55.123414 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jul 10 00:21:55.131053 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jul 10 00:21:55.132107 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 10 00:21:55.135059 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jul 10 00:21:55.137986 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jul 10 00:21:55.138453 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 10 00:21:55.141051 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jul 10 00:21:55.141529 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 10 00:21:55.143110 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 10 00:21:55.150040 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jul 10 00:21:55.153924 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jul 10 00:21:55.154587 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jul 10 00:21:55.155133 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jul 10 00:21:55.162010 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jul 10 00:21:55.173261 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jul 10 00:21:55.173835 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jul 10 00:21:55.180279 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jul 10 00:21:55.186400 systemd-journald[1447]: Time spent on flushing to /var/log/journal/ec207223f90241a2490db0b13b4ab104 is 60.972ms for 1020 entries. Jul 10 00:21:55.186400 systemd-journald[1447]: System Journal (/var/log/journal/ec207223f90241a2490db0b13b4ab104) is 8M, max 195.6M, 187.6M free. Jul 10 00:21:55.254872 systemd-journald[1447]: Received client request to flush runtime journal. Jul 10 00:21:55.254926 kernel: loop0: detected capacity change from 0 to 229808 Jul 10 00:21:55.209155 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:21:55.229570 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 10 00:21:55.258034 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jul 10 00:21:55.269263 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jul 10 00:21:55.279999 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jul 10 00:21:55.284003 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 10 00:21:55.323824 systemd-tmpfiles[1510]: ACLs are not supported, ignoring. Jul 10 00:21:55.324627 systemd-tmpfiles[1510]: ACLs are not supported, ignoring. Jul 10 00:21:55.335100 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 10 00:21:55.378898 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jul 10 00:21:55.408924 kernel: loop1: detected capacity change from 0 to 72352 Jul 10 00:21:55.555898 kernel: loop2: detected capacity change from 0 to 146240 Jul 10 00:21:55.686887 kernel: loop3: detected capacity change from 0 to 113872 Jul 10 00:21:55.801891 kernel: loop4: detected capacity change from 0 to 229808 Jul 10 00:21:55.804605 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jul 10 00:21:55.835897 kernel: loop5: detected capacity change from 0 to 72352 Jul 10 00:21:55.849931 kernel: loop6: detected capacity change from 0 to 146240 Jul 10 00:21:55.865916 kernel: loop7: detected capacity change from 0 to 113872 Jul 10 00:21:55.875272 (sd-merge)[1518]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Jul 10 00:21:55.876946 (sd-merge)[1518]: Merged extensions into '/usr'. Jul 10 00:21:55.881507 systemd[1]: Reload requested from client PID 1494 ('systemd-sysext') (unit systemd-sysext.service)... Jul 10 00:21:55.881635 systemd[1]: Reloading... Jul 10 00:21:55.958898 zram_generator::config[1540]: No configuration found. Jul 10 00:21:56.085543 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:21:56.123661 ldconfig[1489]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jul 10 00:21:56.184684 systemd[1]: Reloading finished in 302 ms. Jul 10 00:21:56.207363 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jul 10 00:21:56.208248 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jul 10 00:21:56.209273 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jul 10 00:21:56.221441 systemd[1]: Starting ensure-sysext.service... Jul 10 00:21:56.226054 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 10 00:21:56.236175 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 10 00:21:56.247365 systemd[1]: Reload requested from client PID 1597 ('systemctl') (unit ensure-sysext.service)... Jul 10 00:21:56.247384 systemd[1]: Reloading... Jul 10 00:21:56.262277 systemd-tmpfiles[1598]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jul 10 00:21:56.262306 systemd-tmpfiles[1598]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jul 10 00:21:56.262539 systemd-tmpfiles[1598]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jul 10 00:21:56.262788 systemd-tmpfiles[1598]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jul 10 00:21:56.263631 systemd-tmpfiles[1598]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jul 10 00:21:56.263903 systemd-tmpfiles[1598]: ACLs are not supported, ignoring. Jul 10 00:21:56.263965 systemd-tmpfiles[1598]: ACLs are not supported, ignoring. Jul 10 00:21:56.273306 systemd-tmpfiles[1598]: Detected autofs mount point /boot during canonicalization of boot. Jul 10 00:21:56.273321 systemd-tmpfiles[1598]: Skipping /boot Jul 10 00:21:56.288165 systemd-udevd[1599]: Using default interface naming scheme 'v255'. Jul 10 00:21:56.293051 systemd-tmpfiles[1598]: Detected autofs mount point /boot during canonicalization of boot. Jul 10 00:21:56.293067 systemd-tmpfiles[1598]: Skipping /boot Jul 10 00:21:56.336932 zram_generator::config[1622]: No configuration found. Jul 10 00:21:56.487173 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:21:56.522154 (udev-worker)[1641]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:21:56.572309 kernel: mousedev: PS/2 mouse device common for all mice Jul 10 00:21:56.591888 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jul 10 00:21:56.643893 kernel: ACPI: button: Power Button [PWRF] Jul 10 00:21:56.643486 systemd[1]: Reloading finished in 395 ms. Jul 10 00:21:56.645885 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input4 Jul 10 00:21:56.654976 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 10 00:21:56.656902 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 10 00:21:56.678877 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Jul 10 00:21:56.686882 kernel: ACPI: button: Sleep Button [SLPF] Jul 10 00:21:56.683519 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jul 10 00:21:56.704937 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 10 00:21:56.708896 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jul 10 00:21:56.713523 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jul 10 00:21:56.716283 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 10 00:21:56.724972 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 10 00:21:56.726839 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jul 10 00:21:56.731557 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:56.731750 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 10 00:21:56.735206 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 10 00:21:56.740673 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 10 00:21:56.749115 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 10 00:21:56.749751 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 10 00:21:56.751234 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 10 00:21:56.751338 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:56.758058 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jul 10 00:21:56.759913 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:56.761167 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 10 00:21:56.761569 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 10 00:21:56.761651 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 10 00:21:56.761735 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:56.767073 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:56.768664 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 10 00:21:56.778031 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 10 00:21:56.779110 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 10 00:21:56.779256 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 10 00:21:56.779420 systemd[1]: Reached target time-set.target - System Time Set. Jul 10 00:21:56.780395 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:21:56.792734 systemd[1]: Finished ensure-sysext.service. Jul 10 00:21:56.796281 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jul 10 00:21:56.813846 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 10 00:21:56.815223 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 10 00:21:56.816349 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 10 00:21:56.825355 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 10 00:21:56.825964 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 10 00:21:56.834417 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jul 10 00:21:56.840410 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jul 10 00:21:56.845264 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 10 00:21:56.845908 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 10 00:21:56.849747 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 10 00:21:56.849924 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 10 00:21:56.850508 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 10 00:21:56.872929 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jul 10 00:21:56.873584 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 10 00:21:56.886206 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jul 10 00:21:56.892665 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Jul 10 00:21:56.901052 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jul 10 00:21:56.902914 augenrules[1846]: No rules Jul 10 00:21:56.904534 systemd[1]: audit-rules.service: Deactivated successfully. Jul 10 00:21:56.904758 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 10 00:21:56.935751 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jul 10 00:21:56.962153 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:21:56.980359 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 10 00:21:56.980558 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:21:56.982136 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:21:57.007768 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jul 10 00:21:57.101749 systemd-networkd[1768]: lo: Link UP Jul 10 00:21:57.102058 systemd-networkd[1768]: lo: Gained carrier Jul 10 00:21:57.103492 systemd-networkd[1768]: Enumeration completed Jul 10 00:21:57.103691 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 10 00:21:57.105320 systemd-networkd[1768]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 10 00:21:57.105407 systemd-networkd[1768]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 10 00:21:57.106913 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jul 10 00:21:57.108382 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jul 10 00:21:57.113210 systemd-networkd[1768]: eth0: Link UP Jul 10 00:21:57.113346 systemd-networkd[1768]: eth0: Gained carrier Jul 10 00:21:57.113374 systemd-networkd[1768]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 10 00:21:57.114963 systemd-resolved[1769]: Positive Trust Anchors: Jul 10 00:21:57.114976 systemd-resolved[1769]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 10 00:21:57.115015 systemd-resolved[1769]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 10 00:21:57.119857 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:21:57.122719 systemd-networkd[1768]: eth0: DHCPv4 address 172.31.20.136/20, gateway 172.31.16.1 acquired from 172.31.16.1 Jul 10 00:21:57.124915 systemd-resolved[1769]: Defaulting to hostname 'linux'. Jul 10 00:21:57.127362 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 10 00:21:57.127979 systemd[1]: Reached target network.target - Network. Jul 10 00:21:57.128474 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 10 00:21:57.129066 systemd[1]: Reached target sysinit.target - System Initialization. Jul 10 00:21:57.129579 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jul 10 00:21:57.130027 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jul 10 00:21:57.130425 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Jul 10 00:21:57.130950 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jul 10 00:21:57.131472 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jul 10 00:21:57.131775 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jul 10 00:21:57.132095 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jul 10 00:21:57.132132 systemd[1]: Reached target paths.target - Path Units. Jul 10 00:21:57.132420 systemd[1]: Reached target timers.target - Timer Units. Jul 10 00:21:57.134771 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jul 10 00:21:57.136616 systemd[1]: Starting docker.socket - Docker Socket for the API... Jul 10 00:21:57.139601 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jul 10 00:21:57.140146 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jul 10 00:21:57.140498 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jul 10 00:21:57.142964 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jul 10 00:21:57.143804 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jul 10 00:21:57.145271 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jul 10 00:21:57.145829 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jul 10 00:21:57.147521 systemd[1]: Reached target sockets.target - Socket Units. Jul 10 00:21:57.148067 systemd[1]: Reached target basic.target - Basic System. Jul 10 00:21:57.148514 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jul 10 00:21:57.148555 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jul 10 00:21:57.150000 systemd[1]: Starting containerd.service - containerd container runtime... Jul 10 00:21:57.153134 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jul 10 00:21:57.156050 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jul 10 00:21:57.158083 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jul 10 00:21:57.165187 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jul 10 00:21:57.169135 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jul 10 00:21:57.170630 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jul 10 00:21:57.173284 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Jul 10 00:21:57.178267 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jul 10 00:21:57.185139 systemd[1]: Started ntpd.service - Network Time Service. Jul 10 00:21:57.192733 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jul 10 00:21:57.196769 google_oslogin_nss_cache[1883]: oslogin_cache_refresh[1883]: Refreshing passwd entry cache Jul 10 00:21:57.196696 oslogin_cache_refresh[1883]: Refreshing passwd entry cache Jul 10 00:21:57.197136 systemd[1]: Starting setup-oem.service - Setup OEM... Jul 10 00:21:57.202118 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jul 10 00:21:57.205403 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jul 10 00:21:57.226940 jq[1881]: false Jul 10 00:21:57.230936 google_oslogin_nss_cache[1883]: oslogin_cache_refresh[1883]: Failure getting users, quitting Jul 10 00:21:57.230936 google_oslogin_nss_cache[1883]: oslogin_cache_refresh[1883]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jul 10 00:21:57.230936 google_oslogin_nss_cache[1883]: oslogin_cache_refresh[1883]: Refreshing group entry cache Jul 10 00:21:57.230115 oslogin_cache_refresh[1883]: Failure getting users, quitting Jul 10 00:21:57.230139 oslogin_cache_refresh[1883]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jul 10 00:21:57.230194 oslogin_cache_refresh[1883]: Refreshing group entry cache Jul 10 00:21:57.232049 systemd[1]: Starting systemd-logind.service - User Login Management... Jul 10 00:21:57.234990 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jul 10 00:21:57.235690 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jul 10 00:21:57.237048 oslogin_cache_refresh[1883]: Failure getting groups, quitting Jul 10 00:21:57.255458 google_oslogin_nss_cache[1883]: oslogin_cache_refresh[1883]: Failure getting groups, quitting Jul 10 00:21:57.255458 google_oslogin_nss_cache[1883]: oslogin_cache_refresh[1883]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jul 10 00:21:57.253809 systemd[1]: Starting update-engine.service - Update Engine... Jul 10 00:21:57.237062 oslogin_cache_refresh[1883]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jul 10 00:21:57.265158 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jul 10 00:21:57.267044 extend-filesystems[1882]: Found /dev/nvme0n1p6 Jul 10 00:21:57.285247 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jul 10 00:21:57.289162 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jul 10 00:21:57.289446 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jul 10 00:21:57.289846 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Jul 10 00:21:57.290405 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Jul 10 00:21:57.296917 extend-filesystems[1882]: Found /dev/nvme0n1p9 Jul 10 00:21:57.316954 jq[1899]: true Jul 10 00:21:57.331220 extend-filesystems[1882]: Checking size of /dev/nvme0n1p9 Jul 10 00:21:57.331892 update_engine[1896]: I20250710 00:21:57.326650 1896 main.cc:92] Flatcar Update Engine starting Jul 10 00:21:57.377578 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jul 10 00:21:57.377900 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jul 10 00:21:57.396324 (ntainerd)[1920]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jul 10 00:21:57.404172 jq[1912]: true Jul 10 00:21:57.411970 extend-filesystems[1882]: Resized partition /dev/nvme0n1p9 Jul 10 00:21:57.426894 extend-filesystems[1931]: resize2fs 1.47.2 (1-Jan-2025) Jul 10 00:21:57.436048 systemd[1]: motdgen.service: Deactivated successfully. Jul 10 00:21:57.436375 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jul 10 00:21:57.452613 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Jul 10 00:21:57.464345 tar[1907]: linux-amd64/LICENSE Jul 10 00:21:57.464345 tar[1907]: linux-amd64/helm Jul 10 00:21:57.459893 systemd[1]: Finished setup-oem.service - Setup OEM. Jul 10 00:21:57.465920 ntpd[1885]: ntpd 4.2.8p17@1.4004-o Wed Jul 9 21:35:50 UTC 2025 (1): Starting Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: ntpd 4.2.8p17@1.4004-o Wed Jul 9 21:35:50 UTC 2025 (1): Starting Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: ---------------------------------------------------- Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: ntp-4 is maintained by Network Time Foundation, Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: corporation. Support and training for ntp-4 are Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: available at https://www.nwtime.org/support Jul 10 00:21:57.467211 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: ---------------------------------------------------- Jul 10 00:21:57.465949 ntpd[1885]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Jul 10 00:21:57.474129 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: proto: precision = 0.082 usec (-23) Jul 10 00:21:57.465959 ntpd[1885]: ---------------------------------------------------- Jul 10 00:21:57.465969 ntpd[1885]: ntp-4 is maintained by Network Time Foundation, Jul 10 00:21:57.465978 ntpd[1885]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Jul 10 00:21:57.465987 ntpd[1885]: corporation. Support and training for ntp-4 are Jul 10 00:21:57.465996 ntpd[1885]: available at https://www.nwtime.org/support Jul 10 00:21:57.466007 ntpd[1885]: ---------------------------------------------------- Jul 10 00:21:57.468986 ntpd[1885]: proto: precision = 0.082 usec (-23) Jul 10 00:21:57.472112 ntpd[1885]: basedate set to 2025-06-27 Jul 10 00:21:57.483197 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: basedate set to 2025-06-27 Jul 10 00:21:57.483197 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: gps base set to 2025-06-29 (week 2373) Jul 10 00:21:57.476537 ntpd[1885]: gps base set to 2025-06-29 (week 2373) Jul 10 00:21:57.495039 ntpd[1885]: Listen and drop on 0 v6wildcard [::]:123 Jul 10 00:21:57.495532 dbus-daemon[1879]: [system] SELinux support is enabled Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Listen and drop on 0 v6wildcard [::]:123 Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Listen normally on 2 lo 127.0.0.1:123 Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Listen normally on 3 eth0 172.31.20.136:123 Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Listen normally on 4 lo [::1]:123 Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: bind(21) AF_INET6 fe80::4ea:18ff:feb3:427d%2#123 flags 0x11 failed: Cannot assign requested address Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: unable to create socket on eth0 (5) for fe80::4ea:18ff:feb3:427d%2#123 Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: failed to init interface for address fe80::4ea:18ff:feb3:427d%2 Jul 10 00:21:57.504415 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: Listening on routing socket on fd #21 for interface updates Jul 10 00:21:57.497884 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jul 10 00:21:57.496007 ntpd[1885]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Jul 10 00:21:57.505555 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jul 10 00:21:57.496199 ntpd[1885]: Listen normally on 2 lo 127.0.0.1:123 Jul 10 00:21:57.505593 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jul 10 00:21:57.496235 ntpd[1885]: Listen normally on 3 eth0 172.31.20.136:123 Jul 10 00:21:57.496275 ntpd[1885]: Listen normally on 4 lo [::1]:123 Jul 10 00:21:57.503458 ntpd[1885]: bind(21) AF_INET6 fe80::4ea:18ff:feb3:427d%2#123 flags 0x11 failed: Cannot assign requested address Jul 10 00:21:57.507832 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jul 10 00:21:57.503491 ntpd[1885]: unable to create socket on eth0 (5) for fe80::4ea:18ff:feb3:427d%2#123 Jul 10 00:21:57.507857 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jul 10 00:21:57.503506 ntpd[1885]: failed to init interface for address fe80::4ea:18ff:feb3:427d%2 Jul 10 00:21:57.503548 ntpd[1885]: Listening on routing socket on fd #21 for interface updates Jul 10 00:21:57.522251 dbus-daemon[1879]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1768 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Jul 10 00:21:57.531834 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 10 00:21:57.531834 ntpd[1885]: 10 Jul 00:21:57 ntpd[1885]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 10 00:21:57.525299 ntpd[1885]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 10 00:21:57.525334 ntpd[1885]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 10 00:21:57.537815 dbus-daemon[1879]: [system] Successfully activated service 'org.freedesktop.systemd1' Jul 10 00:21:57.551070 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Jul 10 00:21:57.553480 systemd[1]: Started update-engine.service - Update Engine. Jul 10 00:21:57.562200 update_engine[1896]: I20250710 00:21:57.562129 1896 update_check_scheduler.cc:74] Next update check in 5m10s Jul 10 00:21:57.588498 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jul 10 00:21:57.632899 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Jul 10 00:21:57.665551 extend-filesystems[1931]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Jul 10 00:21:57.665551 extend-filesystems[1931]: old_desc_blocks = 1, new_desc_blocks = 1 Jul 10 00:21:57.665551 extend-filesystems[1931]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Jul 10 00:21:57.668177 extend-filesystems[1882]: Resized filesystem in /dev/nvme0n1p9 Jul 10 00:21:57.669238 systemd[1]: extend-filesystems.service: Deactivated successfully. Jul 10 00:21:57.670928 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jul 10 00:21:57.714335 bash[1953]: Updated "/home/core/.ssh/authorized_keys" Jul 10 00:21:57.705964 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jul 10 00:21:57.714962 systemd[1]: Starting sshkeys.service... Jul 10 00:21:57.780823 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jul 10 00:21:57.785155 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jul 10 00:21:57.790606 coreos-metadata[1878]: Jul 10 00:21:57.790 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Jul 10 00:21:57.792176 coreos-metadata[1878]: Jul 10 00:21:57.792 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Jul 10 00:21:57.792827 coreos-metadata[1878]: Jul 10 00:21:57.792 INFO Fetch successful Jul 10 00:21:57.792827 coreos-metadata[1878]: Jul 10 00:21:57.792 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Jul 10 00:21:57.795774 coreos-metadata[1878]: Jul 10 00:21:57.795 INFO Fetch successful Jul 10 00:21:57.795774 coreos-metadata[1878]: Jul 10 00:21:57.795 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Jul 10 00:21:57.800843 coreos-metadata[1878]: Jul 10 00:21:57.799 INFO Fetch successful Jul 10 00:21:57.800843 coreos-metadata[1878]: Jul 10 00:21:57.800 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Jul 10 00:21:57.800843 coreos-metadata[1878]: Jul 10 00:21:57.800 INFO Fetch successful Jul 10 00:21:57.801085 coreos-metadata[1878]: Jul 10 00:21:57.800 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Jul 10 00:21:57.803218 coreos-metadata[1878]: Jul 10 00:21:57.802 INFO Fetch failed with 404: resource not found Jul 10 00:21:57.803218 coreos-metadata[1878]: Jul 10 00:21:57.802 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Jul 10 00:21:57.816972 coreos-metadata[1878]: Jul 10 00:21:57.805 INFO Fetch successful Jul 10 00:21:57.816972 coreos-metadata[1878]: Jul 10 00:21:57.805 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Jul 10 00:21:57.816972 coreos-metadata[1878]: Jul 10 00:21:57.814 INFO Fetch successful Jul 10 00:21:57.816972 coreos-metadata[1878]: Jul 10 00:21:57.814 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Jul 10 00:21:57.821167 coreos-metadata[1878]: Jul 10 00:21:57.818 INFO Fetch successful Jul 10 00:21:57.821167 coreos-metadata[1878]: Jul 10 00:21:57.818 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Jul 10 00:21:57.832942 coreos-metadata[1878]: Jul 10 00:21:57.830 INFO Fetch successful Jul 10 00:21:57.832942 coreos-metadata[1878]: Jul 10 00:21:57.830 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Jul 10 00:21:57.832942 coreos-metadata[1878]: Jul 10 00:21:57.831 INFO Fetch successful Jul 10 00:21:57.886224 systemd-logind[1894]: Watching system buttons on /dev/input/event2 (Power Button) Jul 10 00:21:57.899448 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jul 10 00:21:57.899798 systemd-logind[1894]: Watching system buttons on /dev/input/event3 (Sleep Button) Jul 10 00:21:57.899833 systemd-logind[1894]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jul 10 00:21:57.903530 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jul 10 00:21:57.904282 systemd-logind[1894]: New seat seat0. Jul 10 00:21:57.905686 systemd[1]: Started systemd-logind.service - User Login Management. Jul 10 00:21:57.964706 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Jul 10 00:21:57.967675 dbus-daemon[1879]: [system] Successfully activated service 'org.freedesktop.hostname1' Jul 10 00:21:57.969044 dbus-daemon[1879]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=1956 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Jul 10 00:21:57.980775 systemd[1]: Starting polkit.service - Authorization Manager... Jul 10 00:21:58.014018 coreos-metadata[2014]: Jul 10 00:21:58.013 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Jul 10 00:21:58.015439 coreos-metadata[2014]: Jul 10 00:21:58.015 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Jul 10 00:21:58.016118 coreos-metadata[2014]: Jul 10 00:21:58.016 INFO Fetch successful Jul 10 00:21:58.016344 coreos-metadata[2014]: Jul 10 00:21:58.016 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Jul 10 00:21:58.017503 coreos-metadata[2014]: Jul 10 00:21:58.017 INFO Fetch successful Jul 10 00:21:58.019232 unknown[2014]: wrote ssh authorized keys file for user: core Jul 10 00:21:58.036923 locksmithd[1957]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jul 10 00:21:58.057293 update-ssh-keys[2059]: Updated "/home/core/.ssh/authorized_keys" Jul 10 00:21:58.059916 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jul 10 00:21:58.067741 systemd[1]: Finished sshkeys.service. Jul 10 00:21:58.287329 polkitd[2052]: Started polkitd version 126 Jul 10 00:21:58.314939 containerd[1920]: time="2025-07-10T00:21:58Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jul 10 00:21:58.323449 containerd[1920]: time="2025-07-10T00:21:58.323399043Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 Jul 10 00:21:58.325346 polkitd[2052]: Loading rules from directory /etc/polkit-1/rules.d Jul 10 00:21:58.328802 polkitd[2052]: Loading rules from directory /run/polkit-1/rules.d Jul 10 00:21:58.328942 polkitd[2052]: Error opening rules directory: Error opening directory “/run/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Jul 10 00:21:58.329407 polkitd[2052]: Loading rules from directory /usr/local/share/polkit-1/rules.d Jul 10 00:21:58.329446 polkitd[2052]: Error opening rules directory: Error opening directory “/usr/local/share/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Jul 10 00:21:58.329510 polkitd[2052]: Loading rules from directory /usr/share/polkit-1/rules.d Jul 10 00:21:58.336668 polkitd[2052]: Finished loading, compiling and executing 2 rules Jul 10 00:21:58.338133 systemd[1]: Started polkit.service - Authorization Manager. Jul 10 00:21:58.341975 dbus-daemon[1879]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Jul 10 00:21:58.344197 polkitd[2052]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Jul 10 00:21:58.377532 systemd-resolved[1769]: System hostname changed to 'ip-172-31-20-136'. Jul 10 00:21:58.377938 systemd-hostnamed[1956]: Hostname set to (transient) Jul 10 00:21:58.387546 containerd[1920]: time="2025-07-10T00:21:58.387493693Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="12.245µs" Jul 10 00:21:58.387546 containerd[1920]: time="2025-07-10T00:21:58.387542390Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jul 10 00:21:58.387683 containerd[1920]: time="2025-07-10T00:21:58.387565518Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jul 10 00:21:58.387780 containerd[1920]: time="2025-07-10T00:21:58.387759803Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jul 10 00:21:58.387824 containerd[1920]: time="2025-07-10T00:21:58.387788762Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jul 10 00:21:58.387859 containerd[1920]: time="2025-07-10T00:21:58.387824384Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jul 10 00:21:58.389983 containerd[1920]: time="2025-07-10T00:21:58.389946933Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390076 containerd[1920]: time="2025-07-10T00:21:58.389985019Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390353 containerd[1920]: time="2025-07-10T00:21:58.390327100Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390399 containerd[1920]: time="2025-07-10T00:21:58.390353682Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390399 containerd[1920]: time="2025-07-10T00:21:58.390370566Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390399 containerd[1920]: time="2025-07-10T00:21:58.390382981Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390510 containerd[1920]: time="2025-07-10T00:21:58.390492877Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390743 containerd[1920]: time="2025-07-10T00:21:58.390720591Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390794 containerd[1920]: time="2025-07-10T00:21:58.390763005Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jul 10 00:21:58.390794 containerd[1920]: time="2025-07-10T00:21:58.390781600Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jul 10 00:21:58.390906 containerd[1920]: time="2025-07-10T00:21:58.390836253Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jul 10 00:21:58.393337 containerd[1920]: time="2025-07-10T00:21:58.393195078Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jul 10 00:21:58.393337 containerd[1920]: time="2025-07-10T00:21:58.393292200Z" level=info msg="metadata content store policy set" policy=shared Jul 10 00:21:58.398269 containerd[1920]: time="2025-07-10T00:21:58.398196210Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jul 10 00:21:58.398269 containerd[1920]: time="2025-07-10T00:21:58.398259820Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398279128Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398295987Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398311689Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398348451Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398368099Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398383811Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398398411Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398411444Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398423834Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398440624Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398570427Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398592798Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398614073Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jul 10 00:21:58.398985 containerd[1920]: time="2025-07-10T00:21:58.398630095Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398645836Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398663959Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398680362Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398702038Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398718511Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398739403Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398755307Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398827935Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jul 10 00:21:58.399435 containerd[1920]: time="2025-07-10T00:21:58.398845433Z" level=info msg="Start snapshots syncer" Jul 10 00:21:58.400973 containerd[1920]: time="2025-07-10T00:21:58.400291621Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jul 10 00:21:58.400973 containerd[1920]: time="2025-07-10T00:21:58.400657185Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jul 10 00:21:58.401210 containerd[1920]: time="2025-07-10T00:21:58.400721343Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jul 10 00:21:58.405878 containerd[1920]: time="2025-07-10T00:21:58.403592564Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jul 10 00:21:58.406172 containerd[1920]: time="2025-07-10T00:21:58.406147683Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jul 10 00:21:58.406267 containerd[1920]: time="2025-07-10T00:21:58.406254797Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jul 10 00:21:58.406348 containerd[1920]: time="2025-07-10T00:21:58.406333308Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jul 10 00:21:58.406419 containerd[1920]: time="2025-07-10T00:21:58.406405720Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jul 10 00:21:58.406490 containerd[1920]: time="2025-07-10T00:21:58.406478094Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jul 10 00:21:58.406552 containerd[1920]: time="2025-07-10T00:21:58.406541307Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jul 10 00:21:58.406615 containerd[1920]: time="2025-07-10T00:21:58.406603788Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jul 10 00:21:58.406699 containerd[1920]: time="2025-07-10T00:21:58.406689100Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jul 10 00:21:58.406762 containerd[1920]: time="2025-07-10T00:21:58.406748561Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jul 10 00:21:58.406838 containerd[1920]: time="2025-07-10T00:21:58.406824028Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jul 10 00:21:58.406973 containerd[1920]: time="2025-07-10T00:21:58.406957091Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.408895587Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.408921384Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.408946143Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.408959211Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.408975834Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.408991928Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.409015954Z" level=info msg="runtime interface created" Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.409023523Z" level=info msg="created NRI interface" Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.409036403Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.409056261Z" level=info msg="Connect containerd service" Jul 10 00:21:58.409884 containerd[1920]: time="2025-07-10T00:21:58.409103238Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jul 10 00:21:58.412105 containerd[1920]: time="2025-07-10T00:21:58.412075742Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 10 00:21:58.466430 ntpd[1885]: bind(24) AF_INET6 fe80::4ea:18ff:feb3:427d%2#123 flags 0x11 failed: Cannot assign requested address Jul 10 00:21:58.466827 ntpd[1885]: 10 Jul 00:21:58 ntpd[1885]: bind(24) AF_INET6 fe80::4ea:18ff:feb3:427d%2#123 flags 0x11 failed: Cannot assign requested address Jul 10 00:21:58.466930 ntpd[1885]: unable to create socket on eth0 (6) for fe80::4ea:18ff:feb3:427d%2#123 Jul 10 00:21:58.467022 ntpd[1885]: 10 Jul 00:21:58 ntpd[1885]: unable to create socket on eth0 (6) for fe80::4ea:18ff:feb3:427d%2#123 Jul 10 00:21:58.467069 ntpd[1885]: failed to init interface for address fe80::4ea:18ff:feb3:427d%2 Jul 10 00:21:58.467128 ntpd[1885]: 10 Jul 00:21:58 ntpd[1885]: failed to init interface for address fe80::4ea:18ff:feb3:427d%2 Jul 10 00:21:58.853952 containerd[1920]: time="2025-07-10T00:21:58.853903796Z" level=info msg="Start subscribing containerd event" Jul 10 00:21:58.854085 containerd[1920]: time="2025-07-10T00:21:58.853973661Z" level=info msg="Start recovering state" Jul 10 00:21:58.854148 containerd[1920]: time="2025-07-10T00:21:58.854083047Z" level=info msg="Start event monitor" Jul 10 00:21:58.854148 containerd[1920]: time="2025-07-10T00:21:58.854099819Z" level=info msg="Start cni network conf syncer for default" Jul 10 00:21:58.854148 containerd[1920]: time="2025-07-10T00:21:58.854110319Z" level=info msg="Start streaming server" Jul 10 00:21:58.854148 containerd[1920]: time="2025-07-10T00:21:58.854126851Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jul 10 00:21:58.854148 containerd[1920]: time="2025-07-10T00:21:58.854137031Z" level=info msg="runtime interface starting up..." Jul 10 00:21:58.854148 containerd[1920]: time="2025-07-10T00:21:58.854145442Z" level=info msg="starting plugins..." Jul 10 00:21:58.854330 containerd[1920]: time="2025-07-10T00:21:58.854162607Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jul 10 00:21:58.857956 containerd[1920]: time="2025-07-10T00:21:58.855073648Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jul 10 00:21:58.858478 containerd[1920]: time="2025-07-10T00:21:58.858091566Z" level=info msg=serving... address=/run/containerd/containerd.sock Jul 10 00:21:58.866487 containerd[1920]: time="2025-07-10T00:21:58.865914905Z" level=info msg="containerd successfully booted in 0.551698s" Jul 10 00:21:58.866086 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jul 10 00:21:58.867018 systemd-networkd[1768]: eth0: Gained IPv6LL Jul 10 00:21:58.867517 systemd[1]: Started containerd.service - containerd container runtime. Jul 10 00:21:58.874657 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jul 10 00:21:58.877447 systemd[1]: Reached target network-online.target - Network is Online. Jul 10 00:21:58.884995 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Jul 10 00:21:58.893992 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:21:58.903180 tar[1907]: linux-amd64/README.md Jul 10 00:21:58.903084 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jul 10 00:21:58.925954 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jul 10 00:21:58.980191 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jul 10 00:21:59.040575 amazon-ssm-agent[2104]: Initializing new seelog logger Jul 10 00:21:59.041911 amazon-ssm-agent[2104]: New Seelog Logger Creation Complete Jul 10 00:21:59.041911 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.041911 amazon-ssm-agent[2104]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.041911 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 processing appconfig overrides Jul 10 00:21:59.042388 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.042447 amazon-ssm-agent[2104]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.042579 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 processing appconfig overrides Jul 10 00:21:59.042964 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.043025 amazon-ssm-agent[2104]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.043145 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 processing appconfig overrides Jul 10 00:21:59.043675 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.0423 INFO Proxy environment variables: Jul 10 00:21:59.045732 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.046265 amazon-ssm-agent[2104]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.046265 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 processing appconfig overrides Jul 10 00:21:59.090791 sshd_keygen[1902]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jul 10 00:21:59.130109 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jul 10 00:21:59.136172 systemd[1]: Starting issuegen.service - Generate /run/issue... Jul 10 00:21:59.140572 systemd[1]: Started sshd@0-172.31.20.136:22-139.178.89.65:40396.service - OpenSSH per-connection server daemon (139.178.89.65:40396). Jul 10 00:21:59.143244 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.0423 INFO http_proxy: Jul 10 00:21:59.171270 systemd[1]: issuegen.service: Deactivated successfully. Jul 10 00:21:59.171530 systemd[1]: Finished issuegen.service - Generate /run/issue. Jul 10 00:21:59.180457 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jul 10 00:21:59.224221 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jul 10 00:21:59.229371 systemd[1]: Started getty@tty1.service - Getty on tty1. Jul 10 00:21:59.236070 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jul 10 00:21:59.237518 systemd[1]: Reached target getty.target - Login Prompts. Jul 10 00:21:59.241682 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.0423 INFO no_proxy: Jul 10 00:21:59.340264 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.0423 INFO https_proxy: Jul 10 00:21:59.401622 sshd[2134]: Accepted publickey for core from 139.178.89.65 port 40396 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:21:59.406738 sshd-session[2134]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:21:59.421937 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jul 10 00:21:59.428174 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jul 10 00:21:59.438494 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.0426 INFO Checking if agent identity type OnPrem can be assumed Jul 10 00:21:59.448187 systemd-logind[1894]: New session 1 of user core. Jul 10 00:21:59.469614 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jul 10 00:21:59.474194 systemd[1]: Starting user@500.service - User Manager for UID 500... Jul 10 00:21:59.496712 (systemd)[2146]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jul 10 00:21:59.499808 systemd-logind[1894]: New session c1 of user core. Jul 10 00:21:59.501818 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.501818 amazon-ssm-agent[2104]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 10 00:21:59.501818 amazon-ssm-agent[2104]: 2025/07/10 00:21:59 processing appconfig overrides Jul 10 00:21:59.529667 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.0428 INFO Checking if agent identity type EC2 can be assumed Jul 10 00:21:59.529667 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1013 INFO Agent will take identity from EC2 Jul 10 00:21:59.529667 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1059 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.3.0.0 Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1059 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1059 INFO [amazon-ssm-agent] Starting Core Agent Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1059 INFO [amazon-ssm-agent] Registrar detected. Attempting registration Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1059 INFO [Registrar] Starting registrar module Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1102 INFO [EC2Identity] Checking disk for registration info Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1102 INFO [EC2Identity] No registration info found for ec2 instance, attempting registration Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.1102 INFO [EC2Identity] Generating registration keypair Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.4596 INFO [EC2Identity] Checking write access before registering Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.4602 INFO [EC2Identity] Registering EC2 instance with Systems Manager Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.5012 INFO [EC2Identity] EC2 registration was successful. Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.5012 INFO [amazon-ssm-agent] Registration attempted. Resuming core agent startup. Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.5013 INFO [CredentialRefresher] credentialRefresher has started Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.5013 INFO [CredentialRefresher] Starting credentials refresher loop Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.5294 INFO EC2RoleProvider Successfully connected with instance profile role credentials Jul 10 00:21:59.529832 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.5296 INFO [CredentialRefresher] Credentials ready Jul 10 00:21:59.537669 amazon-ssm-agent[2104]: 2025-07-10 00:21:59.5297 INFO [CredentialRefresher] Next credential rotation will be in 29.99999419755 minutes Jul 10 00:21:59.672262 systemd[2146]: Queued start job for default target default.target. Jul 10 00:21:59.682408 systemd[2146]: Created slice app.slice - User Application Slice. Jul 10 00:21:59.682914 systemd[2146]: Reached target paths.target - Paths. Jul 10 00:21:59.682972 systemd[2146]: Reached target timers.target - Timers. Jul 10 00:21:59.684903 systemd[2146]: Starting dbus.socket - D-Bus User Message Bus Socket... Jul 10 00:21:59.711726 systemd[2146]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jul 10 00:21:59.711851 systemd[2146]: Reached target sockets.target - Sockets. Jul 10 00:21:59.712059 systemd[2146]: Reached target basic.target - Basic System. Jul 10 00:21:59.712106 systemd[2146]: Reached target default.target - Main User Target. Jul 10 00:21:59.712136 systemd[2146]: Startup finished in 201ms. Jul 10 00:21:59.712226 systemd[1]: Started user@500.service - User Manager for UID 500. Jul 10 00:21:59.722116 systemd[1]: Started session-1.scope - Session 1 of User core. Jul 10 00:21:59.870249 systemd[1]: Started sshd@1-172.31.20.136:22-139.178.89.65:40398.service - OpenSSH per-connection server daemon (139.178.89.65:40398). Jul 10 00:22:00.050935 sshd[2158]: Accepted publickey for core from 139.178.89.65 port 40398 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:00.055345 sshd-session[2158]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:00.065541 systemd-logind[1894]: New session 2 of user core. Jul 10 00:22:00.071115 systemd[1]: Started session-2.scope - Session 2 of User core. Jul 10 00:22:00.193888 sshd[2160]: Connection closed by 139.178.89.65 port 40398 Jul 10 00:22:00.195587 sshd-session[2158]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:00.203384 systemd[1]: sshd@1-172.31.20.136:22-139.178.89.65:40398.service: Deactivated successfully. Jul 10 00:22:00.206715 systemd[1]: session-2.scope: Deactivated successfully. Jul 10 00:22:00.210613 systemd-logind[1894]: Session 2 logged out. Waiting for processes to exit. Jul 10 00:22:00.213794 systemd-logind[1894]: Removed session 2. Jul 10 00:22:00.232373 systemd[1]: Started sshd@2-172.31.20.136:22-139.178.89.65:40414.service - OpenSSH per-connection server daemon (139.178.89.65:40414). Jul 10 00:22:00.426944 sshd[2166]: Accepted publickey for core from 139.178.89.65 port 40414 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:00.429574 sshd-session[2166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:00.437836 systemd-logind[1894]: New session 3 of user core. Jul 10 00:22:00.445263 systemd[1]: Started session-3.scope - Session 3 of User core. Jul 10 00:22:00.551362 amazon-ssm-agent[2104]: 2025-07-10 00:22:00.5511 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Jul 10 00:22:00.571058 sshd[2168]: Connection closed by 139.178.89.65 port 40414 Jul 10 00:22:00.573338 sshd-session[2166]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:00.582716 systemd[1]: sshd@2-172.31.20.136:22-139.178.89.65:40414.service: Deactivated successfully. Jul 10 00:22:00.586810 systemd[1]: session-3.scope: Deactivated successfully. Jul 10 00:22:00.590035 systemd-logind[1894]: Session 3 logged out. Waiting for processes to exit. Jul 10 00:22:00.591557 systemd-logind[1894]: Removed session 3. Jul 10 00:22:00.652788 amazon-ssm-agent[2104]: 2025-07-10 00:22:00.5541 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2172) started Jul 10 00:22:00.753378 amazon-ssm-agent[2104]: 2025-07-10 00:22:00.5542 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Jul 10 00:22:01.466470 ntpd[1885]: Listen normally on 7 eth0 [fe80::4ea:18ff:feb3:427d%2]:123 Jul 10 00:22:01.467635 ntpd[1885]: 10 Jul 00:22:01 ntpd[1885]: Listen normally on 7 eth0 [fe80::4ea:18ff:feb3:427d%2]:123 Jul 10 00:22:01.595480 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:01.596578 systemd[1]: Reached target multi-user.target - Multi-User System. Jul 10 00:22:01.597891 systemd[1]: Startup finished in 2.850s (kernel) + 9.294s (initrd) + 7.672s (userspace) = 19.816s. Jul 10 00:22:01.606644 (kubelet)[2191]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 10 00:22:02.797298 kubelet[2191]: E0710 00:22:02.797168 2191 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 10 00:22:02.800253 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 10 00:22:02.800412 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 10 00:22:02.800714 systemd[1]: kubelet.service: Consumed 1.104s CPU time, 268.8M memory peak. Jul 10 00:22:05.780239 systemd-resolved[1769]: Clock change detected. Flushing caches. Jul 10 00:22:11.924809 systemd[1]: Started sshd@3-172.31.20.136:22-139.178.89.65:48404.service - OpenSSH per-connection server daemon (139.178.89.65:48404). Jul 10 00:22:12.103198 sshd[2203]: Accepted publickey for core from 139.178.89.65 port 48404 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:12.104614 sshd-session[2203]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:12.110261 systemd-logind[1894]: New session 4 of user core. Jul 10 00:22:12.116370 systemd[1]: Started session-4.scope - Session 4 of User core. Jul 10 00:22:12.238404 sshd[2205]: Connection closed by 139.178.89.65 port 48404 Jul 10 00:22:12.238917 sshd-session[2203]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:12.242598 systemd[1]: sshd@3-172.31.20.136:22-139.178.89.65:48404.service: Deactivated successfully. Jul 10 00:22:12.244586 systemd[1]: session-4.scope: Deactivated successfully. Jul 10 00:22:12.245608 systemd-logind[1894]: Session 4 logged out. Waiting for processes to exit. Jul 10 00:22:12.247560 systemd-logind[1894]: Removed session 4. Jul 10 00:22:12.275397 systemd[1]: Started sshd@4-172.31.20.136:22-139.178.89.65:48420.service - OpenSSH per-connection server daemon (139.178.89.65:48420). Jul 10 00:22:12.446239 sshd[2211]: Accepted publickey for core from 139.178.89.65 port 48420 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:12.447589 sshd-session[2211]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:12.453117 systemd-logind[1894]: New session 5 of user core. Jul 10 00:22:12.459356 systemd[1]: Started session-5.scope - Session 5 of User core. Jul 10 00:22:12.577549 sshd[2213]: Connection closed by 139.178.89.65 port 48420 Jul 10 00:22:12.578310 sshd-session[2211]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:12.582286 systemd[1]: sshd@4-172.31.20.136:22-139.178.89.65:48420.service: Deactivated successfully. Jul 10 00:22:12.583844 systemd[1]: session-5.scope: Deactivated successfully. Jul 10 00:22:12.584678 systemd-logind[1894]: Session 5 logged out. Waiting for processes to exit. Jul 10 00:22:12.586433 systemd-logind[1894]: Removed session 5. Jul 10 00:22:12.608200 systemd[1]: Started sshd@5-172.31.20.136:22-139.178.89.65:48428.service - OpenSSH per-connection server daemon (139.178.89.65:48428). Jul 10 00:22:12.786143 sshd[2219]: Accepted publickey for core from 139.178.89.65 port 48428 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:12.787793 sshd-session[2219]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:12.794958 systemd-logind[1894]: New session 6 of user core. Jul 10 00:22:12.800342 systemd[1]: Started session-6.scope - Session 6 of User core. Jul 10 00:22:12.919070 sshd[2221]: Connection closed by 139.178.89.65 port 48428 Jul 10 00:22:12.919942 sshd-session[2219]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:12.923386 systemd[1]: sshd@5-172.31.20.136:22-139.178.89.65:48428.service: Deactivated successfully. Jul 10 00:22:12.925293 systemd[1]: session-6.scope: Deactivated successfully. Jul 10 00:22:12.927857 systemd-logind[1894]: Session 6 logged out. Waiting for processes to exit. Jul 10 00:22:12.929142 systemd-logind[1894]: Removed session 6. Jul 10 00:22:12.963691 systemd[1]: Started sshd@6-172.31.20.136:22-139.178.89.65:48432.service - OpenSSH per-connection server daemon (139.178.89.65:48432). Jul 10 00:22:13.139078 sshd[2227]: Accepted publickey for core from 139.178.89.65 port 48432 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:13.140488 sshd-session[2227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:13.145944 systemd-logind[1894]: New session 7 of user core. Jul 10 00:22:13.152334 systemd[1]: Started session-7.scope - Session 7 of User core. Jul 10 00:22:13.265205 sudo[2230]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jul 10 00:22:13.265583 sudo[2230]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:22:13.287040 sudo[2230]: pam_unix(sudo:session): session closed for user root Jul 10 00:22:13.310281 sshd[2229]: Connection closed by 139.178.89.65 port 48432 Jul 10 00:22:13.310989 sshd-session[2227]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:13.314593 systemd[1]: sshd@6-172.31.20.136:22-139.178.89.65:48432.service: Deactivated successfully. Jul 10 00:22:13.316233 systemd[1]: session-7.scope: Deactivated successfully. Jul 10 00:22:13.318260 systemd-logind[1894]: Session 7 logged out. Waiting for processes to exit. Jul 10 00:22:13.319195 systemd-logind[1894]: Removed session 7. Jul 10 00:22:13.344222 systemd[1]: Started sshd@7-172.31.20.136:22-139.178.89.65:48434.service - OpenSSH per-connection server daemon (139.178.89.65:48434). Jul 10 00:22:13.513060 sshd[2236]: Accepted publickey for core from 139.178.89.65 port 48434 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:13.514589 sshd-session[2236]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:13.520089 systemd-logind[1894]: New session 8 of user core. Jul 10 00:22:13.526431 systemd[1]: Started session-8.scope - Session 8 of User core. Jul 10 00:22:13.625237 sudo[2240]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jul 10 00:22:13.625505 sudo[2240]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:22:13.630495 sudo[2240]: pam_unix(sudo:session): session closed for user root Jul 10 00:22:13.636266 sudo[2239]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jul 10 00:22:13.636543 sudo[2239]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:22:13.646683 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 10 00:22:13.687062 augenrules[2262]: No rules Jul 10 00:22:13.688250 systemd[1]: audit-rules.service: Deactivated successfully. Jul 10 00:22:13.688458 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 10 00:22:13.689665 sudo[2239]: pam_unix(sudo:session): session closed for user root Jul 10 00:22:13.711650 sshd[2238]: Connection closed by 139.178.89.65 port 48434 Jul 10 00:22:13.712178 sshd-session[2236]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:13.715671 systemd[1]: sshd@7-172.31.20.136:22-139.178.89.65:48434.service: Deactivated successfully. Jul 10 00:22:13.717504 systemd[1]: session-8.scope: Deactivated successfully. Jul 10 00:22:13.719599 systemd-logind[1894]: Session 8 logged out. Waiting for processes to exit. Jul 10 00:22:13.720573 systemd-logind[1894]: Removed session 8. Jul 10 00:22:13.744640 systemd[1]: Started sshd@8-172.31.20.136:22-139.178.89.65:48448.service - OpenSSH per-connection server daemon (139.178.89.65:48448). Jul 10 00:22:13.926204 sshd[2271]: Accepted publickey for core from 139.178.89.65 port 48448 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:22:13.927445 sshd-session[2271]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:22:13.932725 systemd-logind[1894]: New session 9 of user core. Jul 10 00:22:13.935313 systemd[1]: Started session-9.scope - Session 9 of User core. Jul 10 00:22:14.038341 sudo[2274]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jul 10 00:22:14.038619 sudo[2274]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:22:14.275746 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jul 10 00:22:14.279326 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:22:14.483447 systemd[1]: Starting docker.service - Docker Application Container Engine... Jul 10 00:22:14.497702 (dockerd)[2296]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jul 10 00:22:14.575795 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:14.586617 (kubelet)[2301]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 10 00:22:14.643744 kubelet[2301]: E0710 00:22:14.643695 2301 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 10 00:22:14.651768 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 10 00:22:14.651956 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 10 00:22:14.652648 systemd[1]: kubelet.service: Consumed 206ms CPU time, 107.7M memory peak. Jul 10 00:22:14.817627 dockerd[2296]: time="2025-07-10T00:22:14.817565044Z" level=info msg="Starting up" Jul 10 00:22:14.818936 dockerd[2296]: time="2025-07-10T00:22:14.818550623Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jul 10 00:22:14.887500 dockerd[2296]: time="2025-07-10T00:22:14.887388743Z" level=info msg="Loading containers: start." Jul 10 00:22:14.902132 kernel: Initializing XFRM netlink socket Jul 10 00:22:15.134401 (udev-worker)[2329]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:22:15.178275 systemd-networkd[1768]: docker0: Link UP Jul 10 00:22:15.188565 dockerd[2296]: time="2025-07-10T00:22:15.188488664Z" level=info msg="Loading containers: done." Jul 10 00:22:15.211678 dockerd[2296]: time="2025-07-10T00:22:15.211578455Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jul 10 00:22:15.211885 dockerd[2296]: time="2025-07-10T00:22:15.211716652Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 Jul 10 00:22:15.211885 dockerd[2296]: time="2025-07-10T00:22:15.211875099Z" level=info msg="Initializing buildkit" Jul 10 00:22:15.259277 dockerd[2296]: time="2025-07-10T00:22:15.259223374Z" level=info msg="Completed buildkit initialization" Jul 10 00:22:15.263627 dockerd[2296]: time="2025-07-10T00:22:15.263581767Z" level=info msg="Daemon has completed initialization" Jul 10 00:22:15.263627 dockerd[2296]: time="2025-07-10T00:22:15.263641139Z" level=info msg="API listen on /run/docker.sock" Jul 10 00:22:15.263835 systemd[1]: Started docker.service - Docker Application Container Engine. Jul 10 00:22:16.443245 containerd[1920]: time="2025-07-10T00:22:16.443197614Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.2\"" Jul 10 00:22:17.027336 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4209481790.mount: Deactivated successfully. Jul 10 00:22:18.550765 containerd[1920]: time="2025-07-10T00:22:18.550711326Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:18.551669 containerd[1920]: time="2025-07-10T00:22:18.551630681Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.2: active requests=0, bytes read=30079099" Jul 10 00:22:18.552785 containerd[1920]: time="2025-07-10T00:22:18.552739849Z" level=info msg="ImageCreate event name:\"sha256:ee794efa53d856b7e291320be3cd6390fa2e113c3f258a21290bc27fc214233e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:18.555292 containerd[1920]: time="2025-07-10T00:22:18.555242918Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:e8ae58675899e946fabe38425f2b3bfd33120b7930d05b5898de97c81a7f6137\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:18.556204 containerd[1920]: time="2025-07-10T00:22:18.555961591Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.2\" with image id \"sha256:ee794efa53d856b7e291320be3cd6390fa2e113c3f258a21290bc27fc214233e\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:e8ae58675899e946fabe38425f2b3bfd33120b7930d05b5898de97c81a7f6137\", size \"30075899\" in 2.112721364s" Jul 10 00:22:18.556204 containerd[1920]: time="2025-07-10T00:22:18.555992075Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.2\" returns image reference \"sha256:ee794efa53d856b7e291320be3cd6390fa2e113c3f258a21290bc27fc214233e\"" Jul 10 00:22:18.556735 containerd[1920]: time="2025-07-10T00:22:18.556632466Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.2\"" Jul 10 00:22:20.321203 containerd[1920]: time="2025-07-10T00:22:20.321150370Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:20.323587 containerd[1920]: time="2025-07-10T00:22:20.323354643Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.2: active requests=0, bytes read=26018946" Jul 10 00:22:20.326222 containerd[1920]: time="2025-07-10T00:22:20.326077928Z" level=info msg="ImageCreate event name:\"sha256:ff4f56c76b82d6cda0555115a0fe479d5dd612264b85efb9cc14b1b4b937bdf2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:20.333432 containerd[1920]: time="2025-07-10T00:22:20.331200061Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:2236e72a4be5dcc9c04600353ff8849db1557f5364947c520ff05471ae719081\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:20.334874 containerd[1920]: time="2025-07-10T00:22:20.334662171Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.2\" with image id \"sha256:ff4f56c76b82d6cda0555115a0fe479d5dd612264b85efb9cc14b1b4b937bdf2\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:2236e72a4be5dcc9c04600353ff8849db1557f5364947c520ff05471ae719081\", size \"27646507\" in 1.77783619s" Jul 10 00:22:20.334874 containerd[1920]: time="2025-07-10T00:22:20.334711277Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.2\" returns image reference \"sha256:ff4f56c76b82d6cda0555115a0fe479d5dd612264b85efb9cc14b1b4b937bdf2\"" Jul 10 00:22:20.335302 containerd[1920]: time="2025-07-10T00:22:20.335275203Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.2\"" Jul 10 00:22:22.062121 containerd[1920]: time="2025-07-10T00:22:22.061885074Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:22.064346 containerd[1920]: time="2025-07-10T00:22:22.064249419Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.2: active requests=0, bytes read=20155055" Jul 10 00:22:22.067836 containerd[1920]: time="2025-07-10T00:22:22.067237601Z" level=info msg="ImageCreate event name:\"sha256:cfed1ff7489289d4e8d796b0d95fd251990403510563cf843912f42ab9718a7b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:22.072436 containerd[1920]: time="2025-07-10T00:22:22.072373028Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:304c28303133be7d927973bc9bd6c83945b3735c59d283c25b63d5b9ed53bca3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:22.073119 containerd[1920]: time="2025-07-10T00:22:22.073073992Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.2\" with image id \"sha256:cfed1ff7489289d4e8d796b0d95fd251990403510563cf843912f42ab9718a7b\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:304c28303133be7d927973bc9bd6c83945b3735c59d283c25b63d5b9ed53bca3\", size \"21782634\" in 1.737668662s" Jul 10 00:22:22.073211 containerd[1920]: time="2025-07-10T00:22:22.073123591Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.2\" returns image reference \"sha256:cfed1ff7489289d4e8d796b0d95fd251990403510563cf843912f42ab9718a7b\"" Jul 10 00:22:22.073640 containerd[1920]: time="2025-07-10T00:22:22.073614268Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.2\"" Jul 10 00:22:23.218501 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3493839685.mount: Deactivated successfully. Jul 10 00:22:23.831398 containerd[1920]: time="2025-07-10T00:22:23.831340415Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:23.833490 containerd[1920]: time="2025-07-10T00:22:23.833417204Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.2: active requests=0, bytes read=31892746" Jul 10 00:22:23.836138 containerd[1920]: time="2025-07-10T00:22:23.835795611Z" level=info msg="ImageCreate event name:\"sha256:661d404f36f01cd854403fd3540f18dcf0342d22bd9c6516bb9de234ac183b19\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:23.839130 containerd[1920]: time="2025-07-10T00:22:23.839054856Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:4796ef3e43efa5ed2a5b015c18f81d3c2fe3aea36f555ea643cc01827eb65e51\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:23.839945 containerd[1920]: time="2025-07-10T00:22:23.839593935Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.2\" with image id \"sha256:661d404f36f01cd854403fd3540f18dcf0342d22bd9c6516bb9de234ac183b19\", repo tag \"registry.k8s.io/kube-proxy:v1.33.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:4796ef3e43efa5ed2a5b015c18f81d3c2fe3aea36f555ea643cc01827eb65e51\", size \"31891765\" in 1.765945565s" Jul 10 00:22:23.839945 containerd[1920]: time="2025-07-10T00:22:23.839631263Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.2\" returns image reference \"sha256:661d404f36f01cd854403fd3540f18dcf0342d22bd9c6516bb9de234ac183b19\"" Jul 10 00:22:23.840245 containerd[1920]: time="2025-07-10T00:22:23.840223436Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Jul 10 00:22:24.431445 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3106913200.mount: Deactivated successfully. Jul 10 00:22:24.775395 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jul 10 00:22:24.777522 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:22:25.091327 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:25.103923 (kubelet)[2641]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 10 00:22:25.190563 kubelet[2641]: E0710 00:22:25.190518 2641 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 10 00:22:25.196843 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 10 00:22:25.197037 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 10 00:22:25.197763 systemd[1]: kubelet.service: Consumed 205ms CPU time, 109.7M memory peak. Jul 10 00:22:25.674634 containerd[1920]: time="2025-07-10T00:22:25.674581774Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:25.677247 containerd[1920]: time="2025-07-10T00:22:25.677206682Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Jul 10 00:22:25.682303 containerd[1920]: time="2025-07-10T00:22:25.682239065Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:25.686156 containerd[1920]: time="2025-07-10T00:22:25.686096233Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:25.687512 containerd[1920]: time="2025-07-10T00:22:25.687339590Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.84700118s" Jul 10 00:22:25.687512 containerd[1920]: time="2025-07-10T00:22:25.687378815Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Jul 10 00:22:25.688135 containerd[1920]: time="2025-07-10T00:22:25.688117332Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jul 10 00:22:26.206381 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1845758255.mount: Deactivated successfully. Jul 10 00:22:26.221311 containerd[1920]: time="2025-07-10T00:22:26.221233934Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 10 00:22:26.223632 containerd[1920]: time="2025-07-10T00:22:26.223567234Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Jul 10 00:22:26.226084 containerd[1920]: time="2025-07-10T00:22:26.225858305Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 10 00:22:26.229130 containerd[1920]: time="2025-07-10T00:22:26.229043743Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 10 00:22:26.229730 containerd[1920]: time="2025-07-10T00:22:26.229589933Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 541.376619ms" Jul 10 00:22:26.229730 containerd[1920]: time="2025-07-10T00:22:26.229624763Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jul 10 00:22:26.230607 containerd[1920]: time="2025-07-10T00:22:26.230548532Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Jul 10 00:22:26.787748 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount167201026.mount: Deactivated successfully. Jul 10 00:22:29.191381 containerd[1920]: time="2025-07-10T00:22:29.191239043Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:29.193133 containerd[1920]: time="2025-07-10T00:22:29.193060760Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58247175" Jul 10 00:22:29.195526 containerd[1920]: time="2025-07-10T00:22:29.195462921Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:29.199418 containerd[1920]: time="2025-07-10T00:22:29.199351782Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:29.202285 containerd[1920]: time="2025-07-10T00:22:29.202230591Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 2.971649742s" Jul 10 00:22:29.202884 containerd[1920]: time="2025-07-10T00:22:29.202463422Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Jul 10 00:22:29.725509 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Jul 10 00:22:32.911698 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:32.911967 systemd[1]: kubelet.service: Consumed 205ms CPU time, 109.7M memory peak. Jul 10 00:22:32.915126 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:22:32.948895 systemd[1]: Reload requested from client PID 2741 ('systemctl') (unit session-9.scope)... Jul 10 00:22:32.948913 systemd[1]: Reloading... Jul 10 00:22:33.103147 zram_generator::config[2785]: No configuration found. Jul 10 00:22:33.250733 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:22:33.386773 systemd[1]: Reloading finished in 437 ms. Jul 10 00:22:33.454824 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jul 10 00:22:33.454943 systemd[1]: kubelet.service: Failed with result 'signal'. Jul 10 00:22:33.455291 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:33.455356 systemd[1]: kubelet.service: Consumed 142ms CPU time, 98.2M memory peak. Jul 10 00:22:33.457534 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:22:33.713255 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:33.723024 (kubelet)[2848]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 10 00:22:33.791937 kubelet[2848]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:22:33.795232 kubelet[2848]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jul 10 00:22:33.795232 kubelet[2848]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:22:33.799339 kubelet[2848]: I0710 00:22:33.798412 2848 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 10 00:22:34.692127 kubelet[2848]: I0710 00:22:34.691351 2848 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jul 10 00:22:34.692127 kubelet[2848]: I0710 00:22:34.691385 2848 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 10 00:22:34.692127 kubelet[2848]: I0710 00:22:34.691737 2848 server.go:956] "Client rotation is on, will bootstrap in background" Jul 10 00:22:34.772351 kubelet[2848]: I0710 00:22:34.772314 2848 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 10 00:22:34.778785 kubelet[2848]: E0710 00:22:34.778726 2848 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.31.20.136:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jul 10 00:22:34.809618 kubelet[2848]: I0710 00:22:34.809579 2848 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jul 10 00:22:34.821215 kubelet[2848]: I0710 00:22:34.821168 2848 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 10 00:22:34.826489 kubelet[2848]: I0710 00:22:34.826406 2848 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 10 00:22:34.830438 kubelet[2848]: I0710 00:22:34.826480 2848 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-20-136","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 10 00:22:34.830438 kubelet[2848]: I0710 00:22:34.830442 2848 topology_manager.go:138] "Creating topology manager with none policy" Jul 10 00:22:34.830895 kubelet[2848]: I0710 00:22:34.830462 2848 container_manager_linux.go:303] "Creating device plugin manager" Jul 10 00:22:34.831772 kubelet[2848]: I0710 00:22:34.831728 2848 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:22:34.837599 kubelet[2848]: I0710 00:22:34.837429 2848 kubelet.go:480] "Attempting to sync node with API server" Jul 10 00:22:34.837599 kubelet[2848]: I0710 00:22:34.837498 2848 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 10 00:22:34.837599 kubelet[2848]: I0710 00:22:34.837536 2848 kubelet.go:386] "Adding apiserver pod source" Jul 10 00:22:34.840046 kubelet[2848]: I0710 00:22:34.840010 2848 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 10 00:22:34.851198 kubelet[2848]: E0710 00:22:34.850918 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.20.136:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-136&limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jul 10 00:22:34.851198 kubelet[2848]: I0710 00:22:34.851050 2848 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Jul 10 00:22:34.854268 kubelet[2848]: I0710 00:22:34.854240 2848 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jul 10 00:22:34.855328 kubelet[2848]: W0710 00:22:34.855289 2848 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jul 10 00:22:34.860733 kubelet[2848]: I0710 00:22:34.860477 2848 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jul 10 00:22:34.860733 kubelet[2848]: I0710 00:22:34.860558 2848 server.go:1289] "Started kubelet" Jul 10 00:22:34.866488 kubelet[2848]: E0710 00:22:34.866386 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.20.136:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jul 10 00:22:34.867093 kubelet[2848]: I0710 00:22:34.867035 2848 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 10 00:22:34.869935 kubelet[2848]: I0710 00:22:34.869481 2848 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 10 00:22:34.880017 kubelet[2848]: I0710 00:22:34.879969 2848 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 10 00:22:34.881195 kubelet[2848]: E0710 00:22:34.877435 2848 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.20.136:6443/api/v1/namespaces/default/events\": dial tcp 172.31.20.136:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-20-136.1850bbfe0cf3de21 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-20-136,UID:ip-172-31-20-136,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-20-136,},FirstTimestamp:2025-07-10 00:22:34.860518945 +0000 UTC m=+1.132119825,LastTimestamp:2025-07-10 00:22:34.860518945 +0000 UTC m=+1.132119825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-20-136,}" Jul 10 00:22:34.882557 kubelet[2848]: I0710 00:22:34.882301 2848 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jul 10 00:22:34.887726 kubelet[2848]: I0710 00:22:34.887693 2848 server.go:317] "Adding debug handlers to kubelet server" Jul 10 00:22:34.892174 kubelet[2848]: I0710 00:22:34.892134 2848 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 10 00:22:34.896214 kubelet[2848]: I0710 00:22:34.894249 2848 volume_manager.go:297] "Starting Kubelet Volume Manager" Jul 10 00:22:34.896214 kubelet[2848]: E0710 00:22:34.894576 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:34.898476 kubelet[2848]: I0710 00:22:34.898450 2848 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jul 10 00:22:34.898760 kubelet[2848]: I0710 00:22:34.898746 2848 reconciler.go:26] "Reconciler: start to sync state" Jul 10 00:22:34.899352 kubelet[2848]: E0710 00:22:34.899317 2848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.136:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-136?timeout=10s\": dial tcp 172.31.20.136:6443: connect: connection refused" interval="200ms" Jul 10 00:22:34.899623 kubelet[2848]: E0710 00:22:34.899599 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.20.136:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jul 10 00:22:34.916334 kubelet[2848]: I0710 00:22:34.916304 2848 factory.go:223] Registration of the containerd container factory successfully Jul 10 00:22:34.916334 kubelet[2848]: I0710 00:22:34.916328 2848 factory.go:223] Registration of the systemd container factory successfully Jul 10 00:22:34.916524 kubelet[2848]: I0710 00:22:34.916411 2848 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 10 00:22:34.926644 kubelet[2848]: E0710 00:22:34.926611 2848 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 10 00:22:34.928854 kubelet[2848]: I0710 00:22:34.928678 2848 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jul 10 00:22:34.930905 kubelet[2848]: I0710 00:22:34.930873 2848 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jul 10 00:22:34.930905 kubelet[2848]: I0710 00:22:34.930905 2848 status_manager.go:230] "Starting to sync pod status with apiserver" Jul 10 00:22:34.931065 kubelet[2848]: I0710 00:22:34.930930 2848 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jul 10 00:22:34.931065 kubelet[2848]: I0710 00:22:34.930938 2848 kubelet.go:2436] "Starting kubelet main sync loop" Jul 10 00:22:34.931065 kubelet[2848]: E0710 00:22:34.930985 2848 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 10 00:22:34.939679 kubelet[2848]: E0710 00:22:34.939643 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.20.136:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jul 10 00:22:34.948450 kubelet[2848]: I0710 00:22:34.948335 2848 cpu_manager.go:221] "Starting CPU manager" policy="none" Jul 10 00:22:34.948450 kubelet[2848]: I0710 00:22:34.948356 2848 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jul 10 00:22:34.948450 kubelet[2848]: I0710 00:22:34.948376 2848 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:22:34.956073 kubelet[2848]: I0710 00:22:34.956028 2848 policy_none.go:49] "None policy: Start" Jul 10 00:22:34.956073 kubelet[2848]: I0710 00:22:34.956077 2848 memory_manager.go:186] "Starting memorymanager" policy="None" Jul 10 00:22:34.956291 kubelet[2848]: I0710 00:22:34.956116 2848 state_mem.go:35] "Initializing new in-memory state store" Jul 10 00:22:34.974546 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jul 10 00:22:34.987137 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jul 10 00:22:34.991183 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jul 10 00:22:34.995304 kubelet[2848]: E0710 00:22:34.995271 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:34.999215 kubelet[2848]: E0710 00:22:34.998397 2848 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jul 10 00:22:34.999215 kubelet[2848]: I0710 00:22:34.998657 2848 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 10 00:22:34.999215 kubelet[2848]: I0710 00:22:34.998671 2848 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 10 00:22:34.999808 kubelet[2848]: I0710 00:22:34.999793 2848 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 10 00:22:35.003336 kubelet[2848]: E0710 00:22:35.003307 2848 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jul 10 00:22:35.003539 kubelet[2848]: E0710 00:22:35.003355 2848 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-20-136\" not found" Jul 10 00:22:35.082000 systemd[1]: Created slice kubepods-burstable-poddefe058535f9ca4593940d9cc8ce8a2a.slice - libcontainer container kubepods-burstable-poddefe058535f9ca4593940d9cc8ce8a2a.slice. Jul 10 00:22:35.100131 kubelet[2848]: I0710 00:22:35.099929 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/defe058535f9ca4593940d9cc8ce8a2a-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-136\" (UID: \"defe058535f9ca4593940d9cc8ce8a2a\") " pod="kube-system/kube-scheduler-ip-172-31-20-136" Jul 10 00:22:35.100131 kubelet[2848]: I0710 00:22:35.099968 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d381d0aee293bda79b931f4d2b629d35-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-136\" (UID: \"d381d0aee293bda79b931f4d2b629d35\") " pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:35.100131 kubelet[2848]: I0710 00:22:35.099996 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d381d0aee293bda79b931f4d2b629d35-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-136\" (UID: \"d381d0aee293bda79b931f4d2b629d35\") " pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:35.100131 kubelet[2848]: I0710 00:22:35.100022 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:35.100131 kubelet[2848]: I0710 00:22:35.100043 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d381d0aee293bda79b931f4d2b629d35-ca-certs\") pod \"kube-apiserver-ip-172-31-20-136\" (UID: \"d381d0aee293bda79b931f4d2b629d35\") " pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:35.100437 kubelet[2848]: I0710 00:22:35.100065 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:35.100437 kubelet[2848]: I0710 00:22:35.100088 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:35.100840 kubelet[2848]: I0710 00:22:35.100803 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:35.101599 kubelet[2848]: E0710 00:22:35.101149 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:35.101599 kubelet[2848]: E0710 00:22:35.101490 2848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.136:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-136?timeout=10s\": dial tcp 172.31.20.136:6443: connect: connection refused" interval="400ms" Jul 10 00:22:35.102204 kubelet[2848]: I0710 00:22:35.102172 2848 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:35.103666 systemd[1]: Created slice kubepods-burstable-podd381d0aee293bda79b931f4d2b629d35.slice - libcontainer container kubepods-burstable-podd381d0aee293bda79b931f4d2b629d35.slice. Jul 10 00:22:35.105032 kubelet[2848]: I0710 00:22:35.105010 2848 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-20-136" Jul 10 00:22:35.106292 kubelet[2848]: E0710 00:22:35.106253 2848 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.20.136:6443/api/v1/nodes\": dial tcp 172.31.20.136:6443: connect: connection refused" node="ip-172-31-20-136" Jul 10 00:22:35.118173 kubelet[2848]: E0710 00:22:35.118135 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:35.121677 systemd[1]: Created slice kubepods-burstable-podf2d58449835517ae0a9c14b4247086cc.slice - libcontainer container kubepods-burstable-podf2d58449835517ae0a9c14b4247086cc.slice. Jul 10 00:22:35.124155 kubelet[2848]: E0710 00:22:35.124126 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:35.308641 kubelet[2848]: I0710 00:22:35.308384 2848 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-20-136" Jul 10 00:22:35.308874 kubelet[2848]: E0710 00:22:35.308844 2848 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.20.136:6443/api/v1/nodes\": dial tcp 172.31.20.136:6443: connect: connection refused" node="ip-172-31-20-136" Jul 10 00:22:35.403357 containerd[1920]: time="2025-07-10T00:22:35.403303911Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-136,Uid:defe058535f9ca4593940d9cc8ce8a2a,Namespace:kube-system,Attempt:0,}" Jul 10 00:22:35.427019 containerd[1920]: time="2025-07-10T00:22:35.426959382Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-136,Uid:d381d0aee293bda79b931f4d2b629d35,Namespace:kube-system,Attempt:0,}" Jul 10 00:22:35.427305 containerd[1920]: time="2025-07-10T00:22:35.427237986Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-136,Uid:f2d58449835517ae0a9c14b4247086cc,Namespace:kube-system,Attempt:0,}" Jul 10 00:22:35.505290 kubelet[2848]: E0710 00:22:35.505215 2848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.136:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-136?timeout=10s\": dial tcp 172.31.20.136:6443: connect: connection refused" interval="800ms" Jul 10 00:22:35.582714 containerd[1920]: time="2025-07-10T00:22:35.582531854Z" level=info msg="connecting to shim 9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0" address="unix:///run/containerd/s/f606d54e732c51176a2d2fb18bf9c5df71e9e1c753c16246876de19be9e6ba3e" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:22:35.590047 containerd[1920]: time="2025-07-10T00:22:35.589988871Z" level=info msg="connecting to shim bc641d611b62b908bbc754af685026de0a057820a578140febedf6c8c18841a4" address="unix:///run/containerd/s/22053f70dd8bfbfcf3b626a4192aec93fa857b154486e2983ca576a77ea935c6" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:22:35.595450 containerd[1920]: time="2025-07-10T00:22:35.595349015Z" level=info msg="connecting to shim 680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a" address="unix:///run/containerd/s/0dd3687a6cdb49c98a64c39dc9dcf00c182cd6d8d4d31141822e1bd3f9af0e93" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:22:35.713292 kubelet[2848]: I0710 00:22:35.713266 2848 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-20-136" Jul 10 00:22:35.714040 kubelet[2848]: E0710 00:22:35.714003 2848 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.20.136:6443/api/v1/nodes\": dial tcp 172.31.20.136:6443: connect: connection refused" node="ip-172-31-20-136" Jul 10 00:22:35.734397 systemd[1]: Started cri-containerd-680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a.scope - libcontainer container 680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a. Jul 10 00:22:35.736172 systemd[1]: Started cri-containerd-9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0.scope - libcontainer container 9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0. Jul 10 00:22:35.738348 systemd[1]: Started cri-containerd-bc641d611b62b908bbc754af685026de0a057820a578140febedf6c8c18841a4.scope - libcontainer container bc641d611b62b908bbc754af685026de0a057820a578140febedf6c8c18841a4. Jul 10 00:22:35.739604 kubelet[2848]: E0710 00:22:35.739356 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.20.136:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jul 10 00:22:35.836429 containerd[1920]: time="2025-07-10T00:22:35.836221172Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-136,Uid:f2d58449835517ae0a9c14b4247086cc,Namespace:kube-system,Attempt:0,} returns sandbox id \"680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a\"" Jul 10 00:22:35.849730 containerd[1920]: time="2025-07-10T00:22:35.849639303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-136,Uid:d381d0aee293bda79b931f4d2b629d35,Namespace:kube-system,Attempt:0,} returns sandbox id \"bc641d611b62b908bbc754af685026de0a057820a578140febedf6c8c18841a4\"" Jul 10 00:22:35.852918 containerd[1920]: time="2025-07-10T00:22:35.852849378Z" level=info msg="CreateContainer within sandbox \"680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jul 10 00:22:35.860212 containerd[1920]: time="2025-07-10T00:22:35.860142776Z" level=info msg="CreateContainer within sandbox \"bc641d611b62b908bbc754af685026de0a057820a578140febedf6c8c18841a4\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jul 10 00:22:35.875425 containerd[1920]: time="2025-07-10T00:22:35.875371955Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-136,Uid:defe058535f9ca4593940d9cc8ce8a2a,Namespace:kube-system,Attempt:0,} returns sandbox id \"9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0\"" Jul 10 00:22:35.882554 containerd[1920]: time="2025-07-10T00:22:35.882265648Z" level=info msg="CreateContainer within sandbox \"9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jul 10 00:22:35.882916 containerd[1920]: time="2025-07-10T00:22:35.882895016Z" level=info msg="Container 79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:22:35.883598 containerd[1920]: time="2025-07-10T00:22:35.883572927Z" level=info msg="Container 818b34a784d95ce028c2294a92dc2bc9bd3a777958927ffd07e166d860b8f1c2: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:22:35.895779 containerd[1920]: time="2025-07-10T00:22:35.895733716Z" level=info msg="CreateContainer within sandbox \"680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c\"" Jul 10 00:22:35.896591 containerd[1920]: time="2025-07-10T00:22:35.896555390Z" level=info msg="StartContainer for \"79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c\"" Jul 10 00:22:35.899359 containerd[1920]: time="2025-07-10T00:22:35.899274561Z" level=info msg="connecting to shim 79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c" address="unix:///run/containerd/s/0dd3687a6cdb49c98a64c39dc9dcf00c182cd6d8d4d31141822e1bd3f9af0e93" protocol=ttrpc version=3 Jul 10 00:22:35.904661 kubelet[2848]: E0710 00:22:35.903924 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.20.136:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-136&limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jul 10 00:22:35.907451 containerd[1920]: time="2025-07-10T00:22:35.907401056Z" level=info msg="CreateContainer within sandbox \"bc641d611b62b908bbc754af685026de0a057820a578140febedf6c8c18841a4\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"818b34a784d95ce028c2294a92dc2bc9bd3a777958927ffd07e166d860b8f1c2\"" Jul 10 00:22:35.908280 containerd[1920]: time="2025-07-10T00:22:35.908253145Z" level=info msg="StartContainer for \"818b34a784d95ce028c2294a92dc2bc9bd3a777958927ffd07e166d860b8f1c2\"" Jul 10 00:22:35.912972 containerd[1920]: time="2025-07-10T00:22:35.912909695Z" level=info msg="connecting to shim 818b34a784d95ce028c2294a92dc2bc9bd3a777958927ffd07e166d860b8f1c2" address="unix:///run/containerd/s/22053f70dd8bfbfcf3b626a4192aec93fa857b154486e2983ca576a77ea935c6" protocol=ttrpc version=3 Jul 10 00:22:35.914360 containerd[1920]: time="2025-07-10T00:22:35.914320178Z" level=info msg="Container b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:22:35.936025 systemd[1]: Started cri-containerd-79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c.scope - libcontainer container 79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c. Jul 10 00:22:35.941586 containerd[1920]: time="2025-07-10T00:22:35.941520757Z" level=info msg="CreateContainer within sandbox \"9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb\"" Jul 10 00:22:35.943246 containerd[1920]: time="2025-07-10T00:22:35.943012084Z" level=info msg="StartContainer for \"b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb\"" Jul 10 00:22:35.948730 containerd[1920]: time="2025-07-10T00:22:35.948687120Z" level=info msg="connecting to shim b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb" address="unix:///run/containerd/s/f606d54e732c51176a2d2fb18bf9c5df71e9e1c753c16246876de19be9e6ba3e" protocol=ttrpc version=3 Jul 10 00:22:35.970704 systemd[1]: Started cri-containerd-818b34a784d95ce028c2294a92dc2bc9bd3a777958927ffd07e166d860b8f1c2.scope - libcontainer container 818b34a784d95ce028c2294a92dc2bc9bd3a777958927ffd07e166d860b8f1c2. Jul 10 00:22:35.992335 systemd[1]: Started cri-containerd-b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb.scope - libcontainer container b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb. Jul 10 00:22:36.063462 containerd[1920]: time="2025-07-10T00:22:36.063411115Z" level=info msg="StartContainer for \"79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c\" returns successfully" Jul 10 00:22:36.095051 containerd[1920]: time="2025-07-10T00:22:36.094870068Z" level=info msg="StartContainer for \"818b34a784d95ce028c2294a92dc2bc9bd3a777958927ffd07e166d860b8f1c2\" returns successfully" Jul 10 00:22:36.109860 containerd[1920]: time="2025-07-10T00:22:36.109720739Z" level=info msg="StartContainer for \"b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb\" returns successfully" Jul 10 00:22:36.239211 kubelet[2848]: E0710 00:22:36.239170 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.20.136:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jul 10 00:22:36.306455 kubelet[2848]: E0710 00:22:36.306406 2848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.136:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-136?timeout=10s\": dial tcp 172.31.20.136:6443: connect: connection refused" interval="1.6s" Jul 10 00:22:36.421349 kubelet[2848]: E0710 00:22:36.421233 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.20.136:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jul 10 00:22:36.518329 kubelet[2848]: I0710 00:22:36.518300 2848 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-20-136" Jul 10 00:22:36.518717 kubelet[2848]: E0710 00:22:36.518688 2848 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.20.136:6443/api/v1/nodes\": dial tcp 172.31.20.136:6443: connect: connection refused" node="ip-172-31-20-136" Jul 10 00:22:36.848737 kubelet[2848]: E0710 00:22:36.848694 2848 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.31.20.136:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jul 10 00:22:36.998466 kubelet[2848]: E0710 00:22:36.998432 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:37.002522 kubelet[2848]: E0710 00:22:37.002490 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:37.007256 kubelet[2848]: E0710 00:22:37.007098 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:37.907821 kubelet[2848]: E0710 00:22:37.907770 2848 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.136:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-136?timeout=10s\": dial tcp 172.31.20.136:6443: connect: connection refused" interval="3.2s" Jul 10 00:22:38.010147 kubelet[2848]: E0710 00:22:38.009493 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:38.010147 kubelet[2848]: E0710 00:22:38.009901 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:38.010147 kubelet[2848]: E0710 00:22:38.009969 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:38.120685 kubelet[2848]: I0710 00:22:38.120660 2848 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-20-136" Jul 10 00:22:38.120979 kubelet[2848]: E0710 00:22:38.120949 2848 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.20.136:6443/api/v1/nodes\": dial tcp 172.31.20.136:6443: connect: connection refused" node="ip-172-31-20-136" Jul 10 00:22:38.233372 kubelet[2848]: E0710 00:22:38.233261 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.20.136:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jul 10 00:22:38.462248 kubelet[2848]: E0710 00:22:38.462202 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.20.136:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-136&limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jul 10 00:22:38.519564 kubelet[2848]: E0710 00:22:38.519441 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.20.136:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jul 10 00:22:38.805605 kubelet[2848]: E0710 00:22:38.805542 2848 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.20.136:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.20.136:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jul 10 00:22:39.010863 kubelet[2848]: E0710 00:22:39.010700 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:39.010863 kubelet[2848]: E0710 00:22:39.010825 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:40.016699 kubelet[2848]: E0710 00:22:40.016317 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:40.017764 kubelet[2848]: E0710 00:22:40.017602 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:41.323303 kubelet[2848]: I0710 00:22:41.323095 2848 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-20-136" Jul 10 00:22:41.494509 kubelet[2848]: E0710 00:22:41.494315 2848 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:41.551982 kubelet[2848]: I0710 00:22:41.551946 2848 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-20-136" Jul 10 00:22:41.551982 kubelet[2848]: E0710 00:22:41.551991 2848 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ip-172-31-20-136\": node \"ip-172-31-20-136\" not found" Jul 10 00:22:41.586189 kubelet[2848]: E0710 00:22:41.586054 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:41.686831 kubelet[2848]: E0710 00:22:41.686779 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:41.787671 kubelet[2848]: E0710 00:22:41.787634 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:41.888527 kubelet[2848]: E0710 00:22:41.888411 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:41.989335 kubelet[2848]: E0710 00:22:41.989295 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.016282 kubelet[2848]: E0710 00:22:42.016204 2848 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-136\" not found" node="ip-172-31-20-136" Jul 10 00:22:42.090497 kubelet[2848]: E0710 00:22:42.090419 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.190816 kubelet[2848]: E0710 00:22:42.190694 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.291315 kubelet[2848]: E0710 00:22:42.291236 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.392051 kubelet[2848]: E0710 00:22:42.392010 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.493133 kubelet[2848]: E0710 00:22:42.492740 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.594082 kubelet[2848]: E0710 00:22:42.593900 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.695062 kubelet[2848]: E0710 00:22:42.695018 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.802461 kubelet[2848]: E0710 00:22:42.802313 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:42.902542 kubelet[2848]: E0710 00:22:42.902455 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.004803 kubelet[2848]: E0710 00:22:43.004728 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.105942 kubelet[2848]: E0710 00:22:43.105646 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.207018 kubelet[2848]: E0710 00:22:43.206964 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.308248 kubelet[2848]: E0710 00:22:43.308199 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.408854 kubelet[2848]: E0710 00:22:43.408730 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.509527 kubelet[2848]: E0710 00:22:43.509483 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.610674 kubelet[2848]: E0710 00:22:43.610630 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.709838 systemd[1]: Reload requested from client PID 3129 ('systemctl') (unit session-9.scope)... Jul 10 00:22:43.709857 systemd[1]: Reloading... Jul 10 00:22:43.711632 kubelet[2848]: E0710 00:22:43.711582 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.812215 kubelet[2848]: E0710 00:22:43.812176 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.835137 zram_generator::config[3175]: No configuration found. Jul 10 00:22:43.912576 kubelet[2848]: E0710 00:22:43.912511 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:43.937872 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:22:44.013726 kubelet[2848]: E0710 00:22:44.013545 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:44.092610 systemd[1]: Reloading finished in 381 ms. Jul 10 00:22:44.097338 update_engine[1896]: I20250710 00:22:44.097225 1896 update_attempter.cc:509] Updating boot flags... Jul 10 00:22:44.115154 kubelet[2848]: E0710 00:22:44.113728 2848 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-20-136\" not found" Jul 10 00:22:44.126465 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:22:44.153527 systemd[1]: kubelet.service: Deactivated successfully. Jul 10 00:22:44.154092 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:44.154720 systemd[1]: kubelet.service: Consumed 1.512s CPU time, 129.4M memory peak. Jul 10 00:22:44.161833 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:22:44.567485 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:22:44.596746 (kubelet)[3385]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 10 00:22:44.748128 kubelet[3385]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:22:44.755569 kubelet[3385]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jul 10 00:22:44.755569 kubelet[3385]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:22:44.755569 kubelet[3385]: I0710 00:22:44.752551 3385 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 10 00:22:44.787054 kubelet[3385]: I0710 00:22:44.787022 3385 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jul 10 00:22:44.787259 kubelet[3385]: I0710 00:22:44.787244 3385 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 10 00:22:44.787892 kubelet[3385]: I0710 00:22:44.787871 3385 server.go:956] "Client rotation is on, will bootstrap in background" Jul 10 00:22:44.794154 kubelet[3385]: I0710 00:22:44.793337 3385 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Jul 10 00:22:44.816021 sudo[3431]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Jul 10 00:22:44.816374 sudo[3431]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Jul 10 00:22:44.827010 kubelet[3385]: I0710 00:22:44.825985 3385 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 10 00:22:44.842136 kubelet[3385]: I0710 00:22:44.841341 3385 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jul 10 00:22:44.847052 kubelet[3385]: I0710 00:22:44.847024 3385 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 10 00:22:44.852783 kubelet[3385]: I0710 00:22:44.852721 3385 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 10 00:22:44.857199 kubelet[3385]: I0710 00:22:44.856795 3385 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-20-136","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 10 00:22:44.859158 kubelet[3385]: I0710 00:22:44.859077 3385 topology_manager.go:138] "Creating topology manager with none policy" Jul 10 00:22:44.859158 kubelet[3385]: I0710 00:22:44.859138 3385 container_manager_linux.go:303] "Creating device plugin manager" Jul 10 00:22:44.861115 kubelet[3385]: I0710 00:22:44.861051 3385 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:22:44.867872 kubelet[3385]: I0710 00:22:44.867811 3385 kubelet.go:480] "Attempting to sync node with API server" Jul 10 00:22:44.867872 kubelet[3385]: I0710 00:22:44.867846 3385 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 10 00:22:44.869128 kubelet[3385]: I0710 00:22:44.868056 3385 kubelet.go:386] "Adding apiserver pod source" Jul 10 00:22:44.869128 kubelet[3385]: I0710 00:22:44.868077 3385 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 10 00:22:44.882132 kubelet[3385]: I0710 00:22:44.880424 3385 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Jul 10 00:22:44.887726 kubelet[3385]: I0710 00:22:44.887687 3385 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jul 10 00:22:44.909991 kubelet[3385]: I0710 00:22:44.909961 3385 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jul 10 00:22:44.911315 kubelet[3385]: I0710 00:22:44.910205 3385 server.go:1289] "Started kubelet" Jul 10 00:22:44.915259 kubelet[3385]: I0710 00:22:44.915223 3385 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jul 10 00:22:44.919407 kubelet[3385]: I0710 00:22:44.919333 3385 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 10 00:22:44.919971 kubelet[3385]: I0710 00:22:44.919920 3385 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 10 00:22:44.920122 kubelet[3385]: I0710 00:22:44.914872 3385 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 10 00:22:44.923671 kubelet[3385]: I0710 00:22:44.923640 3385 server.go:317] "Adding debug handlers to kubelet server" Jul 10 00:22:44.926361 kubelet[3385]: I0710 00:22:44.915027 3385 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 10 00:22:44.927582 kubelet[3385]: I0710 00:22:44.926409 3385 volume_manager.go:297] "Starting Kubelet Volume Manager" Jul 10 00:22:44.930397 kubelet[3385]: I0710 00:22:44.926423 3385 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jul 10 00:22:44.932373 kubelet[3385]: I0710 00:22:44.932241 3385 reconciler.go:26] "Reconciler: start to sync state" Jul 10 00:22:44.935484 kubelet[3385]: I0710 00:22:44.934941 3385 factory.go:223] Registration of the systemd container factory successfully Jul 10 00:22:44.935770 kubelet[3385]: I0710 00:22:44.935746 3385 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 10 00:22:44.945711 kubelet[3385]: I0710 00:22:44.945668 3385 factory.go:223] Registration of the containerd container factory successfully Jul 10 00:22:44.948848 kubelet[3385]: E0710 00:22:44.948693 3385 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 10 00:22:44.987587 kubelet[3385]: I0710 00:22:44.987546 3385 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jul 10 00:22:45.002212 kubelet[3385]: I0710 00:22:45.002173 3385 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jul 10 00:22:45.002431 kubelet[3385]: I0710 00:22:45.002224 3385 status_manager.go:230] "Starting to sync pod status with apiserver" Jul 10 00:22:45.002431 kubelet[3385]: I0710 00:22:45.002254 3385 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jul 10 00:22:45.002431 kubelet[3385]: I0710 00:22:45.002264 3385 kubelet.go:2436] "Starting kubelet main sync loop" Jul 10 00:22:45.002431 kubelet[3385]: E0710 00:22:45.002321 3385 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 10 00:22:45.098384 kubelet[3385]: I0710 00:22:45.098074 3385 cpu_manager.go:221] "Starting CPU manager" policy="none" Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.098095 3385 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.100136 3385 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.100305 3385 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.100318 3385 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.100340 3385 policy_none.go:49] "None policy: Start" Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.100356 3385 memory_manager.go:186] "Starting memorymanager" policy="None" Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.100368 3385 state_mem.go:35] "Initializing new in-memory state store" Jul 10 00:22:45.100574 kubelet[3385]: I0710 00:22:45.100488 3385 state_mem.go:75] "Updated machine memory state" Jul 10 00:22:45.103209 kubelet[3385]: E0710 00:22:45.103069 3385 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jul 10 00:22:45.112022 kubelet[3385]: E0710 00:22:45.111163 3385 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jul 10 00:22:45.112022 kubelet[3385]: I0710 00:22:45.111372 3385 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 10 00:22:45.112022 kubelet[3385]: I0710 00:22:45.111384 3385 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 10 00:22:45.112022 kubelet[3385]: I0710 00:22:45.111794 3385 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 10 00:22:45.129193 kubelet[3385]: E0710 00:22:45.128360 3385 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jul 10 00:22:45.232965 kubelet[3385]: I0710 00:22:45.232917 3385 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-20-136" Jul 10 00:22:45.253208 kubelet[3385]: I0710 00:22:45.252976 3385 kubelet_node_status.go:124] "Node was previously registered" node="ip-172-31-20-136" Jul 10 00:22:45.254490 kubelet[3385]: I0710 00:22:45.253451 3385 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-20-136" Jul 10 00:22:45.306631 kubelet[3385]: I0710 00:22:45.306597 3385 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:45.308371 kubelet[3385]: I0710 00:22:45.307612 3385 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-20-136" Jul 10 00:22:45.311140 kubelet[3385]: I0710 00:22:45.310995 3385 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:45.340305 kubelet[3385]: I0710 00:22:45.340180 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d381d0aee293bda79b931f4d2b629d35-ca-certs\") pod \"kube-apiserver-ip-172-31-20-136\" (UID: \"d381d0aee293bda79b931f4d2b629d35\") " pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:45.340605 kubelet[3385]: I0710 00:22:45.340488 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d381d0aee293bda79b931f4d2b629d35-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-136\" (UID: \"d381d0aee293bda79b931f4d2b629d35\") " pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:45.340605 kubelet[3385]: I0710 00:22:45.340557 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:45.341217 kubelet[3385]: I0710 00:22:45.340589 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:45.341498 kubelet[3385]: I0710 00:22:45.341184 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:45.341498 kubelet[3385]: I0710 00:22:45.341323 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:45.342007 kubelet[3385]: I0710 00:22:45.341349 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/defe058535f9ca4593940d9cc8ce8a2a-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-136\" (UID: \"defe058535f9ca4593940d9cc8ce8a2a\") " pod="kube-system/kube-scheduler-ip-172-31-20-136" Jul 10 00:22:45.342007 kubelet[3385]: I0710 00:22:45.341676 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d381d0aee293bda79b931f4d2b629d35-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-136\" (UID: \"d381d0aee293bda79b931f4d2b629d35\") " pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:45.344330 kubelet[3385]: I0710 00:22:45.342160 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f2d58449835517ae0a9c14b4247086cc-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-136\" (UID: \"f2d58449835517ae0a9c14b4247086cc\") " pod="kube-system/kube-controller-manager-ip-172-31-20-136" Jul 10 00:22:45.636167 sudo[3431]: pam_unix(sudo:session): session closed for user root Jul 10 00:22:45.879399 kubelet[3385]: I0710 00:22:45.879301 3385 apiserver.go:52] "Watching apiserver" Jul 10 00:22:45.931296 kubelet[3385]: I0710 00:22:45.930758 3385 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jul 10 00:22:46.066018 kubelet[3385]: I0710 00:22:46.064748 3385 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:46.066018 kubelet[3385]: I0710 00:22:46.065227 3385 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-20-136" Jul 10 00:22:46.082827 kubelet[3385]: E0710 00:22:46.082790 3385 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-20-136\" already exists" pod="kube-system/kube-apiserver-ip-172-31-20-136" Jul 10 00:22:46.085339 kubelet[3385]: E0710 00:22:46.085303 3385 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-20-136\" already exists" pod="kube-system/kube-scheduler-ip-172-31-20-136" Jul 10 00:22:46.144877 kubelet[3385]: I0710 00:22:46.144755 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-20-136" podStartSLOduration=1.144631409 podStartE2EDuration="1.144631409s" podCreationTimestamp="2025-07-10 00:22:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:22:46.144045317 +0000 UTC m=+1.527341327" watchObservedRunningTime="2025-07-10 00:22:46.144631409 +0000 UTC m=+1.527927409" Jul 10 00:22:46.145863 kubelet[3385]: I0710 00:22:46.145532 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-20-136" podStartSLOduration=1.145515212 podStartE2EDuration="1.145515212s" podCreationTimestamp="2025-07-10 00:22:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:22:46.125626779 +0000 UTC m=+1.508922785" watchObservedRunningTime="2025-07-10 00:22:46.145515212 +0000 UTC m=+1.528811219" Jul 10 00:22:47.260597 kubelet[3385]: I0710 00:22:47.260442 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-20-136" podStartSLOduration=2.260424745 podStartE2EDuration="2.260424745s" podCreationTimestamp="2025-07-10 00:22:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:22:46.161900555 +0000 UTC m=+1.545196565" watchObservedRunningTime="2025-07-10 00:22:47.260424745 +0000 UTC m=+2.643720751" Jul 10 00:22:47.547026 sudo[2274]: pam_unix(sudo:session): session closed for user root Jul 10 00:22:47.569504 sshd[2273]: Connection closed by 139.178.89.65 port 48448 Jul 10 00:22:47.570827 sshd-session[2271]: pam_unix(sshd:session): session closed for user core Jul 10 00:22:47.576195 systemd[1]: sshd@8-172.31.20.136:22-139.178.89.65:48448.service: Deactivated successfully. Jul 10 00:22:47.579929 systemd[1]: session-9.scope: Deactivated successfully. Jul 10 00:22:47.580340 systemd[1]: session-9.scope: Consumed 6.027s CPU time, 208.2M memory peak. Jul 10 00:22:47.582280 systemd-logind[1894]: Session 9 logged out. Waiting for processes to exit. Jul 10 00:22:47.584214 systemd-logind[1894]: Removed session 9. Jul 10 00:22:49.766086 kubelet[3385]: I0710 00:22:49.765976 3385 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jul 10 00:22:49.767155 containerd[1920]: time="2025-07-10T00:22:49.766768218Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jul 10 00:22:49.767445 kubelet[3385]: I0710 00:22:49.766951 3385 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jul 10 00:22:50.857030 systemd[1]: Created slice kubepods-besteffort-pod2f5522cd_ac7d_496d_b34e_38e0183480a2.slice - libcontainer container kubepods-besteffort-pod2f5522cd_ac7d_496d_b34e_38e0183480a2.slice. Jul 10 00:22:50.877122 kubelet[3385]: I0710 00:22:50.877077 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2f5522cd-ac7d-496d-b34e-38e0183480a2-lib-modules\") pod \"kube-proxy-gxvcc\" (UID: \"2f5522cd-ac7d-496d-b34e-38e0183480a2\") " pod="kube-system/kube-proxy-gxvcc" Jul 10 00:22:50.881620 kubelet[3385]: I0710 00:22:50.877783 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/2f5522cd-ac7d-496d-b34e-38e0183480a2-kube-proxy\") pod \"kube-proxy-gxvcc\" (UID: \"2f5522cd-ac7d-496d-b34e-38e0183480a2\") " pod="kube-system/kube-proxy-gxvcc" Jul 10 00:22:50.883826 kubelet[3385]: I0710 00:22:50.883795 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2f5522cd-ac7d-496d-b34e-38e0183480a2-xtables-lock\") pod \"kube-proxy-gxvcc\" (UID: \"2f5522cd-ac7d-496d-b34e-38e0183480a2\") " pod="kube-system/kube-proxy-gxvcc" Jul 10 00:22:50.884051 kubelet[3385]: I0710 00:22:50.884032 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhknq\" (UniqueName: \"kubernetes.io/projected/2f5522cd-ac7d-496d-b34e-38e0183480a2-kube-api-access-bhknq\") pod \"kube-proxy-gxvcc\" (UID: \"2f5522cd-ac7d-496d-b34e-38e0183480a2\") " pod="kube-system/kube-proxy-gxvcc" Jul 10 00:22:50.893485 systemd[1]: Created slice kubepods-burstable-pod307418df_217d_4289_85e5_04a0e4401fa8.slice - libcontainer container kubepods-burstable-pod307418df_217d_4289_85e5_04a0e4401fa8.slice. Jul 10 00:22:50.964444 systemd[1]: Created slice kubepods-besteffort-pod2d957c09_00b3_4d7e_9065_2aa6cd27ab93.slice - libcontainer container kubepods-besteffort-pod2d957c09_00b3_4d7e_9065_2aa6cd27ab93.slice. Jul 10 00:22:50.987430 kubelet[3385]: I0710 00:22:50.986381 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngmrp\" (UniqueName: \"kubernetes.io/projected/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-kube-api-access-ngmrp\") pod \"cilium-operator-6c4d7847fc-f8d7d\" (UID: \"2d957c09-00b3-4d7e-9065-2aa6cd27ab93\") " pod="kube-system/cilium-operator-6c4d7847fc-f8d7d" Jul 10 00:22:50.987430 kubelet[3385]: I0710 00:22:50.986436 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-bpf-maps\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987430 kubelet[3385]: I0710 00:22:50.986487 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-xtables-lock\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987430 kubelet[3385]: I0710 00:22:50.986510 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-kernel\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987430 kubelet[3385]: I0710 00:22:50.986533 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-f8d7d\" (UID: \"2d957c09-00b3-4d7e-9065-2aa6cd27ab93\") " pod="kube-system/cilium-operator-6c4d7847fc-f8d7d" Jul 10 00:22:50.987754 kubelet[3385]: I0710 00:22:50.986574 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-hostproc\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987754 kubelet[3385]: I0710 00:22:50.986599 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/307418df-217d-4289-85e5-04a0e4401fa8-cilium-config-path\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987754 kubelet[3385]: I0710 00:22:50.986620 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-etc-cni-netd\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987754 kubelet[3385]: I0710 00:22:50.986657 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-cgroup\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987754 kubelet[3385]: I0710 00:22:50.986692 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-lib-modules\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987754 kubelet[3385]: I0710 00:22:50.986714 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hxhd\" (UniqueName: \"kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-kube-api-access-7hxhd\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987990 kubelet[3385]: I0710 00:22:50.986735 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cni-path\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987990 kubelet[3385]: I0710 00:22:50.986755 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/307418df-217d-4289-85e5-04a0e4401fa8-clustermesh-secrets\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987990 kubelet[3385]: I0710 00:22:50.986805 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-net\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987990 kubelet[3385]: I0710 00:22:50.986840 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-run\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:50.987990 kubelet[3385]: I0710 00:22:50.986874 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-hubble-tls\") pod \"cilium-4sqxz\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " pod="kube-system/cilium-4sqxz" Jul 10 00:22:51.183991 containerd[1920]: time="2025-07-10T00:22:51.183866303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gxvcc,Uid:2f5522cd-ac7d-496d-b34e-38e0183480a2,Namespace:kube-system,Attempt:0,}" Jul 10 00:22:51.206731 containerd[1920]: time="2025-07-10T00:22:51.206683705Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4sqxz,Uid:307418df-217d-4289-85e5-04a0e4401fa8,Namespace:kube-system,Attempt:0,}" Jul 10 00:22:51.222646 containerd[1920]: time="2025-07-10T00:22:51.222601309Z" level=info msg="connecting to shim b2a21f92a730becee838b00dbb807724ba6805f5ef59bcc82521ccf20a47a896" address="unix:///run/containerd/s/ee53be0b31f54c84a855d123674495bd0c53cb0560c48e1ac074f70f4049a913" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:22:51.248347 systemd[1]: Started cri-containerd-b2a21f92a730becee838b00dbb807724ba6805f5ef59bcc82521ccf20a47a896.scope - libcontainer container b2a21f92a730becee838b00dbb807724ba6805f5ef59bcc82521ccf20a47a896. Jul 10 00:22:51.259734 containerd[1920]: time="2025-07-10T00:22:51.259555845Z" level=info msg="connecting to shim dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae" address="unix:///run/containerd/s/74406c1f37a4245e26e41c8d20a7a2501dfd1701a48822e51c1260fb62aace92" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:22:51.272882 containerd[1920]: time="2025-07-10T00:22:51.272760997Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-f8d7d,Uid:2d957c09-00b3-4d7e-9065-2aa6cd27ab93,Namespace:kube-system,Attempt:0,}" Jul 10 00:22:51.295303 containerd[1920]: time="2025-07-10T00:22:51.295249482Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gxvcc,Uid:2f5522cd-ac7d-496d-b34e-38e0183480a2,Namespace:kube-system,Attempt:0,} returns sandbox id \"b2a21f92a730becee838b00dbb807724ba6805f5ef59bcc82521ccf20a47a896\"" Jul 10 00:22:51.298502 systemd[1]: Started cri-containerd-dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae.scope - libcontainer container dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae. Jul 10 00:22:51.308433 containerd[1920]: time="2025-07-10T00:22:51.308378116Z" level=info msg="CreateContainer within sandbox \"b2a21f92a730becee838b00dbb807724ba6805f5ef59bcc82521ccf20a47a896\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jul 10 00:22:51.326121 containerd[1920]: time="2025-07-10T00:22:51.325946032Z" level=info msg="connecting to shim f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc" address="unix:///run/containerd/s/620af7ace588975c0009c7a958ce7e15a7682626ea5cf262b3e1b5c202f968d5" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:22:51.337854 containerd[1920]: time="2025-07-10T00:22:51.337812927Z" level=info msg="Container 48db14dcb8b2815485abae931ebbf698cbe12b5fd5bec596bfa8eb5930f4c3bf: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:22:51.355864 containerd[1920]: time="2025-07-10T00:22:51.355762342Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4sqxz,Uid:307418df-217d-4289-85e5-04a0e4401fa8,Namespace:kube-system,Attempt:0,} returns sandbox id \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\"" Jul 10 00:22:51.359036 containerd[1920]: time="2025-07-10T00:22:51.358992789Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Jul 10 00:22:51.360056 containerd[1920]: time="2025-07-10T00:22:51.360018769Z" level=info msg="CreateContainer within sandbox \"b2a21f92a730becee838b00dbb807724ba6805f5ef59bcc82521ccf20a47a896\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"48db14dcb8b2815485abae931ebbf698cbe12b5fd5bec596bfa8eb5930f4c3bf\"" Jul 10 00:22:51.360880 containerd[1920]: time="2025-07-10T00:22:51.360635119Z" level=info msg="StartContainer for \"48db14dcb8b2815485abae931ebbf698cbe12b5fd5bec596bfa8eb5930f4c3bf\"" Jul 10 00:22:51.362929 containerd[1920]: time="2025-07-10T00:22:51.362880710Z" level=info msg="connecting to shim 48db14dcb8b2815485abae931ebbf698cbe12b5fd5bec596bfa8eb5930f4c3bf" address="unix:///run/containerd/s/ee53be0b31f54c84a855d123674495bd0c53cb0560c48e1ac074f70f4049a913" protocol=ttrpc version=3 Jul 10 00:22:51.385407 systemd[1]: Started cri-containerd-f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc.scope - libcontainer container f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc. Jul 10 00:22:51.396631 systemd[1]: Started cri-containerd-48db14dcb8b2815485abae931ebbf698cbe12b5fd5bec596bfa8eb5930f4c3bf.scope - libcontainer container 48db14dcb8b2815485abae931ebbf698cbe12b5fd5bec596bfa8eb5930f4c3bf. Jul 10 00:22:51.468193 containerd[1920]: time="2025-07-10T00:22:51.467778799Z" level=info msg="StartContainer for \"48db14dcb8b2815485abae931ebbf698cbe12b5fd5bec596bfa8eb5930f4c3bf\" returns successfully" Jul 10 00:22:51.482652 containerd[1920]: time="2025-07-10T00:22:51.482554927Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-f8d7d,Uid:2d957c09-00b3-4d7e-9065-2aa6cd27ab93,Namespace:kube-system,Attempt:0,} returns sandbox id \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\"" Jul 10 00:22:52.098940 kubelet[3385]: I0710 00:22:52.098873 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-gxvcc" podStartSLOduration=2.098856137 podStartE2EDuration="2.098856137s" podCreationTimestamp="2025-07-10 00:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:22:52.098584825 +0000 UTC m=+7.481880832" watchObservedRunningTime="2025-07-10 00:22:52.098856137 +0000 UTC m=+7.482152140" Jul 10 00:22:56.015244 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount622920666.mount: Deactivated successfully. Jul 10 00:22:58.594787 containerd[1920]: time="2025-07-10T00:22:58.594604070Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Jul 10 00:22:58.594787 containerd[1920]: time="2025-07-10T00:22:58.594724623Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:58.595700 containerd[1920]: time="2025-07-10T00:22:58.595538067Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 7.236500964s" Jul 10 00:22:58.595700 containerd[1920]: time="2025-07-10T00:22:58.595567502Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Jul 10 00:22:58.596221 containerd[1920]: time="2025-07-10T00:22:58.596199553Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:22:58.597871 containerd[1920]: time="2025-07-10T00:22:58.597844284Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Jul 10 00:22:58.602152 containerd[1920]: time="2025-07-10T00:22:58.602119864Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 10 00:22:58.629243 containerd[1920]: time="2025-07-10T00:22:58.628663483Z" level=info msg="Container e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:22:58.652371 containerd[1920]: time="2025-07-10T00:22:58.652322718Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\"" Jul 10 00:22:58.653280 containerd[1920]: time="2025-07-10T00:22:58.653246424Z" level=info msg="StartContainer for \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\"" Jul 10 00:22:58.654625 containerd[1920]: time="2025-07-10T00:22:58.654590604Z" level=info msg="connecting to shim e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3" address="unix:///run/containerd/s/74406c1f37a4245e26e41c8d20a7a2501dfd1701a48822e51c1260fb62aace92" protocol=ttrpc version=3 Jul 10 00:22:58.724322 systemd[1]: Started cri-containerd-e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3.scope - libcontainer container e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3. Jul 10 00:22:58.783718 containerd[1920]: time="2025-07-10T00:22:58.783665049Z" level=info msg="StartContainer for \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" returns successfully" Jul 10 00:22:58.793820 systemd[1]: cri-containerd-e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3.scope: Deactivated successfully. Jul 10 00:22:58.868778 containerd[1920]: time="2025-07-10T00:22:58.868383784Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" id:\"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" pid:3838 exited_at:{seconds:1752106978 nanos:797261215}" Jul 10 00:22:58.873725 containerd[1920]: time="2025-07-10T00:22:58.873682131Z" level=info msg="received exit event container_id:\"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" id:\"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" pid:3838 exited_at:{seconds:1752106978 nanos:797261215}" Jul 10 00:22:58.900026 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3-rootfs.mount: Deactivated successfully. Jul 10 00:22:59.117404 containerd[1920]: time="2025-07-10T00:22:59.117350549Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 10 00:22:59.129342 containerd[1920]: time="2025-07-10T00:22:59.129238137Z" level=info msg="Container a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:22:59.152369 containerd[1920]: time="2025-07-10T00:22:59.152327551Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\"" Jul 10 00:22:59.153404 containerd[1920]: time="2025-07-10T00:22:59.153208922Z" level=info msg="StartContainer for \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\"" Jul 10 00:22:59.156118 containerd[1920]: time="2025-07-10T00:22:59.155097510Z" level=info msg="connecting to shim a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224" address="unix:///run/containerd/s/74406c1f37a4245e26e41c8d20a7a2501dfd1701a48822e51c1260fb62aace92" protocol=ttrpc version=3 Jul 10 00:22:59.212004 systemd[1]: Started cri-containerd-a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224.scope - libcontainer container a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224. Jul 10 00:22:59.249749 containerd[1920]: time="2025-07-10T00:22:59.249570029Z" level=info msg="StartContainer for \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" returns successfully" Jul 10 00:22:59.263827 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 10 00:22:59.264205 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:22:59.265453 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Jul 10 00:22:59.268429 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 10 00:22:59.270480 systemd[1]: cri-containerd-a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224.scope: Deactivated successfully. Jul 10 00:22:59.272128 containerd[1920]: time="2025-07-10T00:22:59.271999075Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" id:\"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" pid:3881 exited_at:{seconds:1752106979 nanos:270624440}" Jul 10 00:22:59.272128 containerd[1920]: time="2025-07-10T00:22:59.272058464Z" level=info msg="received exit event container_id:\"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" id:\"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" pid:3881 exited_at:{seconds:1752106979 nanos:270624440}" Jul 10 00:22:59.314857 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:22:59.896080 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1748334388.mount: Deactivated successfully. Jul 10 00:23:00.127276 containerd[1920]: time="2025-07-10T00:23:00.127223411Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 10 00:23:00.213203 containerd[1920]: time="2025-07-10T00:23:00.211258183Z" level=info msg="Container 2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:23:00.251693 containerd[1920]: time="2025-07-10T00:23:00.251647937Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\"" Jul 10 00:23:00.258395 containerd[1920]: time="2025-07-10T00:23:00.257879150Z" level=info msg="StartContainer for \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\"" Jul 10 00:23:00.262326 containerd[1920]: time="2025-07-10T00:23:00.261922099Z" level=info msg="connecting to shim 2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd" address="unix:///run/containerd/s/74406c1f37a4245e26e41c8d20a7a2501dfd1701a48822e51c1260fb62aace92" protocol=ttrpc version=3 Jul 10 00:23:00.304595 systemd[1]: Started cri-containerd-2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd.scope - libcontainer container 2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd. Jul 10 00:23:00.386994 containerd[1920]: time="2025-07-10T00:23:00.386867224Z" level=info msg="StartContainer for \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" returns successfully" Jul 10 00:23:00.398824 systemd[1]: cri-containerd-2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd.scope: Deactivated successfully. Jul 10 00:23:00.399176 systemd[1]: cri-containerd-2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd.scope: Consumed 33ms CPU time, 4.2M memory peak, 1M read from disk. Jul 10 00:23:00.404780 containerd[1920]: time="2025-07-10T00:23:00.404731522Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" id:\"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" pid:3942 exited_at:{seconds:1752106980 nanos:404362487}" Jul 10 00:23:00.405997 containerd[1920]: time="2025-07-10T00:23:00.405915175Z" level=info msg="received exit event container_id:\"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" id:\"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" pid:3942 exited_at:{seconds:1752106980 nanos:404362487}" Jul 10 00:23:00.627284 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2005641757.mount: Deactivated successfully. Jul 10 00:23:00.743473 containerd[1920]: time="2025-07-10T00:23:00.743121779Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:23:00.745617 containerd[1920]: time="2025-07-10T00:23:00.745565757Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Jul 10 00:23:00.747144 containerd[1920]: time="2025-07-10T00:23:00.746502860Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:23:00.747914 containerd[1920]: time="2025-07-10T00:23:00.747768557Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.149892359s" Jul 10 00:23:00.747914 containerd[1920]: time="2025-07-10T00:23:00.747813331Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Jul 10 00:23:00.754770 containerd[1920]: time="2025-07-10T00:23:00.754733203Z" level=info msg="CreateContainer within sandbox \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Jul 10 00:23:00.773609 containerd[1920]: time="2025-07-10T00:23:00.771650149Z" level=info msg="Container 59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:23:00.778258 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2143907829.mount: Deactivated successfully. Jul 10 00:23:00.788081 containerd[1920]: time="2025-07-10T00:23:00.788027803Z" level=info msg="CreateContainer within sandbox \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\"" Jul 10 00:23:00.788837 containerd[1920]: time="2025-07-10T00:23:00.788791725Z" level=info msg="StartContainer for \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\"" Jul 10 00:23:00.790799 containerd[1920]: time="2025-07-10T00:23:00.790752212Z" level=info msg="connecting to shim 59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161" address="unix:///run/containerd/s/620af7ace588975c0009c7a958ce7e15a7682626ea5cf262b3e1b5c202f968d5" protocol=ttrpc version=3 Jul 10 00:23:00.820363 systemd[1]: Started cri-containerd-59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161.scope - libcontainer container 59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161. Jul 10 00:23:00.858447 containerd[1920]: time="2025-07-10T00:23:00.858411603Z" level=info msg="StartContainer for \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" returns successfully" Jul 10 00:23:01.143844 containerd[1920]: time="2025-07-10T00:23:01.143804123Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 10 00:23:01.169122 containerd[1920]: time="2025-07-10T00:23:01.168306844Z" level=info msg="Container 3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:23:01.188894 containerd[1920]: time="2025-07-10T00:23:01.188844919Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\"" Jul 10 00:23:01.193118 containerd[1920]: time="2025-07-10T00:23:01.193051254Z" level=info msg="StartContainer for \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\"" Jul 10 00:23:01.195612 containerd[1920]: time="2025-07-10T00:23:01.194493701Z" level=info msg="connecting to shim 3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b" address="unix:///run/containerd/s/74406c1f37a4245e26e41c8d20a7a2501dfd1701a48822e51c1260fb62aace92" protocol=ttrpc version=3 Jul 10 00:23:01.257731 kubelet[3385]: I0710 00:23:01.257636 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-f8d7d" podStartSLOduration=1.992923621 podStartE2EDuration="11.257615206s" podCreationTimestamp="2025-07-10 00:22:50 +0000 UTC" firstStartedPulling="2025-07-10 00:22:51.484194793 +0000 UTC m=+6.867490789" lastFinishedPulling="2025-07-10 00:23:00.748886392 +0000 UTC m=+16.132182374" observedRunningTime="2025-07-10 00:23:01.236530403 +0000 UTC m=+16.619826429" watchObservedRunningTime="2025-07-10 00:23:01.257615206 +0000 UTC m=+16.640911211" Jul 10 00:23:01.262247 systemd[1]: Started cri-containerd-3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b.scope - libcontainer container 3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b. Jul 10 00:23:01.347943 systemd[1]: cri-containerd-3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b.scope: Deactivated successfully. Jul 10 00:23:01.351718 containerd[1920]: time="2025-07-10T00:23:01.351639421Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" id:\"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" pid:4020 exited_at:{seconds:1752106981 nanos:347879035}" Jul 10 00:23:01.357138 containerd[1920]: time="2025-07-10T00:23:01.356442845Z" level=info msg="received exit event container_id:\"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" id:\"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" pid:4020 exited_at:{seconds:1752106981 nanos:347879035}" Jul 10 00:23:01.384572 containerd[1920]: time="2025-07-10T00:23:01.384533112Z" level=info msg="StartContainer for \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" returns successfully" Jul 10 00:23:02.159606 containerd[1920]: time="2025-07-10T00:23:02.159559317Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 10 00:23:02.206966 containerd[1920]: time="2025-07-10T00:23:02.199761264Z" level=info msg="Container 4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:23:02.258306 containerd[1920]: time="2025-07-10T00:23:02.258256357Z" level=info msg="CreateContainer within sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\"" Jul 10 00:23:02.260272 containerd[1920]: time="2025-07-10T00:23:02.260051037Z" level=info msg="StartContainer for \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\"" Jul 10 00:23:02.263474 containerd[1920]: time="2025-07-10T00:23:02.263367437Z" level=info msg="connecting to shim 4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2" address="unix:///run/containerd/s/74406c1f37a4245e26e41c8d20a7a2501dfd1701a48822e51c1260fb62aace92" protocol=ttrpc version=3 Jul 10 00:23:02.319130 systemd[1]: Started cri-containerd-4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2.scope - libcontainer container 4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2. Jul 10 00:23:02.404558 containerd[1920]: time="2025-07-10T00:23:02.404509147Z" level=info msg="StartContainer for \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" returns successfully" Jul 10 00:23:02.583764 kubelet[3385]: I0710 00:23:02.583714 3385 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Jul 10 00:23:02.588361 containerd[1920]: time="2025-07-10T00:23:02.587415041Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" id:\"0a4a9cfaf77e3ec985afbc3ac020342b89846ec83090c0f913ffc93775125873\" pid:4086 exited_at:{seconds:1752106982 nanos:585364725}" Jul 10 00:23:02.681839 systemd[1]: Created slice kubepods-burstable-podaa6372db_0667_4c8d_bb2d_de410e983504.slice - libcontainer container kubepods-burstable-podaa6372db_0667_4c8d_bb2d_de410e983504.slice. Jul 10 00:23:02.695465 systemd[1]: Created slice kubepods-burstable-pod9d3ee75b_9a02_41ba_a4ef_81f7dda86154.slice - libcontainer container kubepods-burstable-pod9d3ee75b_9a02_41ba_a4ef_81f7dda86154.slice. Jul 10 00:23:02.784715 kubelet[3385]: I0710 00:23:02.784662 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9d3ee75b-9a02-41ba-a4ef-81f7dda86154-config-volume\") pod \"coredns-674b8bbfcf-59p2z\" (UID: \"9d3ee75b-9a02-41ba-a4ef-81f7dda86154\") " pod="kube-system/coredns-674b8bbfcf-59p2z" Jul 10 00:23:02.784715 kubelet[3385]: I0710 00:23:02.784704 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5k96\" (UniqueName: \"kubernetes.io/projected/9d3ee75b-9a02-41ba-a4ef-81f7dda86154-kube-api-access-s5k96\") pod \"coredns-674b8bbfcf-59p2z\" (UID: \"9d3ee75b-9a02-41ba-a4ef-81f7dda86154\") " pod="kube-system/coredns-674b8bbfcf-59p2z" Jul 10 00:23:02.784715 kubelet[3385]: I0710 00:23:02.784723 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa6372db-0667-4c8d-bb2d-de410e983504-config-volume\") pod \"coredns-674b8bbfcf-dqpm9\" (UID: \"aa6372db-0667-4c8d-bb2d-de410e983504\") " pod="kube-system/coredns-674b8bbfcf-dqpm9" Jul 10 00:23:02.784715 kubelet[3385]: I0710 00:23:02.784742 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w8pc\" (UniqueName: \"kubernetes.io/projected/aa6372db-0667-4c8d-bb2d-de410e983504-kube-api-access-6w8pc\") pod \"coredns-674b8bbfcf-dqpm9\" (UID: \"aa6372db-0667-4c8d-bb2d-de410e983504\") " pod="kube-system/coredns-674b8bbfcf-dqpm9" Jul 10 00:23:02.997721 containerd[1920]: time="2025-07-10T00:23:02.996988268Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dqpm9,Uid:aa6372db-0667-4c8d-bb2d-de410e983504,Namespace:kube-system,Attempt:0,}" Jul 10 00:23:03.003555 containerd[1920]: time="2025-07-10T00:23:03.003205174Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-59p2z,Uid:9d3ee75b-9a02-41ba-a4ef-81f7dda86154,Namespace:kube-system,Attempt:0,}" Jul 10 00:23:03.193671 kubelet[3385]: I0710 00:23:03.193594 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-4sqxz" podStartSLOduration=5.955188015 podStartE2EDuration="13.193572474s" podCreationTimestamp="2025-07-10 00:22:50 +0000 UTC" firstStartedPulling="2025-07-10 00:22:51.358014212 +0000 UTC m=+6.741310217" lastFinishedPulling="2025-07-10 00:22:58.596398679 +0000 UTC m=+13.979694676" observedRunningTime="2025-07-10 00:23:03.191765253 +0000 UTC m=+18.575061264" watchObservedRunningTime="2025-07-10 00:23:03.193572474 +0000 UTC m=+18.576868480" Jul 10 00:23:05.256681 systemd-networkd[1768]: cilium_host: Link UP Jul 10 00:23:05.257335 systemd-networkd[1768]: cilium_net: Link UP Jul 10 00:23:05.257504 systemd-networkd[1768]: cilium_net: Gained carrier Jul 10 00:23:05.257646 systemd-networkd[1768]: cilium_host: Gained carrier Jul 10 00:23:05.258753 (udev-worker)[4145]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:23:05.259079 (udev-worker)[4180]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:23:05.396384 systemd-networkd[1768]: cilium_host: Gained IPv6LL Jul 10 00:23:05.398737 (udev-worker)[4190]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:23:05.407835 systemd-networkd[1768]: cilium_vxlan: Link UP Jul 10 00:23:05.407846 systemd-networkd[1768]: cilium_vxlan: Gained carrier Jul 10 00:23:05.972260 systemd-networkd[1768]: cilium_net: Gained IPv6LL Jul 10 00:23:06.378212 kernel: NET: Registered PF_ALG protocol family Jul 10 00:23:06.996292 systemd-networkd[1768]: cilium_vxlan: Gained IPv6LL Jul 10 00:23:07.139045 systemd-networkd[1768]: lxc_health: Link UP Jul 10 00:23:07.153615 systemd-networkd[1768]: lxc_health: Gained carrier Jul 10 00:23:07.596145 kernel: eth0: renamed from tmp079e6 Jul 10 00:23:07.605530 kernel: eth0: renamed from tmp43ed5 Jul 10 00:23:07.611719 systemd-networkd[1768]: lxca97b95c711da: Link UP Jul 10 00:23:07.612050 systemd-networkd[1768]: lxc002e7c933dc9: Link UP Jul 10 00:23:07.617528 systemd-networkd[1768]: lxca97b95c711da: Gained carrier Jul 10 00:23:07.620909 systemd-networkd[1768]: lxc002e7c933dc9: Gained carrier Jul 10 00:23:08.596360 systemd-networkd[1768]: lxc_health: Gained IPv6LL Jul 10 00:23:09.110309 systemd-networkd[1768]: lxc002e7c933dc9: Gained IPv6LL Jul 10 00:23:09.620340 systemd-networkd[1768]: lxca97b95c711da: Gained IPv6LL Jul 10 00:23:11.779756 ntpd[1885]: Listen normally on 8 cilium_host 192.168.0.72:123 Jul 10 00:23:11.781950 ntpd[1885]: 10 Jul 00:23:11 ntpd[1885]: Listen normally on 8 cilium_host 192.168.0.72:123 Jul 10 00:23:11.781950 ntpd[1885]: 10 Jul 00:23:11 ntpd[1885]: Listen normally on 9 cilium_net [fe80::c07e:ecff:fe50:d33b%4]:123 Jul 10 00:23:11.781950 ntpd[1885]: 10 Jul 00:23:11 ntpd[1885]: Listen normally on 10 cilium_host [fe80::94a9:bff:fe9c:146e%5]:123 Jul 10 00:23:11.781950 ntpd[1885]: 10 Jul 00:23:11 ntpd[1885]: Listen normally on 11 cilium_vxlan [fe80::2847:4aff:fec6:a725%6]:123 Jul 10 00:23:11.781950 ntpd[1885]: 10 Jul 00:23:11 ntpd[1885]: Listen normally on 12 lxc_health [fe80::7049:c9ff:fe12:86bb%8]:123 Jul 10 00:23:11.781950 ntpd[1885]: 10 Jul 00:23:11 ntpd[1885]: Listen normally on 13 lxc002e7c933dc9 [fe80::ec33:81ff:fe40:8f6f%10]:123 Jul 10 00:23:11.781950 ntpd[1885]: 10 Jul 00:23:11 ntpd[1885]: Listen normally on 14 lxca97b95c711da [fe80::ec7b:1aff:fe52:7b3f%12]:123 Jul 10 00:23:11.779856 ntpd[1885]: Listen normally on 9 cilium_net [fe80::c07e:ecff:fe50:d33b%4]:123 Jul 10 00:23:11.779914 ntpd[1885]: Listen normally on 10 cilium_host [fe80::94a9:bff:fe9c:146e%5]:123 Jul 10 00:23:11.779960 ntpd[1885]: Listen normally on 11 cilium_vxlan [fe80::2847:4aff:fec6:a725%6]:123 Jul 10 00:23:11.779997 ntpd[1885]: Listen normally on 12 lxc_health [fe80::7049:c9ff:fe12:86bb%8]:123 Jul 10 00:23:11.780036 ntpd[1885]: Listen normally on 13 lxc002e7c933dc9 [fe80::ec33:81ff:fe40:8f6f%10]:123 Jul 10 00:23:11.780075 ntpd[1885]: Listen normally on 14 lxca97b95c711da [fe80::ec7b:1aff:fe52:7b3f%12]:123 Jul 10 00:23:12.198221 containerd[1920]: time="2025-07-10T00:23:12.197723092Z" level=info msg="connecting to shim 079e63bc577b038fe382f79d518f01954a58d623653d49fc0b65b5d688ce8c74" address="unix:///run/containerd/s/7e2d091f610ee46288636dd6a093396b4dbb4a9b2c50c8a9dd7f72a87c246fcd" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:23:12.204880 containerd[1920]: time="2025-07-10T00:23:12.204280510Z" level=info msg="connecting to shim 43ed532f9cab2a439633e06434026d5c78e41080de0805f69a9ee500c973a9e0" address="unix:///run/containerd/s/3ff0370ead4f38f96bc49c86a3ac0c4f9707d87a6743caccbb8bac7209bfb6d2" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:23:12.305523 systemd[1]: Started cri-containerd-079e63bc577b038fe382f79d518f01954a58d623653d49fc0b65b5d688ce8c74.scope - libcontainer container 079e63bc577b038fe382f79d518f01954a58d623653d49fc0b65b5d688ce8c74. Jul 10 00:23:12.320489 systemd[1]: Started cri-containerd-43ed532f9cab2a439633e06434026d5c78e41080de0805f69a9ee500c973a9e0.scope - libcontainer container 43ed532f9cab2a439633e06434026d5c78e41080de0805f69a9ee500c973a9e0. Jul 10 00:23:12.425127 containerd[1920]: time="2025-07-10T00:23:12.425006669Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-dqpm9,Uid:aa6372db-0667-4c8d-bb2d-de410e983504,Namespace:kube-system,Attempt:0,} returns sandbox id \"079e63bc577b038fe382f79d518f01954a58d623653d49fc0b65b5d688ce8c74\"" Jul 10 00:23:12.428542 containerd[1920]: time="2025-07-10T00:23:12.427683090Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-59p2z,Uid:9d3ee75b-9a02-41ba-a4ef-81f7dda86154,Namespace:kube-system,Attempt:0,} returns sandbox id \"43ed532f9cab2a439633e06434026d5c78e41080de0805f69a9ee500c973a9e0\"" Jul 10 00:23:12.460405 containerd[1920]: time="2025-07-10T00:23:12.460316622Z" level=info msg="CreateContainer within sandbox \"43ed532f9cab2a439633e06434026d5c78e41080de0805f69a9ee500c973a9e0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 10 00:23:12.464584 containerd[1920]: time="2025-07-10T00:23:12.464528941Z" level=info msg="CreateContainer within sandbox \"079e63bc577b038fe382f79d518f01954a58d623653d49fc0b65b5d688ce8c74\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 10 00:23:12.489210 containerd[1920]: time="2025-07-10T00:23:12.489165510Z" level=info msg="Container 747c6be753c6d2db38620fd42ac8848bc9f96daf9d060ade7d0be18181672910: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:23:12.490056 containerd[1920]: time="2025-07-10T00:23:12.490026426Z" level=info msg="Container 3c4e16bac72749c73a2fbee4dd0162f18114616adef3014361fad5055d0d7db2: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:23:12.504055 containerd[1920]: time="2025-07-10T00:23:12.503998895Z" level=info msg="CreateContainer within sandbox \"43ed532f9cab2a439633e06434026d5c78e41080de0805f69a9ee500c973a9e0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"747c6be753c6d2db38620fd42ac8848bc9f96daf9d060ade7d0be18181672910\"" Jul 10 00:23:12.507140 containerd[1920]: time="2025-07-10T00:23:12.507009985Z" level=info msg="StartContainer for \"747c6be753c6d2db38620fd42ac8848bc9f96daf9d060ade7d0be18181672910\"" Jul 10 00:23:12.507393 containerd[1920]: time="2025-07-10T00:23:12.507199392Z" level=info msg="CreateContainer within sandbox \"079e63bc577b038fe382f79d518f01954a58d623653d49fc0b65b5d688ce8c74\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3c4e16bac72749c73a2fbee4dd0162f18114616adef3014361fad5055d0d7db2\"" Jul 10 00:23:12.508310 containerd[1920]: time="2025-07-10T00:23:12.508280248Z" level=info msg="StartContainer for \"3c4e16bac72749c73a2fbee4dd0162f18114616adef3014361fad5055d0d7db2\"" Jul 10 00:23:12.509666 containerd[1920]: time="2025-07-10T00:23:12.509562675Z" level=info msg="connecting to shim 3c4e16bac72749c73a2fbee4dd0162f18114616adef3014361fad5055d0d7db2" address="unix:///run/containerd/s/7e2d091f610ee46288636dd6a093396b4dbb4a9b2c50c8a9dd7f72a87c246fcd" protocol=ttrpc version=3 Jul 10 00:23:12.509666 containerd[1920]: time="2025-07-10T00:23:12.509616573Z" level=info msg="connecting to shim 747c6be753c6d2db38620fd42ac8848bc9f96daf9d060ade7d0be18181672910" address="unix:///run/containerd/s/3ff0370ead4f38f96bc49c86a3ac0c4f9707d87a6743caccbb8bac7209bfb6d2" protocol=ttrpc version=3 Jul 10 00:23:12.540347 systemd[1]: Started cri-containerd-3c4e16bac72749c73a2fbee4dd0162f18114616adef3014361fad5055d0d7db2.scope - libcontainer container 3c4e16bac72749c73a2fbee4dd0162f18114616adef3014361fad5055d0d7db2. Jul 10 00:23:12.541694 systemd[1]: Started cri-containerd-747c6be753c6d2db38620fd42ac8848bc9f96daf9d060ade7d0be18181672910.scope - libcontainer container 747c6be753c6d2db38620fd42ac8848bc9f96daf9d060ade7d0be18181672910. Jul 10 00:23:12.610405 containerd[1920]: time="2025-07-10T00:23:12.610364211Z" level=info msg="StartContainer for \"3c4e16bac72749c73a2fbee4dd0162f18114616adef3014361fad5055d0d7db2\" returns successfully" Jul 10 00:23:12.615588 containerd[1920]: time="2025-07-10T00:23:12.615549899Z" level=info msg="StartContainer for \"747c6be753c6d2db38620fd42ac8848bc9f96daf9d060ade7d0be18181672910\" returns successfully" Jul 10 00:23:13.134377 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1377514316.mount: Deactivated successfully. Jul 10 00:23:13.245906 kubelet[3385]: I0710 00:23:13.243712 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-59p2z" podStartSLOduration=23.24366333 podStartE2EDuration="23.24366333s" podCreationTimestamp="2025-07-10 00:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:23:13.226260401 +0000 UTC m=+28.609556404" watchObservedRunningTime="2025-07-10 00:23:13.24366333 +0000 UTC m=+28.626959335" Jul 10 00:23:32.816909 systemd[1]: Started sshd@9-172.31.20.136:22-139.178.89.65:59048.service - OpenSSH per-connection server daemon (139.178.89.65:59048). Jul 10 00:23:33.032577 sshd[4718]: Accepted publickey for core from 139.178.89.65 port 59048 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:23:33.034865 sshd-session[4718]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:23:33.047682 systemd-logind[1894]: New session 10 of user core. Jul 10 00:23:33.059562 systemd[1]: Started session-10.scope - Session 10 of User core. Jul 10 00:23:33.891482 sshd[4720]: Connection closed by 139.178.89.65 port 59048 Jul 10 00:23:33.892050 sshd-session[4718]: pam_unix(sshd:session): session closed for user core Jul 10 00:23:33.896530 systemd[1]: sshd@9-172.31.20.136:22-139.178.89.65:59048.service: Deactivated successfully. Jul 10 00:23:33.898988 systemd[1]: session-10.scope: Deactivated successfully. Jul 10 00:23:33.900327 systemd-logind[1894]: Session 10 logged out. Waiting for processes to exit. Jul 10 00:23:33.902762 systemd-logind[1894]: Removed session 10. Jul 10 00:23:38.928614 systemd[1]: Started sshd@10-172.31.20.136:22-139.178.89.65:59062.service - OpenSSH per-connection server daemon (139.178.89.65:59062). Jul 10 00:23:39.113951 sshd[4735]: Accepted publickey for core from 139.178.89.65 port 59062 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:23:39.115482 sshd-session[4735]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:23:39.121268 systemd-logind[1894]: New session 11 of user core. Jul 10 00:23:39.130433 systemd[1]: Started session-11.scope - Session 11 of User core. Jul 10 00:23:39.348128 sshd[4737]: Connection closed by 139.178.89.65 port 59062 Jul 10 00:23:39.348842 sshd-session[4735]: pam_unix(sshd:session): session closed for user core Jul 10 00:23:39.353712 systemd[1]: sshd@10-172.31.20.136:22-139.178.89.65:59062.service: Deactivated successfully. Jul 10 00:23:39.356067 systemd[1]: session-11.scope: Deactivated successfully. Jul 10 00:23:39.357471 systemd-logind[1894]: Session 11 logged out. Waiting for processes to exit. Jul 10 00:23:39.359164 systemd-logind[1894]: Removed session 11. Jul 10 00:23:44.387750 systemd[1]: Started sshd@11-172.31.20.136:22-139.178.89.65:49176.service - OpenSSH per-connection server daemon (139.178.89.65:49176). Jul 10 00:23:44.572165 sshd[4751]: Accepted publickey for core from 139.178.89.65 port 49176 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:23:44.574253 sshd-session[4751]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:23:44.581937 systemd-logind[1894]: New session 12 of user core. Jul 10 00:23:44.589391 systemd[1]: Started session-12.scope - Session 12 of User core. Jul 10 00:23:44.779747 sshd[4753]: Connection closed by 139.178.89.65 port 49176 Jul 10 00:23:44.780553 sshd-session[4751]: pam_unix(sshd:session): session closed for user core Jul 10 00:23:44.785358 systemd[1]: sshd@11-172.31.20.136:22-139.178.89.65:49176.service: Deactivated successfully. Jul 10 00:23:44.788015 systemd[1]: session-12.scope: Deactivated successfully. Jul 10 00:23:44.789987 systemd-logind[1894]: Session 12 logged out. Waiting for processes to exit. Jul 10 00:23:44.791756 systemd-logind[1894]: Removed session 12. Jul 10 00:23:49.811366 systemd[1]: Started sshd@12-172.31.20.136:22-139.178.89.65:46572.service - OpenSSH per-connection server daemon (139.178.89.65:46572). Jul 10 00:23:49.978898 sshd[4768]: Accepted publickey for core from 139.178.89.65 port 46572 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:23:49.980354 sshd-session[4768]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:23:49.986165 systemd-logind[1894]: New session 13 of user core. Jul 10 00:23:49.991334 systemd[1]: Started session-13.scope - Session 13 of User core. Jul 10 00:23:50.208234 sshd[4770]: Connection closed by 139.178.89.65 port 46572 Jul 10 00:23:50.208250 sshd-session[4768]: pam_unix(sshd:session): session closed for user core Jul 10 00:23:50.213125 systemd[1]: sshd@12-172.31.20.136:22-139.178.89.65:46572.service: Deactivated successfully. Jul 10 00:23:50.216114 systemd[1]: session-13.scope: Deactivated successfully. Jul 10 00:23:50.217377 systemd-logind[1894]: Session 13 logged out. Waiting for processes to exit. Jul 10 00:23:50.219327 systemd-logind[1894]: Removed session 13. Jul 10 00:23:50.244793 systemd[1]: Started sshd@13-172.31.20.136:22-139.178.89.65:46584.service - OpenSSH per-connection server daemon (139.178.89.65:46584). Jul 10 00:23:50.440072 sshd[4782]: Accepted publickey for core from 139.178.89.65 port 46584 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:23:50.441819 sshd-session[4782]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:23:50.448905 systemd-logind[1894]: New session 14 of user core. Jul 10 00:23:50.455335 systemd[1]: Started session-14.scope - Session 14 of User core. Jul 10 00:23:50.693140 sshd[4784]: Connection closed by 139.178.89.65 port 46584 Jul 10 00:23:50.695005 sshd-session[4782]: pam_unix(sshd:session): session closed for user core Jul 10 00:23:50.704386 systemd[1]: sshd@13-172.31.20.136:22-139.178.89.65:46584.service: Deactivated successfully. Jul 10 00:23:50.704832 systemd-logind[1894]: Session 14 logged out. Waiting for processes to exit. Jul 10 00:23:50.710380 systemd[1]: session-14.scope: Deactivated successfully. Jul 10 00:23:50.714176 systemd-logind[1894]: Removed session 14. Jul 10 00:23:50.728139 systemd[1]: Started sshd@14-172.31.20.136:22-139.178.89.65:46590.service - OpenSSH per-connection server daemon (139.178.89.65:46590). Jul 10 00:23:50.918240 sshd[4794]: Accepted publickey for core from 139.178.89.65 port 46590 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:23:50.919658 sshd-session[4794]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:23:50.925916 systemd-logind[1894]: New session 15 of user core. Jul 10 00:23:50.936349 systemd[1]: Started session-15.scope - Session 15 of User core. Jul 10 00:23:51.192219 sshd[4796]: Connection closed by 139.178.89.65 port 46590 Jul 10 00:23:51.193517 sshd-session[4794]: pam_unix(sshd:session): session closed for user core Jul 10 00:23:51.199200 systemd[1]: sshd@14-172.31.20.136:22-139.178.89.65:46590.service: Deactivated successfully. Jul 10 00:23:51.201924 systemd[1]: session-15.scope: Deactivated successfully. Jul 10 00:23:51.204091 systemd-logind[1894]: Session 15 logged out. Waiting for processes to exit. Jul 10 00:23:51.206563 systemd-logind[1894]: Removed session 15. Jul 10 00:23:56.233562 systemd[1]: Started sshd@15-172.31.20.136:22-139.178.89.65:46594.service - OpenSSH per-connection server daemon (139.178.89.65:46594). Jul 10 00:23:56.411777 sshd[4811]: Accepted publickey for core from 139.178.89.65 port 46594 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:23:56.413488 sshd-session[4811]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:23:56.419189 systemd-logind[1894]: New session 16 of user core. Jul 10 00:23:56.427361 systemd[1]: Started session-16.scope - Session 16 of User core. Jul 10 00:23:56.616341 sshd[4813]: Connection closed by 139.178.89.65 port 46594 Jul 10 00:23:56.616724 sshd-session[4811]: pam_unix(sshd:session): session closed for user core Jul 10 00:23:56.621657 systemd[1]: sshd@15-172.31.20.136:22-139.178.89.65:46594.service: Deactivated successfully. Jul 10 00:23:56.623964 systemd[1]: session-16.scope: Deactivated successfully. Jul 10 00:23:56.625472 systemd-logind[1894]: Session 16 logged out. Waiting for processes to exit. Jul 10 00:23:56.627624 systemd-logind[1894]: Removed session 16. Jul 10 00:24:01.673547 systemd[1]: Started sshd@16-172.31.20.136:22-139.178.89.65:37394.service - OpenSSH per-connection server daemon (139.178.89.65:37394). Jul 10 00:24:01.911154 sshd[4824]: Accepted publickey for core from 139.178.89.65 port 37394 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:01.913283 sshd-session[4824]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:01.939909 systemd-logind[1894]: New session 17 of user core. Jul 10 00:24:01.950419 systemd[1]: Started session-17.scope - Session 17 of User core. Jul 10 00:24:02.415256 sshd[4826]: Connection closed by 139.178.89.65 port 37394 Jul 10 00:24:02.417460 sshd-session[4824]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:02.423237 systemd-logind[1894]: Session 17 logged out. Waiting for processes to exit. Jul 10 00:24:02.424395 systemd[1]: sshd@16-172.31.20.136:22-139.178.89.65:37394.service: Deactivated successfully. Jul 10 00:24:02.428255 systemd[1]: session-17.scope: Deactivated successfully. Jul 10 00:24:02.429829 systemd-logind[1894]: Removed session 17. Jul 10 00:24:02.469676 systemd[1]: Started sshd@17-172.31.20.136:22-139.178.89.65:37404.service - OpenSSH per-connection server daemon (139.178.89.65:37404). Jul 10 00:24:02.729751 sshd[4839]: Accepted publickey for core from 139.178.89.65 port 37404 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:02.731494 sshd-session[4839]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:02.740208 systemd-logind[1894]: New session 18 of user core. Jul 10 00:24:02.747496 systemd[1]: Started session-18.scope - Session 18 of User core. Jul 10 00:24:03.621301 sshd[4841]: Connection closed by 139.178.89.65 port 37404 Jul 10 00:24:03.622914 sshd-session[4839]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:03.631927 systemd[1]: sshd@17-172.31.20.136:22-139.178.89.65:37404.service: Deactivated successfully. Jul 10 00:24:03.634879 systemd[1]: session-18.scope: Deactivated successfully. Jul 10 00:24:03.637040 systemd-logind[1894]: Session 18 logged out. Waiting for processes to exit. Jul 10 00:24:03.640977 systemd-logind[1894]: Removed session 18. Jul 10 00:24:03.659546 systemd[1]: Started sshd@18-172.31.20.136:22-139.178.89.65:37416.service - OpenSSH per-connection server daemon (139.178.89.65:37416). Jul 10 00:24:03.859894 sshd[4851]: Accepted publickey for core from 139.178.89.65 port 37416 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:03.861327 sshd-session[4851]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:03.869121 systemd-logind[1894]: New session 19 of user core. Jul 10 00:24:03.873788 systemd[1]: Started session-19.scope - Session 19 of User core. Jul 10 00:24:04.988485 sshd[4853]: Connection closed by 139.178.89.65 port 37416 Jul 10 00:24:04.989762 sshd-session[4851]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:04.994961 systemd-logind[1894]: Session 19 logged out. Waiting for processes to exit. Jul 10 00:24:04.997858 systemd[1]: sshd@18-172.31.20.136:22-139.178.89.65:37416.service: Deactivated successfully. Jul 10 00:24:05.001526 systemd[1]: session-19.scope: Deactivated successfully. Jul 10 00:24:05.009390 systemd-logind[1894]: Removed session 19. Jul 10 00:24:05.024711 systemd[1]: Started sshd@19-172.31.20.136:22-139.178.89.65:37428.service - OpenSSH per-connection server daemon (139.178.89.65:37428). Jul 10 00:24:05.207703 sshd[4871]: Accepted publickey for core from 139.178.89.65 port 37428 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:05.209699 sshd-session[4871]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:05.215439 systemd-logind[1894]: New session 20 of user core. Jul 10 00:24:05.226360 systemd[1]: Started session-20.scope - Session 20 of User core. Jul 10 00:24:05.589670 sshd[4873]: Connection closed by 139.178.89.65 port 37428 Jul 10 00:24:05.591299 sshd-session[4871]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:05.595113 systemd[1]: sshd@19-172.31.20.136:22-139.178.89.65:37428.service: Deactivated successfully. Jul 10 00:24:05.600993 systemd[1]: session-20.scope: Deactivated successfully. Jul 10 00:24:05.604665 systemd-logind[1894]: Session 20 logged out. Waiting for processes to exit. Jul 10 00:24:05.607967 systemd-logind[1894]: Removed session 20. Jul 10 00:24:05.622360 systemd[1]: Started sshd@20-172.31.20.136:22-139.178.89.65:37442.service - OpenSSH per-connection server daemon (139.178.89.65:37442). Jul 10 00:24:05.799510 sshd[4883]: Accepted publickey for core from 139.178.89.65 port 37442 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:05.801258 sshd-session[4883]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:05.807081 systemd-logind[1894]: New session 21 of user core. Jul 10 00:24:05.815365 systemd[1]: Started session-21.scope - Session 21 of User core. Jul 10 00:24:06.002329 sshd[4885]: Connection closed by 139.178.89.65 port 37442 Jul 10 00:24:06.003238 sshd-session[4883]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:06.008210 systemd-logind[1894]: Session 21 logged out. Waiting for processes to exit. Jul 10 00:24:06.008432 systemd[1]: sshd@20-172.31.20.136:22-139.178.89.65:37442.service: Deactivated successfully. Jul 10 00:24:06.011901 systemd[1]: session-21.scope: Deactivated successfully. Jul 10 00:24:06.014275 systemd-logind[1894]: Removed session 21. Jul 10 00:24:11.040386 systemd[1]: Started sshd@21-172.31.20.136:22-139.178.89.65:43356.service - OpenSSH per-connection server daemon (139.178.89.65:43356). Jul 10 00:24:11.215768 sshd[4899]: Accepted publickey for core from 139.178.89.65 port 43356 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:11.217910 sshd-session[4899]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:11.225838 systemd-logind[1894]: New session 22 of user core. Jul 10 00:24:11.236368 systemd[1]: Started session-22.scope - Session 22 of User core. Jul 10 00:24:11.421909 sshd[4901]: Connection closed by 139.178.89.65 port 43356 Jul 10 00:24:11.422466 sshd-session[4899]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:11.426878 systemd-logind[1894]: Session 22 logged out. Waiting for processes to exit. Jul 10 00:24:11.427415 systemd[1]: sshd@21-172.31.20.136:22-139.178.89.65:43356.service: Deactivated successfully. Jul 10 00:24:11.429866 systemd[1]: session-22.scope: Deactivated successfully. Jul 10 00:24:11.432223 systemd-logind[1894]: Removed session 22. Jul 10 00:24:16.460890 systemd[1]: Started sshd@22-172.31.20.136:22-139.178.89.65:43358.service - OpenSSH per-connection server daemon (139.178.89.65:43358). Jul 10 00:24:16.650669 sshd[4914]: Accepted publickey for core from 139.178.89.65 port 43358 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:16.651695 sshd-session[4914]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:16.657989 systemd-logind[1894]: New session 23 of user core. Jul 10 00:24:16.663399 systemd[1]: Started session-23.scope - Session 23 of User core. Jul 10 00:24:16.874433 sshd[4916]: Connection closed by 139.178.89.65 port 43358 Jul 10 00:24:16.876201 sshd-session[4914]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:16.880511 systemd-logind[1894]: Session 23 logged out. Waiting for processes to exit. Jul 10 00:24:16.881633 systemd[1]: sshd@22-172.31.20.136:22-139.178.89.65:43358.service: Deactivated successfully. Jul 10 00:24:16.884479 systemd[1]: session-23.scope: Deactivated successfully. Jul 10 00:24:16.886420 systemd-logind[1894]: Removed session 23. Jul 10 00:24:16.909336 systemd[1]: Started sshd@23-172.31.20.136:22-139.178.89.65:43364.service - OpenSSH per-connection server daemon (139.178.89.65:43364). Jul 10 00:24:17.084772 sshd[4928]: Accepted publickey for core from 139.178.89.65 port 43364 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:17.086569 sshd-session[4928]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:17.091912 systemd-logind[1894]: New session 24 of user core. Jul 10 00:24:17.097312 systemd[1]: Started session-24.scope - Session 24 of User core. Jul 10 00:24:18.502848 kubelet[3385]: I0710 00:24:18.502561 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-dqpm9" podStartSLOduration=88.502546275 podStartE2EDuration="1m28.502546275s" podCreationTimestamp="2025-07-10 00:22:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:23:13.266323784 +0000 UTC m=+28.649619796" watchObservedRunningTime="2025-07-10 00:24:18.502546275 +0000 UTC m=+93.885842312" Jul 10 00:24:18.523865 containerd[1920]: time="2025-07-10T00:24:18.523776067Z" level=info msg="StopContainer for \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" with timeout 30 (s)" Jul 10 00:24:18.536889 containerd[1920]: time="2025-07-10T00:24:18.536847194Z" level=info msg="Stop container \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" with signal terminated" Jul 10 00:24:18.560754 systemd[1]: cri-containerd-59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161.scope: Deactivated successfully. Jul 10 00:24:18.564692 containerd[1920]: time="2025-07-10T00:24:18.564587343Z" level=info msg="received exit event container_id:\"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" id:\"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" pid:3987 exited_at:{seconds:1752107058 nanos:564167391}" Jul 10 00:24:18.566485 containerd[1920]: time="2025-07-10T00:24:18.566419796Z" level=info msg="TaskExit event in podsandbox handler container_id:\"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" id:\"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" pid:3987 exited_at:{seconds:1752107058 nanos:564167391}" Jul 10 00:24:18.577924 containerd[1920]: time="2025-07-10T00:24:18.577766645Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 10 00:24:18.590950 containerd[1920]: time="2025-07-10T00:24:18.586740162Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" id:\"1fe500b2ff5289a4ed9c5a54bcde17a6f0f28bdcfea86ac05ebb7eddeb0143bf\" pid:4954 exited_at:{seconds:1752107058 nanos:585648153}" Jul 10 00:24:18.595556 containerd[1920]: time="2025-07-10T00:24:18.595523060Z" level=info msg="StopContainer for \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" with timeout 2 (s)" Jul 10 00:24:18.596097 containerd[1920]: time="2025-07-10T00:24:18.596051957Z" level=info msg="Stop container \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" with signal terminated" Jul 10 00:24:18.596889 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161-rootfs.mount: Deactivated successfully. Jul 10 00:24:18.607863 systemd-networkd[1768]: lxc_health: Link DOWN Jul 10 00:24:18.607877 systemd-networkd[1768]: lxc_health: Lost carrier Jul 10 00:24:18.626252 containerd[1920]: time="2025-07-10T00:24:18.626069647Z" level=info msg="StopContainer for \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" returns successfully" Jul 10 00:24:18.627159 containerd[1920]: time="2025-07-10T00:24:18.627121259Z" level=info msg="StopPodSandbox for \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\"" Jul 10 00:24:18.627326 containerd[1920]: time="2025-07-10T00:24:18.627203373Z" level=info msg="Container to stop \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:24:18.634626 systemd[1]: cri-containerd-4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2.scope: Deactivated successfully. Jul 10 00:24:18.635005 systemd[1]: cri-containerd-4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2.scope: Consumed 8.206s CPU time, 240M memory peak, 121.1M read from disk, 13.3M written to disk. Jul 10 00:24:18.638683 containerd[1920]: time="2025-07-10T00:24:18.638639856Z" level=info msg="received exit event container_id:\"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" id:\"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" pid:4058 exited_at:{seconds:1752107058 nanos:638116410}" Jul 10 00:24:18.639078 containerd[1920]: time="2025-07-10T00:24:18.638775309Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" id:\"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" pid:4058 exited_at:{seconds:1752107058 nanos:638116410}" Jul 10 00:24:18.641680 systemd[1]: cri-containerd-f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc.scope: Deactivated successfully. Jul 10 00:24:18.645868 containerd[1920]: time="2025-07-10T00:24:18.645829240Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" id:\"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" pid:3633 exit_status:137 exited_at:{seconds:1752107058 nanos:645535166}" Jul 10 00:24:18.689678 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2-rootfs.mount: Deactivated successfully. Jul 10 00:24:18.707535 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc-rootfs.mount: Deactivated successfully. Jul 10 00:24:18.711866 containerd[1920]: time="2025-07-10T00:24:18.711641793Z" level=info msg="shim disconnected" id=f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc namespace=k8s.io Jul 10 00:24:18.711866 containerd[1920]: time="2025-07-10T00:24:18.711678259Z" level=warning msg="cleaning up after shim disconnected" id=f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc namespace=k8s.io Jul 10 00:24:18.711866 containerd[1920]: time="2025-07-10T00:24:18.711689722Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:24:18.717060 containerd[1920]: time="2025-07-10T00:24:18.717013901Z" level=info msg="StopContainer for \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" returns successfully" Jul 10 00:24:18.718718 containerd[1920]: time="2025-07-10T00:24:18.718665572Z" level=info msg="StopPodSandbox for \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\"" Jul 10 00:24:18.719064 containerd[1920]: time="2025-07-10T00:24:18.719040363Z" level=info msg="Container to stop \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:24:18.719834 containerd[1920]: time="2025-07-10T00:24:18.719495238Z" level=info msg="Container to stop \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:24:18.719834 containerd[1920]: time="2025-07-10T00:24:18.719521475Z" level=info msg="Container to stop \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:24:18.719834 containerd[1920]: time="2025-07-10T00:24:18.719536075Z" level=info msg="Container to stop \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:24:18.719834 containerd[1920]: time="2025-07-10T00:24:18.719560292Z" level=info msg="Container to stop \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:24:18.732256 systemd[1]: cri-containerd-dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae.scope: Deactivated successfully. Jul 10 00:24:18.744495 containerd[1920]: time="2025-07-10T00:24:18.744264189Z" level=error msg="Failed to handle event container_id:\"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" id:\"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" pid:3633 exit_status:137 exited_at:{seconds:1752107058 nanos:645535166} for f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc" error="failed to handle container TaskExit event: failed to stop sandbox: failed to delete task: ttrpc: closed" Jul 10 00:24:18.744495 containerd[1920]: time="2025-07-10T00:24:18.744346089Z" level=info msg="TaskExit event in podsandbox handler container_id:\"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" id:\"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" pid:3578 exit_status:137 exited_at:{seconds:1752107058 nanos:735089161}" Jul 10 00:24:18.746478 containerd[1920]: time="2025-07-10T00:24:18.746439364Z" level=info msg="received exit event sandbox_id:\"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" exit_status:137 exited_at:{seconds:1752107058 nanos:645535166}" Jul 10 00:24:18.751425 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc-shm.mount: Deactivated successfully. Jul 10 00:24:18.774843 containerd[1920]: time="2025-07-10T00:24:18.774573365Z" level=info msg="TearDown network for sandbox \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" successfully" Jul 10 00:24:18.774843 containerd[1920]: time="2025-07-10T00:24:18.774616856Z" level=info msg="StopPodSandbox for \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" returns successfully" Jul 10 00:24:18.789781 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae-rootfs.mount: Deactivated successfully. Jul 10 00:24:18.805550 containerd[1920]: time="2025-07-10T00:24:18.805442171Z" level=info msg="received exit event sandbox_id:\"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" exit_status:137 exited_at:{seconds:1752107058 nanos:735089161}" Jul 10 00:24:18.807287 containerd[1920]: time="2025-07-10T00:24:18.807244407Z" level=info msg="shim disconnected" id=dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae namespace=k8s.io Jul 10 00:24:18.807549 containerd[1920]: time="2025-07-10T00:24:18.807409171Z" level=warning msg="cleaning up after shim disconnected" id=dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae namespace=k8s.io Jul 10 00:24:18.807549 containerd[1920]: time="2025-07-10T00:24:18.807421499Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:24:18.808548 containerd[1920]: time="2025-07-10T00:24:18.808414740Z" level=info msg="TearDown network for sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" successfully" Jul 10 00:24:18.808548 containerd[1920]: time="2025-07-10T00:24:18.808439375Z" level=info msg="StopPodSandbox for \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" returns successfully" Jul 10 00:24:18.872955 kubelet[3385]: I0710 00:24:18.872479 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngmrp\" (UniqueName: \"kubernetes.io/projected/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-kube-api-access-ngmrp\") pod \"2d957c09-00b3-4d7e-9065-2aa6cd27ab93\" (UID: \"2d957c09-00b3-4d7e-9065-2aa6cd27ab93\") " Jul 10 00:24:18.872955 kubelet[3385]: I0710 00:24:18.872530 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-cilium-config-path\") pod \"2d957c09-00b3-4d7e-9065-2aa6cd27ab93\" (UID: \"2d957c09-00b3-4d7e-9065-2aa6cd27ab93\") " Jul 10 00:24:18.875121 kubelet[3385]: I0710 00:24:18.874756 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "2d957c09-00b3-4d7e-9065-2aa6cd27ab93" (UID: "2d957c09-00b3-4d7e-9065-2aa6cd27ab93"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jul 10 00:24:18.877143 kubelet[3385]: I0710 00:24:18.877080 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-kube-api-access-ngmrp" (OuterVolumeSpecName: "kube-api-access-ngmrp") pod "2d957c09-00b3-4d7e-9065-2aa6cd27ab93" (UID: "2d957c09-00b3-4d7e-9065-2aa6cd27ab93"). InnerVolumeSpecName "kube-api-access-ngmrp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jul 10 00:24:18.973775 kubelet[3385]: I0710 00:24:18.973737 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-etc-cni-netd\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.973775 kubelet[3385]: I0710 00:24:18.973777 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-run\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.973991 kubelet[3385]: I0710 00:24:18.973802 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/307418df-217d-4289-85e5-04a0e4401fa8-clustermesh-secrets\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.973991 kubelet[3385]: I0710 00:24:18.973816 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-net\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.973991 kubelet[3385]: I0710 00:24:18.973836 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-bpf-maps\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.973991 kubelet[3385]: I0710 00:24:18.973859 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hxhd\" (UniqueName: \"kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-kube-api-access-7hxhd\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.973991 kubelet[3385]: I0710 00:24:18.973873 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cni-path\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.973991 kubelet[3385]: I0710 00:24:18.973888 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-cgroup\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.974605 kubelet[3385]: I0710 00:24:18.973902 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-xtables-lock\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.974605 kubelet[3385]: I0710 00:24:18.973915 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-hostproc\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.974605 kubelet[3385]: I0710 00:24:18.973932 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/307418df-217d-4289-85e5-04a0e4401fa8-cilium-config-path\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.974605 kubelet[3385]: I0710 00:24:18.973970 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-kernel\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.974605 kubelet[3385]: I0710 00:24:18.973984 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-lib-modules\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.974605 kubelet[3385]: I0710 00:24:18.974003 3385 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-hubble-tls\") pod \"307418df-217d-4289-85e5-04a0e4401fa8\" (UID: \"307418df-217d-4289-85e5-04a0e4401fa8\") " Jul 10 00:24:18.974772 kubelet[3385]: I0710 00:24:18.974045 3385 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ngmrp\" (UniqueName: \"kubernetes.io/projected/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-kube-api-access-ngmrp\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:18.974772 kubelet[3385]: I0710 00:24:18.974055 3385 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2d957c09-00b3-4d7e-9065-2aa6cd27ab93-cilium-config-path\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:18.974772 kubelet[3385]: I0710 00:24:18.974175 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cni-path" (OuterVolumeSpecName: "cni-path") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.974772 kubelet[3385]: I0710 00:24:18.974221 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.974772 kubelet[3385]: I0710 00:24:18.974240 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.977584 kubelet[3385]: I0710 00:24:18.977313 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/307418df-217d-4289-85e5-04a0e4401fa8-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jul 10 00:24:18.977584 kubelet[3385]: I0710 00:24:18.977369 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.977584 kubelet[3385]: I0710 00:24:18.977385 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.979091 kubelet[3385]: I0710 00:24:18.979054 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jul 10 00:24:18.979091 kubelet[3385]: I0710 00:24:18.979141 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.979259 kubelet[3385]: I0710 00:24:18.979158 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.979259 kubelet[3385]: I0710 00:24:18.979172 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-hostproc" (OuterVolumeSpecName: "hostproc") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.980805 kubelet[3385]: I0710 00:24:18.980782 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-kube-api-access-7hxhd" (OuterVolumeSpecName: "kube-api-access-7hxhd") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "kube-api-access-7hxhd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jul 10 00:24:18.980951 kubelet[3385]: I0710 00:24:18.980932 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.981043 kubelet[3385]: I0710 00:24:18.981008 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:24:18.982741 kubelet[3385]: I0710 00:24:18.982694 3385 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/307418df-217d-4289-85e5-04a0e4401fa8-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "307418df-217d-4289-85e5-04a0e4401fa8" (UID: "307418df-217d-4289-85e5-04a0e4401fa8"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jul 10 00:24:19.012398 systemd[1]: Removed slice kubepods-besteffort-pod2d957c09_00b3_4d7e_9065_2aa6cd27ab93.slice - libcontainer container kubepods-besteffort-pod2d957c09_00b3_4d7e_9065_2aa6cd27ab93.slice. Jul 10 00:24:19.014567 systemd[1]: Removed slice kubepods-burstable-pod307418df_217d_4289_85e5_04a0e4401fa8.slice - libcontainer container kubepods-burstable-pod307418df_217d_4289_85e5_04a0e4401fa8.slice. Jul 10 00:24:19.014766 systemd[1]: kubepods-burstable-pod307418df_217d_4289_85e5_04a0e4401fa8.slice: Consumed 8.325s CPU time, 240.3M memory peak, 122.2M read from disk, 13.3M written to disk. Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074483 3385 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7hxhd\" (UniqueName: \"kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-kube-api-access-7hxhd\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074522 3385 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cni-path\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074531 3385 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-cgroup\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074541 3385 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-xtables-lock\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074549 3385 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-hostproc\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074557 3385 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/307418df-217d-4289-85e5-04a0e4401fa8-cilium-config-path\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074566 3385 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-kernel\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074607 kubelet[3385]: I0710 00:24:19.074574 3385 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-lib-modules\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074916 kubelet[3385]: I0710 00:24:19.074581 3385 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/307418df-217d-4289-85e5-04a0e4401fa8-hubble-tls\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074916 kubelet[3385]: I0710 00:24:19.074589 3385 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-etc-cni-netd\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074916 kubelet[3385]: I0710 00:24:19.074597 3385 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-cilium-run\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074916 kubelet[3385]: I0710 00:24:19.074604 3385 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/307418df-217d-4289-85e5-04a0e4401fa8-clustermesh-secrets\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074916 kubelet[3385]: I0710 00:24:19.074612 3385 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-host-proc-sys-net\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.074916 kubelet[3385]: I0710 00:24:19.074620 3385 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/307418df-217d-4289-85e5-04a0e4401fa8-bpf-maps\") on node \"ip-172-31-20-136\" DevicePath \"\"" Jul 10 00:24:19.387529 kubelet[3385]: I0710 00:24:19.387426 3385 scope.go:117] "RemoveContainer" containerID="59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161" Jul 10 00:24:19.391791 containerd[1920]: time="2025-07-10T00:24:19.391670230Z" level=info msg="RemoveContainer for \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\"" Jul 10 00:24:19.403251 containerd[1920]: time="2025-07-10T00:24:19.403013883Z" level=info msg="RemoveContainer for \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" returns successfully" Jul 10 00:24:19.403405 kubelet[3385]: I0710 00:24:19.403298 3385 scope.go:117] "RemoveContainer" containerID="59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161" Jul 10 00:24:19.403919 containerd[1920]: time="2025-07-10T00:24:19.403811933Z" level=error msg="ContainerStatus for \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\": not found" Jul 10 00:24:19.404169 kubelet[3385]: E0710 00:24:19.404141 3385 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\": not found" containerID="59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161" Jul 10 00:24:19.405257 kubelet[3385]: I0710 00:24:19.405196 3385 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161"} err="failed to get container status \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\": rpc error: code = NotFound desc = an error occurred when try to find container \"59bc9e9165e95fb5fdd68d4e51526b9c2f720420baa92b467fafbcfe54871161\": not found" Jul 10 00:24:19.412125 kubelet[3385]: I0710 00:24:19.411320 3385 scope.go:117] "RemoveContainer" containerID="4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2" Jul 10 00:24:19.424514 containerd[1920]: time="2025-07-10T00:24:19.424361295Z" level=info msg="RemoveContainer for \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\"" Jul 10 00:24:19.432216 containerd[1920]: time="2025-07-10T00:24:19.432174484Z" level=info msg="RemoveContainer for \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" returns successfully" Jul 10 00:24:19.432477 kubelet[3385]: I0710 00:24:19.432439 3385 scope.go:117] "RemoveContainer" containerID="3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b" Jul 10 00:24:19.434268 containerd[1920]: time="2025-07-10T00:24:19.434214032Z" level=info msg="RemoveContainer for \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\"" Jul 10 00:24:19.441942 containerd[1920]: time="2025-07-10T00:24:19.441895914Z" level=info msg="RemoveContainer for \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" returns successfully" Jul 10 00:24:19.442347 kubelet[3385]: I0710 00:24:19.442225 3385 scope.go:117] "RemoveContainer" containerID="2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd" Jul 10 00:24:19.446285 containerd[1920]: time="2025-07-10T00:24:19.446248044Z" level=info msg="RemoveContainer for \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\"" Jul 10 00:24:19.452877 containerd[1920]: time="2025-07-10T00:24:19.452806862Z" level=info msg="RemoveContainer for \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" returns successfully" Jul 10 00:24:19.453089 kubelet[3385]: I0710 00:24:19.453046 3385 scope.go:117] "RemoveContainer" containerID="a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224" Jul 10 00:24:19.455283 containerd[1920]: time="2025-07-10T00:24:19.455245561Z" level=info msg="RemoveContainer for \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\"" Jul 10 00:24:19.462566 containerd[1920]: time="2025-07-10T00:24:19.462527129Z" level=info msg="RemoveContainer for \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" returns successfully" Jul 10 00:24:19.462867 kubelet[3385]: I0710 00:24:19.462833 3385 scope.go:117] "RemoveContainer" containerID="e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3" Jul 10 00:24:19.464954 containerd[1920]: time="2025-07-10T00:24:19.464919802Z" level=info msg="RemoveContainer for \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\"" Jul 10 00:24:19.475592 containerd[1920]: time="2025-07-10T00:24:19.475545594Z" level=info msg="RemoveContainer for \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" returns successfully" Jul 10 00:24:19.475950 kubelet[3385]: I0710 00:24:19.475865 3385 scope.go:117] "RemoveContainer" containerID="4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2" Jul 10 00:24:19.476363 containerd[1920]: time="2025-07-10T00:24:19.476321658Z" level=error msg="ContainerStatus for \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\": not found" Jul 10 00:24:19.476563 kubelet[3385]: E0710 00:24:19.476485 3385 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\": not found" containerID="4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2" Jul 10 00:24:19.476563 kubelet[3385]: I0710 00:24:19.476517 3385 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2"} err="failed to get container status \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\": rpc error: code = NotFound desc = an error occurred when try to find container \"4d0f72f1cdbcf54d4804aca2284b59b7945fc8cd8c2591d8113c9908629f54d2\": not found" Jul 10 00:24:19.476563 kubelet[3385]: I0710 00:24:19.476546 3385 scope.go:117] "RemoveContainer" containerID="3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b" Jul 10 00:24:19.477195 kubelet[3385]: E0710 00:24:19.477135 3385 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\": not found" containerID="3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b" Jul 10 00:24:19.477195 kubelet[3385]: I0710 00:24:19.477165 3385 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b"} err="failed to get container status \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\": rpc error: code = NotFound desc = an error occurred when try to find container \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\": not found" Jul 10 00:24:19.477195 kubelet[3385]: I0710 00:24:19.477187 3385 scope.go:117] "RemoveContainer" containerID="2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd" Jul 10 00:24:19.477328 containerd[1920]: time="2025-07-10T00:24:19.476984476Z" level=error msg="ContainerStatus for \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3321be55569aa8c5901eafdc1bd97b660954baed535b153b36225c92f61f8c2b\": not found" Jul 10 00:24:19.477465 containerd[1920]: time="2025-07-10T00:24:19.477424892Z" level=error msg="ContainerStatus for \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\": not found" Jul 10 00:24:19.477630 kubelet[3385]: E0710 00:24:19.477564 3385 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\": not found" containerID="2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd" Jul 10 00:24:19.477630 kubelet[3385]: I0710 00:24:19.477588 3385 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd"} err="failed to get container status \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\": rpc error: code = NotFound desc = an error occurred when try to find container \"2ccbe2bcda2762a39098db0d2067648c6b31849b3addcaf569ba990531d9f5dd\": not found" Jul 10 00:24:19.477630 kubelet[3385]: I0710 00:24:19.477609 3385 scope.go:117] "RemoveContainer" containerID="a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224" Jul 10 00:24:19.478152 kubelet[3385]: E0710 00:24:19.477937 3385 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\": not found" containerID="a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224" Jul 10 00:24:19.478152 kubelet[3385]: I0710 00:24:19.477992 3385 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224"} err="failed to get container status \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\": rpc error: code = NotFound desc = an error occurred when try to find container \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\": not found" Jul 10 00:24:19.478152 kubelet[3385]: I0710 00:24:19.478014 3385 scope.go:117] "RemoveContainer" containerID="e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3" Jul 10 00:24:19.478505 containerd[1920]: time="2025-07-10T00:24:19.477808854Z" level=error msg="ContainerStatus for \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a17a059a234cf59a5925911800fd476afea22eb24a224e887f43bef09c21a224\": not found" Jul 10 00:24:19.478505 containerd[1920]: time="2025-07-10T00:24:19.478248931Z" level=error msg="ContainerStatus for \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\": not found" Jul 10 00:24:19.478563 kubelet[3385]: E0710 00:24:19.478379 3385 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\": not found" containerID="e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3" Jul 10 00:24:19.478563 kubelet[3385]: I0710 00:24:19.478399 3385 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3"} err="failed to get container status \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\": rpc error: code = NotFound desc = an error occurred when try to find container \"e5c711bfb9e01b54569946cdcbda3b222b2fc347d66277bc3b46888dbe7bfdf3\": not found" Jul 10 00:24:19.593988 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae-shm.mount: Deactivated successfully. Jul 10 00:24:19.594127 systemd[1]: var-lib-kubelet-pods-307418df\x2d217d\x2d4289\x2d85e5\x2d04a0e4401fa8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d7hxhd.mount: Deactivated successfully. Jul 10 00:24:19.594196 systemd[1]: var-lib-kubelet-pods-2d957c09\x2d00b3\x2d4d7e\x2d9065\x2d2aa6cd27ab93-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dngmrp.mount: Deactivated successfully. Jul 10 00:24:19.594257 systemd[1]: var-lib-kubelet-pods-307418df\x2d217d\x2d4289\x2d85e5\x2d04a0e4401fa8-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jul 10 00:24:19.594316 systemd[1]: var-lib-kubelet-pods-307418df\x2d217d\x2d4289\x2d85e5\x2d04a0e4401fa8-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jul 10 00:24:20.172912 kubelet[3385]: E0710 00:24:20.172866 3385 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 10 00:24:20.483371 sshd[4930]: Connection closed by 139.178.89.65 port 43364 Jul 10 00:24:20.484133 sshd-session[4928]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:20.489424 systemd-logind[1894]: Session 24 logged out. Waiting for processes to exit. Jul 10 00:24:20.490337 systemd[1]: sshd@23-172.31.20.136:22-139.178.89.65:43364.service: Deactivated successfully. Jul 10 00:24:20.492837 systemd[1]: session-24.scope: Deactivated successfully. Jul 10 00:24:20.495271 systemd-logind[1894]: Removed session 24. Jul 10 00:24:20.515952 systemd[1]: Started sshd@24-172.31.20.136:22-139.178.89.65:52206.service - OpenSSH per-connection server daemon (139.178.89.65:52206). Jul 10 00:24:20.713453 sshd[5090]: Accepted publickey for core from 139.178.89.65 port 52206 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:20.714029 sshd-session[5090]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:20.714547 containerd[1920]: time="2025-07-10T00:24:20.714236030Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" id:\"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" pid:3633 exit_status:137 exited_at:{seconds:1752107058 nanos:645535166}" Jul 10 00:24:20.721286 systemd-logind[1894]: New session 25 of user core. Jul 10 00:24:20.725340 systemd[1]: Started session-25.scope - Session 25 of User core. Jul 10 00:24:20.779639 ntpd[1885]: Deleting interface #12 lxc_health, fe80::7049:c9ff:fe12:86bb%8#123, interface stats: received=0, sent=0, dropped=0, active_time=69 secs Jul 10 00:24:20.780258 ntpd[1885]: 10 Jul 00:24:20 ntpd[1885]: Deleting interface #12 lxc_health, fe80::7049:c9ff:fe12:86bb%8#123, interface stats: received=0, sent=0, dropped=0, active_time=69 secs Jul 10 00:24:21.006939 kubelet[3385]: I0710 00:24:21.006901 3385 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d957c09-00b3-4d7e-9065-2aa6cd27ab93" path="/var/lib/kubelet/pods/2d957c09-00b3-4d7e-9065-2aa6cd27ab93/volumes" Jul 10 00:24:21.007829 kubelet[3385]: I0710 00:24:21.007798 3385 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="307418df-217d-4289-85e5-04a0e4401fa8" path="/var/lib/kubelet/pods/307418df-217d-4289-85e5-04a0e4401fa8/volumes" Jul 10 00:24:21.242197 sshd[5092]: Connection closed by 139.178.89.65 port 52206 Jul 10 00:24:21.242312 sshd-session[5090]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:21.249296 systemd-logind[1894]: Session 25 logged out. Waiting for processes to exit. Jul 10 00:24:21.250720 systemd[1]: sshd@24-172.31.20.136:22-139.178.89.65:52206.service: Deactivated successfully. Jul 10 00:24:21.254511 systemd[1]: session-25.scope: Deactivated successfully. Jul 10 00:24:21.259622 systemd-logind[1894]: Removed session 25. Jul 10 00:24:21.274337 systemd[1]: Started sshd@25-172.31.20.136:22-139.178.89.65:52212.service - OpenSSH per-connection server daemon (139.178.89.65:52212). Jul 10 00:24:21.294390 systemd[1]: Created slice kubepods-burstable-podc291a48b_16aa_4ff2_a8cd_c0ae6744a016.slice - libcontainer container kubepods-burstable-podc291a48b_16aa_4ff2_a8cd_c0ae6744a016.slice. Jul 10 00:24:21.393363 kubelet[3385]: I0710 00:24:21.393278 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-clustermesh-secrets\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393363 kubelet[3385]: I0710 00:24:21.393336 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-cilium-ipsec-secrets\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393363 kubelet[3385]: I0710 00:24:21.393368 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-bpf-maps\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393800 kubelet[3385]: I0710 00:24:21.393384 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-hostproc\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393800 kubelet[3385]: I0710 00:24:21.393398 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-etc-cni-netd\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393800 kubelet[3385]: I0710 00:24:21.393415 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-cilium-config-path\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393800 kubelet[3385]: I0710 00:24:21.393435 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p96l\" (UniqueName: \"kubernetes.io/projected/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-kube-api-access-9p96l\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393800 kubelet[3385]: I0710 00:24:21.393452 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-xtables-lock\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393800 kubelet[3385]: I0710 00:24:21.393468 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-cilium-run\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393962 kubelet[3385]: I0710 00:24:21.393485 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-host-proc-sys-net\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393962 kubelet[3385]: I0710 00:24:21.393500 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-host-proc-sys-kernel\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393962 kubelet[3385]: I0710 00:24:21.393519 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-lib-modules\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393962 kubelet[3385]: I0710 00:24:21.393534 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-hubble-tls\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393962 kubelet[3385]: I0710 00:24:21.393552 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-cni-path\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.393962 kubelet[3385]: I0710 00:24:21.393566 3385 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/c291a48b-16aa-4ff2-a8cd-c0ae6744a016-cilium-cgroup\") pod \"cilium-4cjwz\" (UID: \"c291a48b-16aa-4ff2-a8cd-c0ae6744a016\") " pod="kube-system/cilium-4cjwz" Jul 10 00:24:21.467673 sshd[5103]: Accepted publickey for core from 139.178.89.65 port 52212 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:21.469229 sshd-session[5103]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:21.475132 systemd-logind[1894]: New session 26 of user core. Jul 10 00:24:21.479338 systemd[1]: Started session-26.scope - Session 26 of User core. Jul 10 00:24:21.597754 sshd[5105]: Connection closed by 139.178.89.65 port 52212 Jul 10 00:24:21.598669 sshd-session[5103]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:21.600426 containerd[1920]: time="2025-07-10T00:24:21.600314388Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4cjwz,Uid:c291a48b-16aa-4ff2-a8cd-c0ae6744a016,Namespace:kube-system,Attempt:0,}" Jul 10 00:24:21.604492 systemd[1]: sshd@25-172.31.20.136:22-139.178.89.65:52212.service: Deactivated successfully. Jul 10 00:24:21.607615 systemd[1]: session-26.scope: Deactivated successfully. Jul 10 00:24:21.609708 systemd-logind[1894]: Session 26 logged out. Waiting for processes to exit. Jul 10 00:24:21.611706 systemd-logind[1894]: Removed session 26. Jul 10 00:24:21.632134 containerd[1920]: time="2025-07-10T00:24:21.631969508Z" level=info msg="connecting to shim f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc" address="unix:///run/containerd/s/4da342f01d5ae30adf1257f90aa03cd9e6d854b0647b1ba49bdba668f172ce69" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:24:21.644988 systemd[1]: Started sshd@26-172.31.20.136:22-139.178.89.65:52228.service - OpenSSH per-connection server daemon (139.178.89.65:52228). Jul 10 00:24:21.659659 systemd[1]: Started cri-containerd-f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc.scope - libcontainer container f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc. Jul 10 00:24:21.706806 containerd[1920]: time="2025-07-10T00:24:21.706762949Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4cjwz,Uid:c291a48b-16aa-4ff2-a8cd-c0ae6744a016,Namespace:kube-system,Attempt:0,} returns sandbox id \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\"" Jul 10 00:24:21.717222 containerd[1920]: time="2025-07-10T00:24:21.717163107Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 10 00:24:21.729121 containerd[1920]: time="2025-07-10T00:24:21.728160799Z" level=info msg="Container 03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:24:21.741214 containerd[1920]: time="2025-07-10T00:24:21.741166028Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6\"" Jul 10 00:24:21.744020 containerd[1920]: time="2025-07-10T00:24:21.743978069Z" level=info msg="StartContainer for \"03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6\"" Jul 10 00:24:21.745448 containerd[1920]: time="2025-07-10T00:24:21.745391756Z" level=info msg="connecting to shim 03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6" address="unix:///run/containerd/s/4da342f01d5ae30adf1257f90aa03cd9e6d854b0647b1ba49bdba668f172ce69" protocol=ttrpc version=3 Jul 10 00:24:21.769349 systemd[1]: Started cri-containerd-03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6.scope - libcontainer container 03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6. Jul 10 00:24:21.810323 containerd[1920]: time="2025-07-10T00:24:21.810264115Z" level=info msg="StartContainer for \"03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6\" returns successfully" Jul 10 00:24:21.823712 systemd[1]: cri-containerd-03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6.scope: Deactivated successfully. Jul 10 00:24:21.823972 systemd[1]: cri-containerd-03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6.scope: Consumed 22ms CPU time, 9.6M memory peak, 3.2M read from disk. Jul 10 00:24:21.826140 containerd[1920]: time="2025-07-10T00:24:21.825871842Z" level=info msg="received exit event container_id:\"03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6\" id:\"03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6\" pid:5175 exited_at:{seconds:1752107061 nanos:825629044}" Jul 10 00:24:21.826568 containerd[1920]: time="2025-07-10T00:24:21.826548851Z" level=info msg="TaskExit event in podsandbox handler container_id:\"03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6\" id:\"03adcfbcda723a57cecac69d71531f2e7164c17427b99222138e77fd9c19e8c6\" pid:5175 exited_at:{seconds:1752107061 nanos:825629044}" Jul 10 00:24:21.831461 sshd[5140]: Accepted publickey for core from 139.178.89.65 port 52228 ssh2: RSA SHA256:8gcBu3X/zjMKtjKrMkKIwTrYfDQG3sNa69IzDxa0i3U Jul 10 00:24:21.833607 sshd-session[5140]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:24:21.840700 systemd-logind[1894]: New session 27 of user core. Jul 10 00:24:21.846557 systemd[1]: Started session-27.scope - Session 27 of User core. Jul 10 00:24:22.426862 containerd[1920]: time="2025-07-10T00:24:22.426728789Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 10 00:24:22.446551 containerd[1920]: time="2025-07-10T00:24:22.446404731Z" level=info msg="Container 7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:24:22.457367 containerd[1920]: time="2025-07-10T00:24:22.457322266Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd\"" Jul 10 00:24:22.458120 containerd[1920]: time="2025-07-10T00:24:22.458031097Z" level=info msg="StartContainer for \"7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd\"" Jul 10 00:24:22.459133 containerd[1920]: time="2025-07-10T00:24:22.459022329Z" level=info msg="connecting to shim 7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd" address="unix:///run/containerd/s/4da342f01d5ae30adf1257f90aa03cd9e6d854b0647b1ba49bdba668f172ce69" protocol=ttrpc version=3 Jul 10 00:24:22.481352 systemd[1]: Started cri-containerd-7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd.scope - libcontainer container 7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd. Jul 10 00:24:22.529580 containerd[1920]: time="2025-07-10T00:24:22.529538231Z" level=info msg="StartContainer for \"7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd\" returns successfully" Jul 10 00:24:22.539195 systemd[1]: cri-containerd-7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd.scope: Deactivated successfully. Jul 10 00:24:22.539524 systemd[1]: cri-containerd-7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd.scope: Consumed 21ms CPU time, 7.2M memory peak, 2.1M read from disk. Jul 10 00:24:22.540797 containerd[1920]: time="2025-07-10T00:24:22.540754771Z" level=info msg="received exit event container_id:\"7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd\" id:\"7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd\" pid:5230 exited_at:{seconds:1752107062 nanos:539949109}" Jul 10 00:24:22.542284 containerd[1920]: time="2025-07-10T00:24:22.542085060Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd\" id:\"7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd\" pid:5230 exited_at:{seconds:1752107062 nanos:539949109}" Jul 10 00:24:22.563996 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7e81aeace442fc8cd2deaf12141e0dd34006392460c5a73b5f103b467e9aeabd-rootfs.mount: Deactivated successfully. Jul 10 00:24:23.432224 containerd[1920]: time="2025-07-10T00:24:23.432083380Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 10 00:24:23.454494 containerd[1920]: time="2025-07-10T00:24:23.454445163Z" level=info msg="Container 46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:24:23.466518 containerd[1920]: time="2025-07-10T00:24:23.466476506Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef\"" Jul 10 00:24:23.467154 containerd[1920]: time="2025-07-10T00:24:23.467132297Z" level=info msg="StartContainer for \"46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef\"" Jul 10 00:24:23.469119 containerd[1920]: time="2025-07-10T00:24:23.468983205Z" level=info msg="connecting to shim 46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef" address="unix:///run/containerd/s/4da342f01d5ae30adf1257f90aa03cd9e6d854b0647b1ba49bdba668f172ce69" protocol=ttrpc version=3 Jul 10 00:24:23.496318 systemd[1]: Started cri-containerd-46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef.scope - libcontainer container 46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef. Jul 10 00:24:23.544904 containerd[1920]: time="2025-07-10T00:24:23.543400048Z" level=info msg="StartContainer for \"46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef\" returns successfully" Jul 10 00:24:23.551012 systemd[1]: cri-containerd-46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef.scope: Deactivated successfully. Jul 10 00:24:23.551511 systemd[1]: cri-containerd-46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef.scope: Consumed 25ms CPU time, 5.9M memory peak, 1.1M read from disk. Jul 10 00:24:23.552374 containerd[1920]: time="2025-07-10T00:24:23.552336997Z" level=info msg="received exit event container_id:\"46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef\" id:\"46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef\" pid:5274 exited_at:{seconds:1752107063 nanos:552156281}" Jul 10 00:24:23.553649 containerd[1920]: time="2025-07-10T00:24:23.553609404Z" level=info msg="TaskExit event in podsandbox handler container_id:\"46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef\" id:\"46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef\" pid:5274 exited_at:{seconds:1752107063 nanos:552156281}" Jul 10 00:24:23.578853 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-46085b120c23622bb1e2609f14c12897ab2d7a803c96560603e073c352050bef-rootfs.mount: Deactivated successfully. Jul 10 00:24:24.437581 containerd[1920]: time="2025-07-10T00:24:24.437536568Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 10 00:24:24.452030 containerd[1920]: time="2025-07-10T00:24:24.451982157Z" level=info msg="Container a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:24:24.463964 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2389231337.mount: Deactivated successfully. Jul 10 00:24:24.470054 containerd[1920]: time="2025-07-10T00:24:24.470014868Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565\"" Jul 10 00:24:24.470720 containerd[1920]: time="2025-07-10T00:24:24.470678847Z" level=info msg="StartContainer for \"a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565\"" Jul 10 00:24:24.471960 containerd[1920]: time="2025-07-10T00:24:24.471887904Z" level=info msg="connecting to shim a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565" address="unix:///run/containerd/s/4da342f01d5ae30adf1257f90aa03cd9e6d854b0647b1ba49bdba668f172ce69" protocol=ttrpc version=3 Jul 10 00:24:24.497495 systemd[1]: Started cri-containerd-a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565.scope - libcontainer container a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565. Jul 10 00:24:24.531525 systemd[1]: cri-containerd-a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565.scope: Deactivated successfully. Jul 10 00:24:24.534143 containerd[1920]: time="2025-07-10T00:24:24.533090518Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565\" id:\"a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565\" pid:5314 exited_at:{seconds:1752107064 nanos:532733661}" Jul 10 00:24:24.534943 containerd[1920]: time="2025-07-10T00:24:24.534910241Z" level=info msg="received exit event container_id:\"a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565\" id:\"a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565\" pid:5314 exited_at:{seconds:1752107064 nanos:532733661}" Jul 10 00:24:24.545088 containerd[1920]: time="2025-07-10T00:24:24.545048390Z" level=info msg="StartContainer for \"a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565\" returns successfully" Jul 10 00:24:24.562545 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a864010159207d55df7ab5b74538174dda357bb6e22d98e0cb3c35cf0f4c5565-rootfs.mount: Deactivated successfully. Jul 10 00:24:25.173816 kubelet[3385]: E0710 00:24:25.173525 3385 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 10 00:24:25.446809 containerd[1920]: time="2025-07-10T00:24:25.446588597Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 10 00:24:25.474198 containerd[1920]: time="2025-07-10T00:24:25.474123578Z" level=info msg="Container 5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:24:25.477703 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount84198171.mount: Deactivated successfully. Jul 10 00:24:25.488685 containerd[1920]: time="2025-07-10T00:24:25.488486107Z" level=info msg="CreateContainer within sandbox \"f618997de78fb35cd4c690ce4d5eba50c88a283cb86f89695623eaa137f42bfc\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\"" Jul 10 00:24:25.489407 containerd[1920]: time="2025-07-10T00:24:25.489383628Z" level=info msg="StartContainer for \"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\"" Jul 10 00:24:25.490480 containerd[1920]: time="2025-07-10T00:24:25.490421142Z" level=info msg="connecting to shim 5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e" address="unix:///run/containerd/s/4da342f01d5ae30adf1257f90aa03cd9e6d854b0647b1ba49bdba668f172ce69" protocol=ttrpc version=3 Jul 10 00:24:25.518363 systemd[1]: Started cri-containerd-5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e.scope - libcontainer container 5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e. Jul 10 00:24:25.565400 containerd[1920]: time="2025-07-10T00:24:25.565332365Z" level=info msg="StartContainer for \"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\" returns successfully" Jul 10 00:24:25.708864 containerd[1920]: time="2025-07-10T00:24:25.708282403Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\" id:\"150b3300e42473881486cab446e05d8c7c09e215186d799bd706f2c94cb23f6d\" pid:5387 exited_at:{seconds:1752107065 nanos:702397894}" Jul 10 00:24:26.147140 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni-avx)) Jul 10 00:24:26.463235 kubelet[3385]: I0710 00:24:26.462883 3385 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-4cjwz" podStartSLOduration=5.462869366 podStartE2EDuration="5.462869366s" podCreationTimestamp="2025-07-10 00:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:24:26.462538624 +0000 UTC m=+101.845834642" watchObservedRunningTime="2025-07-10 00:24:26.462869366 +0000 UTC m=+101.846165370" Jul 10 00:24:27.105853 kubelet[3385]: I0710 00:24:27.105532 3385 setters.go:618] "Node became not ready" node="ip-172-31-20-136" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-07-10T00:24:27Z","lastTransitionTime":"2025-07-10T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Jul 10 00:24:28.534637 containerd[1920]: time="2025-07-10T00:24:28.534198509Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\" id:\"defdd87148b053b9cd37caee33e453dfc7095b384a5749d493d1e471791c8ef6\" pid:5632 exit_status:1 exited_at:{seconds:1752107068 nanos:533699381}" Jul 10 00:24:29.005211 kubelet[3385]: E0710 00:24:29.003490 3385 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-674b8bbfcf-dqpm9" podUID="aa6372db-0667-4c8d-bb2d-de410e983504" Jul 10 00:24:29.308034 (udev-worker)[5871]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:24:29.310605 (udev-worker)[5872]: Network interface NamePolicy= disabled on kernel command line. Jul 10 00:24:29.311597 systemd-networkd[1768]: lxc_health: Link UP Jul 10 00:24:29.323278 systemd-networkd[1768]: lxc_health: Gained carrier Jul 10 00:24:30.452317 systemd-networkd[1768]: lxc_health: Gained IPv6LL Jul 10 00:24:30.741178 containerd[1920]: time="2025-07-10T00:24:30.741030069Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\" id:\"2171ca4e434b43c324e6a8479a689cb7a3e3f083157a99096b3795558efa4953\" pid:5912 exited_at:{seconds:1752107070 nanos:740438359}" Jul 10 00:24:30.754716 kubelet[3385]: E0710 00:24:30.754648 3385 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:50186->127.0.0.1:39055: write tcp 127.0.0.1:50186->127.0.0.1:39055: write: broken pipe Jul 10 00:24:32.780500 ntpd[1885]: Listen normally on 15 lxc_health [fe80::6c95:2ff:fe18:9ddb%14]:123 Jul 10 00:24:32.781011 ntpd[1885]: 10 Jul 00:24:32 ntpd[1885]: Listen normally on 15 lxc_health [fe80::6c95:2ff:fe18:9ddb%14]:123 Jul 10 00:24:33.002141 containerd[1920]: time="2025-07-10T00:24:33.001700211Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\" id:\"8032432f9d7707b73bdae76bbac82d7c43a5f195c2c5a329aebba8c308c29bea\" pid:5937 exited_at:{seconds:1752107073 nanos:87609}" Jul 10 00:24:35.151904 containerd[1920]: time="2025-07-10T00:24:35.151866730Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5d7c1a65408136f20f5e2529417f0366a807f3c62fc0b3b1b2380c57a5344f7e\" id:\"cc168764cfceb423779209ce736c3397bbba8a838eb480f4e8e521f008b70d25\" pid:5969 exited_at:{seconds:1752107075 nanos:151312195}" Jul 10 00:24:35.179082 sshd[5210]: Connection closed by 139.178.89.65 port 52228 Jul 10 00:24:35.181735 sshd-session[5140]: pam_unix(sshd:session): session closed for user core Jul 10 00:24:35.187498 systemd[1]: sshd@26-172.31.20.136:22-139.178.89.65:52228.service: Deactivated successfully. Jul 10 00:24:35.190688 systemd[1]: session-27.scope: Deactivated successfully. Jul 10 00:24:35.192737 systemd-logind[1894]: Session 27 logged out. Waiting for processes to exit. Jul 10 00:24:35.194505 systemd-logind[1894]: Removed session 27. Jul 10 00:24:44.954262 containerd[1920]: time="2025-07-10T00:24:44.954217724Z" level=info msg="StopPodSandbox for \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\"" Jul 10 00:24:44.954893 containerd[1920]: time="2025-07-10T00:24:44.954805884Z" level=info msg="TearDown network for sandbox \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" successfully" Jul 10 00:24:44.954893 containerd[1920]: time="2025-07-10T00:24:44.954828473Z" level=info msg="StopPodSandbox for \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" returns successfully" Jul 10 00:24:44.955273 containerd[1920]: time="2025-07-10T00:24:44.955239418Z" level=info msg="RemovePodSandbox for \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\"" Jul 10 00:24:44.967576 containerd[1920]: time="2025-07-10T00:24:44.967527964Z" level=info msg="Forcibly stopping sandbox \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\"" Jul 10 00:24:44.967714 containerd[1920]: time="2025-07-10T00:24:44.967684838Z" level=info msg="TearDown network for sandbox \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" successfully" Jul 10 00:24:44.972958 containerd[1920]: time="2025-07-10T00:24:44.972902012Z" level=info msg="Ensure that sandbox f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc in task-service has been cleanup successfully" Jul 10 00:24:44.978850 containerd[1920]: time="2025-07-10T00:24:44.978808759Z" level=info msg="RemovePodSandbox \"f934c41af0c1cfb8ac378795d809301cdd6812b3c4b4a10c7a1808313fc240cc\" returns successfully" Jul 10 00:24:44.979314 containerd[1920]: time="2025-07-10T00:24:44.979279830Z" level=info msg="StopPodSandbox for \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\"" Jul 10 00:24:44.979409 containerd[1920]: time="2025-07-10T00:24:44.979395728Z" level=info msg="TearDown network for sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" successfully" Jul 10 00:24:44.979409 containerd[1920]: time="2025-07-10T00:24:44.979406776Z" level=info msg="StopPodSandbox for \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" returns successfully" Jul 10 00:24:44.979800 containerd[1920]: time="2025-07-10T00:24:44.979777598Z" level=info msg="RemovePodSandbox for \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\"" Jul 10 00:24:44.979800 containerd[1920]: time="2025-07-10T00:24:44.979800097Z" level=info msg="Forcibly stopping sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\"" Jul 10 00:24:44.979910 containerd[1920]: time="2025-07-10T00:24:44.979876322Z" level=info msg="TearDown network for sandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" successfully" Jul 10 00:24:44.981930 containerd[1920]: time="2025-07-10T00:24:44.981624822Z" level=info msg="Ensure that sandbox dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae in task-service has been cleanup successfully" Jul 10 00:24:44.988994 containerd[1920]: time="2025-07-10T00:24:44.988947731Z" level=info msg="RemovePodSandbox \"dea987c2c56cbe52bb562d6d7afb21aaa516c2ffb902c9c1ab62c949bef935ae\" returns successfully" Jul 10 00:24:49.364977 systemd[1]: cri-containerd-79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c.scope: Deactivated successfully. Jul 10 00:24:49.365886 systemd[1]: cri-containerd-79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c.scope: Consumed 3.106s CPU time, 76.7M memory peak, 28.1M read from disk. Jul 10 00:24:49.370656 containerd[1920]: time="2025-07-10T00:24:49.370603443Z" level=info msg="received exit event container_id:\"79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c\" id:\"79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c\" pid:3049 exit_status:1 exited_at:{seconds:1752107089 nanos:369775157}" Jul 10 00:24:49.371831 containerd[1920]: time="2025-07-10T00:24:49.371792036Z" level=info msg="TaskExit event in podsandbox handler container_id:\"79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c\" id:\"79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c\" pid:3049 exit_status:1 exited_at:{seconds:1752107089 nanos:369775157}" Jul 10 00:24:49.397772 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c-rootfs.mount: Deactivated successfully. Jul 10 00:24:49.500375 kubelet[3385]: I0710 00:24:49.500216 3385 scope.go:117] "RemoveContainer" containerID="79e971eed17a1f39da7c47795cbf18b246317ddb49ef1ed271660a4a7a24c78c" Jul 10 00:24:49.504668 containerd[1920]: time="2025-07-10T00:24:49.504613654Z" level=info msg="CreateContainer within sandbox \"680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Jul 10 00:24:49.521717 containerd[1920]: time="2025-07-10T00:24:49.521485246Z" level=info msg="Container 659fe7e6829384ba32d9e627fe4ebda8f10914b2373ac80e3717e8f608d27dbb: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:24:49.528862 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount962882507.mount: Deactivated successfully. Jul 10 00:24:49.533562 containerd[1920]: time="2025-07-10T00:24:49.533520790Z" level=info msg="CreateContainer within sandbox \"680c105793b5bbd3d7c5a7e3b503cb9f56b34ef38b38298ef569af93854a025a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"659fe7e6829384ba32d9e627fe4ebda8f10914b2373ac80e3717e8f608d27dbb\"" Jul 10 00:24:49.535509 containerd[1920]: time="2025-07-10T00:24:49.534068360Z" level=info msg="StartContainer for \"659fe7e6829384ba32d9e627fe4ebda8f10914b2373ac80e3717e8f608d27dbb\"" Jul 10 00:24:49.535509 containerd[1920]: time="2025-07-10T00:24:49.535189786Z" level=info msg="connecting to shim 659fe7e6829384ba32d9e627fe4ebda8f10914b2373ac80e3717e8f608d27dbb" address="unix:///run/containerd/s/0dd3687a6cdb49c98a64c39dc9dcf00c182cd6d8d4d31141822e1bd3f9af0e93" protocol=ttrpc version=3 Jul 10 00:24:49.561319 systemd[1]: Started cri-containerd-659fe7e6829384ba32d9e627fe4ebda8f10914b2373ac80e3717e8f608d27dbb.scope - libcontainer container 659fe7e6829384ba32d9e627fe4ebda8f10914b2373ac80e3717e8f608d27dbb. Jul 10 00:24:49.623382 containerd[1920]: time="2025-07-10T00:24:49.623273691Z" level=info msg="StartContainer for \"659fe7e6829384ba32d9e627fe4ebda8f10914b2373ac80e3717e8f608d27dbb\" returns successfully" Jul 10 00:24:54.198832 systemd[1]: cri-containerd-b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb.scope: Deactivated successfully. Jul 10 00:24:54.199928 systemd[1]: cri-containerd-b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb.scope: Consumed 2.802s CPU time, 29.8M memory peak, 10.7M read from disk. Jul 10 00:24:54.203698 containerd[1920]: time="2025-07-10T00:24:54.203663380Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb\" id:\"b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb\" pid:3078 exit_status:1 exited_at:{seconds:1752107094 nanos:203264375}" Jul 10 00:24:54.204079 containerd[1920]: time="2025-07-10T00:24:54.204014283Z" level=info msg="received exit event container_id:\"b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb\" id:\"b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb\" pid:3078 exit_status:1 exited_at:{seconds:1752107094 nanos:203264375}" Jul 10 00:24:54.231827 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb-rootfs.mount: Deactivated successfully. Jul 10 00:24:54.515510 kubelet[3385]: I0710 00:24:54.515389 3385 scope.go:117] "RemoveContainer" containerID="b66983a13a13c5fa1d29918c379834b288d2d4f782b05c0b512c52076611e0bb" Jul 10 00:24:54.518777 containerd[1920]: time="2025-07-10T00:24:54.518290093Z" level=info msg="CreateContainer within sandbox \"9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Jul 10 00:24:54.537836 containerd[1920]: time="2025-07-10T00:24:54.537693820Z" level=info msg="Container ab0f6003610f433b041f99db62927ef3b548e0c2ec97dd7c2559674d472bb5bd: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:24:54.553089 containerd[1920]: time="2025-07-10T00:24:54.553009541Z" level=info msg="CreateContainer within sandbox \"9bf2e64cc8ff2b73279179f7d5a2b6551bc96f2f61d34a5301f7e42060a4ebe0\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"ab0f6003610f433b041f99db62927ef3b548e0c2ec97dd7c2559674d472bb5bd\"" Jul 10 00:24:54.553546 containerd[1920]: time="2025-07-10T00:24:54.553504383Z" level=info msg="StartContainer for \"ab0f6003610f433b041f99db62927ef3b548e0c2ec97dd7c2559674d472bb5bd\"" Jul 10 00:24:54.554532 containerd[1920]: time="2025-07-10T00:24:54.554500545Z" level=info msg="connecting to shim ab0f6003610f433b041f99db62927ef3b548e0c2ec97dd7c2559674d472bb5bd" address="unix:///run/containerd/s/f606d54e732c51176a2d2fb18bf9c5df71e9e1c753c16246876de19be9e6ba3e" protocol=ttrpc version=3 Jul 10 00:24:54.582365 systemd[1]: Started cri-containerd-ab0f6003610f433b041f99db62927ef3b548e0c2ec97dd7c2559674d472bb5bd.scope - libcontainer container ab0f6003610f433b041f99db62927ef3b548e0c2ec97dd7c2559674d472bb5bd. Jul 10 00:24:54.640285 containerd[1920]: time="2025-07-10T00:24:54.640093293Z" level=info msg="StartContainer for \"ab0f6003610f433b041f99db62927ef3b548e0c2ec97dd7c2559674d472bb5bd\" returns successfully" Jul 10 00:24:57.485152 kubelet[3385]: E0710 00:24:57.485028 3385 controller.go:195] "Failed to update lease" err="Put \"https://172.31.20.136:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-136?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jul 10 00:25:07.485400 kubelet[3385]: E0710 00:25:07.485320 3385 controller.go:195] "Failed to update lease" err="Put \"https://172.31.20.136:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-136?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"