Jul 12 00:08:34.239014 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd083] Jul 12 00:08:34.239059 kernel: Linux version 6.6.96-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Fri Jul 11 22:42:11 -00 2025 Jul 12 00:08:34.239085 kernel: KASLR disabled due to lack of seed Jul 12 00:08:34.239101 kernel: efi: EFI v2.7 by EDK II Jul 12 00:08:34.239117 kernel: efi: SMBIOS=0x7bed0000 SMBIOS 3.0=0x7beb0000 ACPI=0x786e0000 ACPI 2.0=0x786e0014 MEMATTR=0x7affea98 MEMRESERVE=0x7852ee18 Jul 12 00:08:34.239134 kernel: ACPI: Early table checksum verification disabled Jul 12 00:08:34.239153 kernel: ACPI: RSDP 0x00000000786E0014 000024 (v02 AMAZON) Jul 12 00:08:34.239168 kernel: ACPI: XSDT 0x00000000786D00E8 000064 (v01 AMAZON AMZNFACP 00000001 01000013) Jul 12 00:08:34.239184 kernel: ACPI: FACP 0x00000000786B0000 000114 (v06 AMAZON AMZNFACP 00000001 AMZN 00000001) Jul 12 00:08:34.239199 kernel: ACPI: DSDT 0x0000000078640000 00159D (v02 AMAZON AMZNDSDT 00000001 INTL 20160527) Jul 12 00:08:34.239220 kernel: ACPI: APIC 0x00000000786C0000 000108 (v04 AMAZON AMZNAPIC 00000001 AMZN 00000001) Jul 12 00:08:34.239235 kernel: ACPI: SPCR 0x00000000786A0000 000050 (v02 AMAZON AMZNSPCR 00000001 AMZN 00000001) Jul 12 00:08:34.239251 kernel: ACPI: GTDT 0x0000000078690000 000060 (v02 AMAZON AMZNGTDT 00000001 AMZN 00000001) Jul 12 00:08:34.239266 kernel: ACPI: MCFG 0x0000000078680000 00003C (v02 AMAZON AMZNMCFG 00000001 AMZN 00000001) Jul 12 00:08:34.239285 kernel: ACPI: SLIT 0x0000000078670000 00002D (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Jul 12 00:08:34.239306 kernel: ACPI: IORT 0x0000000078660000 000078 (v01 AMAZON AMZNIORT 00000001 AMZN 00000001) Jul 12 00:08:34.239324 kernel: ACPI: PPTT 0x0000000078650000 0000EC (v01 AMAZON AMZNPPTT 00000001 AMZN 00000001) Jul 12 00:08:34.239340 kernel: ACPI: SPCR: console: uart,mmio,0x90a0000,115200 Jul 12 00:08:34.239356 kernel: earlycon: uart0 at MMIO 0x00000000090a0000 (options '115200') Jul 12 00:08:34.239372 kernel: printk: bootconsole [uart0] enabled Jul 12 00:08:34.239389 kernel: NUMA: Failed to initialise from firmware Jul 12 00:08:34.239405 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000004b5ffffff] Jul 12 00:08:34.239422 kernel: NUMA: NODE_DATA [mem 0x4b583f800-0x4b5844fff] Jul 12 00:08:34.240529 kernel: Zone ranges: Jul 12 00:08:34.240554 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Jul 12 00:08:34.240571 kernel: DMA32 empty Jul 12 00:08:34.240596 kernel: Normal [mem 0x0000000100000000-0x00000004b5ffffff] Jul 12 00:08:34.240613 kernel: Movable zone start for each node Jul 12 00:08:34.240630 kernel: Early memory node ranges Jul 12 00:08:34.240646 kernel: node 0: [mem 0x0000000040000000-0x000000007862ffff] Jul 12 00:08:34.240662 kernel: node 0: [mem 0x0000000078630000-0x000000007863ffff] Jul 12 00:08:34.240678 kernel: node 0: [mem 0x0000000078640000-0x00000000786effff] Jul 12 00:08:34.240695 kernel: node 0: [mem 0x00000000786f0000-0x000000007872ffff] Jul 12 00:08:34.240711 kernel: node 0: [mem 0x0000000078730000-0x000000007bbfffff] Jul 12 00:08:34.240727 kernel: node 0: [mem 0x000000007bc00000-0x000000007bfdffff] Jul 12 00:08:34.240743 kernel: node 0: [mem 0x000000007bfe0000-0x000000007fffffff] Jul 12 00:08:34.240759 kernel: node 0: [mem 0x0000000400000000-0x00000004b5ffffff] Jul 12 00:08:34.240775 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000004b5ffffff] Jul 12 00:08:34.240796 kernel: On node 0, zone Normal: 8192 pages in unavailable ranges Jul 12 00:08:34.240813 kernel: psci: probing for conduit method from ACPI. Jul 12 00:08:34.240836 kernel: psci: PSCIv1.0 detected in firmware. Jul 12 00:08:34.240854 kernel: psci: Using standard PSCI v0.2 function IDs Jul 12 00:08:34.240872 kernel: psci: Trusted OS migration not required Jul 12 00:08:34.240893 kernel: psci: SMC Calling Convention v1.1 Jul 12 00:08:34.240911 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000001) Jul 12 00:08:34.240928 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Jul 12 00:08:34.240945 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Jul 12 00:08:34.240962 kernel: pcpu-alloc: [0] 0 [0] 1 Jul 12 00:08:34.240979 kernel: Detected PIPT I-cache on CPU0 Jul 12 00:08:34.240997 kernel: CPU features: detected: GIC system register CPU interface Jul 12 00:08:34.241013 kernel: CPU features: detected: Spectre-v2 Jul 12 00:08:34.241031 kernel: CPU features: detected: Spectre-v3a Jul 12 00:08:34.241048 kernel: CPU features: detected: Spectre-BHB Jul 12 00:08:34.241065 kernel: CPU features: detected: ARM erratum 1742098 Jul 12 00:08:34.241086 kernel: CPU features: detected: ARM errata 1165522, 1319367, or 1530923 Jul 12 00:08:34.241104 kernel: alternatives: applying boot alternatives Jul 12 00:08:34.241123 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=52e0eba0325ad9e58f7b221f0132165c94b480ebf93a398f4fe935660ba9e15c Jul 12 00:08:34.241142 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 12 00:08:34.241159 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jul 12 00:08:34.241177 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jul 12 00:08:34.241194 kernel: Fallback order for Node 0: 0 Jul 12 00:08:34.241211 kernel: Built 1 zonelists, mobility grouping on. Total pages: 991872 Jul 12 00:08:34.241228 kernel: Policy zone: Normal Jul 12 00:08:34.241245 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 12 00:08:34.241262 kernel: software IO TLB: area num 2. Jul 12 00:08:34.241284 kernel: software IO TLB: mapped [mem 0x000000007c000000-0x0000000080000000] (64MB) Jul 12 00:08:34.241302 kernel: Memory: 3820088K/4030464K available (10304K kernel code, 2186K rwdata, 8108K rodata, 39424K init, 897K bss, 210376K reserved, 0K cma-reserved) Jul 12 00:08:34.241319 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jul 12 00:08:34.241336 kernel: rcu: Preemptible hierarchical RCU implementation. Jul 12 00:08:34.241354 kernel: rcu: RCU event tracing is enabled. Jul 12 00:08:34.241372 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jul 12 00:08:34.241389 kernel: Trampoline variant of Tasks RCU enabled. Jul 12 00:08:34.241407 kernel: Tracing variant of Tasks RCU enabled. Jul 12 00:08:34.241424 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 12 00:08:34.242132 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jul 12 00:08:34.242151 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Jul 12 00:08:34.242176 kernel: GICv3: 96 SPIs implemented Jul 12 00:08:34.242194 kernel: GICv3: 0 Extended SPIs implemented Jul 12 00:08:34.242211 kernel: Root IRQ handler: gic_handle_irq Jul 12 00:08:34.242228 kernel: GICv3: GICv3 features: 16 PPIs Jul 12 00:08:34.242245 kernel: GICv3: CPU0: found redistributor 0 region 0:0x0000000010200000 Jul 12 00:08:34.242262 kernel: ITS [mem 0x10080000-0x1009ffff] Jul 12 00:08:34.242279 kernel: ITS@0x0000000010080000: allocated 8192 Devices @4000b0000 (indirect, esz 8, psz 64K, shr 1) Jul 12 00:08:34.242296 kernel: ITS@0x0000000010080000: allocated 8192 Interrupt Collections @4000c0000 (flat, esz 8, psz 64K, shr 1) Jul 12 00:08:34.242314 kernel: GICv3: using LPI property table @0x00000004000d0000 Jul 12 00:08:34.242331 kernel: ITS: Using hypervisor restricted LPI range [128] Jul 12 00:08:34.242348 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000004000e0000 Jul 12 00:08:34.242365 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jul 12 00:08:34.242413 kernel: arch_timer: cp15 timer(s) running at 83.33MHz (virt). Jul 12 00:08:34.244121 kernel: clocksource: arch_sys_counter: mask: 0x1ffffffffffffff max_cycles: 0x13381ebeec, max_idle_ns: 440795203145 ns Jul 12 00:08:34.244155 kernel: sched_clock: 57 bits at 83MHz, resolution 12ns, wraps every 4398046511100ns Jul 12 00:08:34.244173 kernel: Console: colour dummy device 80x25 Jul 12 00:08:34.244191 kernel: printk: console [tty1] enabled Jul 12 00:08:34.244209 kernel: ACPI: Core revision 20230628 Jul 12 00:08:34.244238 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 166.66 BogoMIPS (lpj=83333) Jul 12 00:08:34.244261 kernel: pid_max: default: 32768 minimum: 301 Jul 12 00:08:34.244281 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jul 12 00:08:34.244307 kernel: landlock: Up and running. Jul 12 00:08:34.244325 kernel: SELinux: Initializing. Jul 12 00:08:34.244343 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 12 00:08:34.244361 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jul 12 00:08:34.244379 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 12 00:08:34.244397 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jul 12 00:08:34.244415 kernel: rcu: Hierarchical SRCU implementation. Jul 12 00:08:34.244476 kernel: rcu: Max phase no-delay instances is 400. Jul 12 00:08:34.244501 kernel: Platform MSI: ITS@0x10080000 domain created Jul 12 00:08:34.244525 kernel: PCI/MSI: ITS@0x10080000 domain created Jul 12 00:08:34.244543 kernel: Remapping and enabling EFI services. Jul 12 00:08:34.244560 kernel: smp: Bringing up secondary CPUs ... Jul 12 00:08:34.244578 kernel: Detected PIPT I-cache on CPU1 Jul 12 00:08:34.244596 kernel: GICv3: CPU1: found redistributor 1 region 0:0x0000000010220000 Jul 12 00:08:34.244614 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000004000f0000 Jul 12 00:08:34.244631 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd083] Jul 12 00:08:34.244649 kernel: smp: Brought up 1 node, 2 CPUs Jul 12 00:08:34.244666 kernel: SMP: Total of 2 processors activated. Jul 12 00:08:34.244683 kernel: CPU features: detected: 32-bit EL0 Support Jul 12 00:08:34.244706 kernel: CPU features: detected: 32-bit EL1 Support Jul 12 00:08:34.244723 kernel: CPU features: detected: CRC32 instructions Jul 12 00:08:34.244752 kernel: CPU: All CPU(s) started at EL1 Jul 12 00:08:34.244775 kernel: alternatives: applying system-wide alternatives Jul 12 00:08:34.244793 kernel: devtmpfs: initialized Jul 12 00:08:34.244811 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 12 00:08:34.244844 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jul 12 00:08:34.244865 kernel: pinctrl core: initialized pinctrl subsystem Jul 12 00:08:34.244883 kernel: SMBIOS 3.0.0 present. Jul 12 00:08:34.244907 kernel: DMI: Amazon EC2 a1.large/, BIOS 1.0 11/1/2018 Jul 12 00:08:34.244926 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 12 00:08:34.244944 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Jul 12 00:08:34.244963 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Jul 12 00:08:34.244981 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Jul 12 00:08:34.245000 kernel: audit: initializing netlink subsys (disabled) Jul 12 00:08:34.245018 kernel: audit: type=2000 audit(0.286:1): state=initialized audit_enabled=0 res=1 Jul 12 00:08:34.245041 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 12 00:08:34.245060 kernel: cpuidle: using governor menu Jul 12 00:08:34.245078 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Jul 12 00:08:34.245096 kernel: ASID allocator initialised with 65536 entries Jul 12 00:08:34.245114 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 12 00:08:34.245132 kernel: Serial: AMBA PL011 UART driver Jul 12 00:08:34.245150 kernel: Modules: 17488 pages in range for non-PLT usage Jul 12 00:08:34.245169 kernel: Modules: 509008 pages in range for PLT usage Jul 12 00:08:34.245187 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jul 12 00:08:34.245209 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Jul 12 00:08:34.245228 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Jul 12 00:08:34.245246 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Jul 12 00:08:34.245265 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jul 12 00:08:34.245283 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Jul 12 00:08:34.245301 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Jul 12 00:08:34.245319 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Jul 12 00:08:34.245337 kernel: ACPI: Added _OSI(Module Device) Jul 12 00:08:34.245355 kernel: ACPI: Added _OSI(Processor Device) Jul 12 00:08:34.245378 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 12 00:08:34.245396 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jul 12 00:08:34.245414 kernel: ACPI: Interpreter enabled Jul 12 00:08:34.250471 kernel: ACPI: Using GIC for interrupt routing Jul 12 00:08:34.250512 kernel: ACPI: MCFG table detected, 1 entries Jul 12 00:08:34.250531 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-0f]) Jul 12 00:08:34.250857 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jul 12 00:08:34.251068 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Jul 12 00:08:34.251276 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Jul 12 00:08:34.252882 kernel: acpi PNP0A08:00: ECAM area [mem 0x20000000-0x20ffffff] reserved by PNP0C02:00 Jul 12 00:08:34.253124 kernel: acpi PNP0A08:00: ECAM at [mem 0x20000000-0x20ffffff] for [bus 00-0f] Jul 12 00:08:34.253151 kernel: ACPI: Remapped I/O 0x000000001fff0000 to [io 0x0000-0xffff window] Jul 12 00:08:34.253171 kernel: acpiphp: Slot [1] registered Jul 12 00:08:34.253190 kernel: acpiphp: Slot [2] registered Jul 12 00:08:34.253208 kernel: acpiphp: Slot [3] registered Jul 12 00:08:34.253227 kernel: acpiphp: Slot [4] registered Jul 12 00:08:34.253254 kernel: acpiphp: Slot [5] registered Jul 12 00:08:34.253273 kernel: acpiphp: Slot [6] registered Jul 12 00:08:34.253292 kernel: acpiphp: Slot [7] registered Jul 12 00:08:34.253310 kernel: acpiphp: Slot [8] registered Jul 12 00:08:34.253328 kernel: acpiphp: Slot [9] registered Jul 12 00:08:34.253346 kernel: acpiphp: Slot [10] registered Jul 12 00:08:34.253364 kernel: acpiphp: Slot [11] registered Jul 12 00:08:34.253382 kernel: acpiphp: Slot [12] registered Jul 12 00:08:34.253400 kernel: acpiphp: Slot [13] registered Jul 12 00:08:34.253418 kernel: acpiphp: Slot [14] registered Jul 12 00:08:34.253488 kernel: acpiphp: Slot [15] registered Jul 12 00:08:34.253507 kernel: acpiphp: Slot [16] registered Jul 12 00:08:34.253526 kernel: acpiphp: Slot [17] registered Jul 12 00:08:34.253544 kernel: acpiphp: Slot [18] registered Jul 12 00:08:34.253562 kernel: acpiphp: Slot [19] registered Jul 12 00:08:34.253580 kernel: acpiphp: Slot [20] registered Jul 12 00:08:34.253599 kernel: acpiphp: Slot [21] registered Jul 12 00:08:34.253617 kernel: acpiphp: Slot [22] registered Jul 12 00:08:34.253635 kernel: acpiphp: Slot [23] registered Jul 12 00:08:34.253659 kernel: acpiphp: Slot [24] registered Jul 12 00:08:34.253678 kernel: acpiphp: Slot [25] registered Jul 12 00:08:34.253696 kernel: acpiphp: Slot [26] registered Jul 12 00:08:34.253714 kernel: acpiphp: Slot [27] registered Jul 12 00:08:34.253732 kernel: acpiphp: Slot [28] registered Jul 12 00:08:34.253750 kernel: acpiphp: Slot [29] registered Jul 12 00:08:34.253768 kernel: acpiphp: Slot [30] registered Jul 12 00:08:34.253786 kernel: acpiphp: Slot [31] registered Jul 12 00:08:34.253804 kernel: PCI host bridge to bus 0000:00 Jul 12 00:08:34.254024 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xffffffff window] Jul 12 00:08:34.254244 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Jul 12 00:08:34.255587 kernel: pci_bus 0000:00: root bus resource [mem 0x400000000000-0x407fffffffff window] Jul 12 00:08:34.255803 kernel: pci_bus 0000:00: root bus resource [bus 00-0f] Jul 12 00:08:34.256033 kernel: pci 0000:00:00.0: [1d0f:0200] type 00 class 0x060000 Jul 12 00:08:34.256253 kernel: pci 0000:00:01.0: [1d0f:8250] type 00 class 0x070003 Jul 12 00:08:34.256483 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80118000-0x80118fff] Jul 12 00:08:34.256717 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Jul 12 00:08:34.256933 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80114000-0x80117fff] Jul 12 00:08:34.257143 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D2 D3hot D3cold Jul 12 00:08:34.257367 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Jul 12 00:08:34.259696 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80110000-0x80113fff] Jul 12 00:08:34.259944 kernel: pci 0000:00:05.0: reg 0x18: [mem 0x80000000-0x800fffff pref] Jul 12 00:08:34.260162 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x80100000-0x8010ffff] Jul 12 00:08:34.260375 kernel: pci 0000:00:05.0: PME# supported from D0 D1 D2 D3hot D3cold Jul 12 00:08:34.263174 kernel: pci 0000:00:05.0: BAR 2: assigned [mem 0x80000000-0x800fffff pref] Jul 12 00:08:34.267775 kernel: pci 0000:00:05.0: BAR 4: assigned [mem 0x80100000-0x8010ffff] Jul 12 00:08:34.268026 kernel: pci 0000:00:04.0: BAR 0: assigned [mem 0x80110000-0x80113fff] Jul 12 00:08:34.268246 kernel: pci 0000:00:05.0: BAR 0: assigned [mem 0x80114000-0x80117fff] Jul 12 00:08:34.268501 kernel: pci 0000:00:01.0: BAR 0: assigned [mem 0x80118000-0x80118fff] Jul 12 00:08:34.268707 kernel: pci_bus 0000:00: resource 4 [mem 0x80000000-0xffffffff window] Jul 12 00:08:34.268914 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Jul 12 00:08:34.269109 kernel: pci_bus 0000:00: resource 6 [mem 0x400000000000-0x407fffffffff window] Jul 12 00:08:34.269135 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Jul 12 00:08:34.269154 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Jul 12 00:08:34.269174 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Jul 12 00:08:34.269192 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Jul 12 00:08:34.269210 kernel: iommu: Default domain type: Translated Jul 12 00:08:34.269229 kernel: iommu: DMA domain TLB invalidation policy: strict mode Jul 12 00:08:34.269253 kernel: efivars: Registered efivars operations Jul 12 00:08:34.269271 kernel: vgaarb: loaded Jul 12 00:08:34.269289 kernel: clocksource: Switched to clocksource arch_sys_counter Jul 12 00:08:34.269307 kernel: VFS: Disk quotas dquot_6.6.0 Jul 12 00:08:34.269326 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 12 00:08:34.269344 kernel: pnp: PnP ACPI init Jul 12 00:08:34.269604 kernel: system 00:00: [mem 0x20000000-0x2fffffff] could not be reserved Jul 12 00:08:34.269633 kernel: pnp: PnP ACPI: found 1 devices Jul 12 00:08:34.269658 kernel: NET: Registered PF_INET protocol family Jul 12 00:08:34.269678 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jul 12 00:08:34.269697 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jul 12 00:08:34.269715 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 12 00:08:34.269734 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jul 12 00:08:34.269753 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jul 12 00:08:34.269771 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jul 12 00:08:34.269790 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 12 00:08:34.269808 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jul 12 00:08:34.269831 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 12 00:08:34.269850 kernel: PCI: CLS 0 bytes, default 64 Jul 12 00:08:34.269868 kernel: kvm [1]: HYP mode not available Jul 12 00:08:34.269886 kernel: Initialise system trusted keyrings Jul 12 00:08:34.269905 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jul 12 00:08:34.269923 kernel: Key type asymmetric registered Jul 12 00:08:34.269941 kernel: Asymmetric key parser 'x509' registered Jul 12 00:08:34.269959 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jul 12 00:08:34.269977 kernel: io scheduler mq-deadline registered Jul 12 00:08:34.270000 kernel: io scheduler kyber registered Jul 12 00:08:34.270018 kernel: io scheduler bfq registered Jul 12 00:08:34.270255 kernel: pl061_gpio ARMH0061:00: PL061 GPIO chip registered Jul 12 00:08:34.270285 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Jul 12 00:08:34.270305 kernel: ACPI: button: Power Button [PWRB] Jul 12 00:08:34.270325 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input1 Jul 12 00:08:34.270343 kernel: ACPI: button: Sleep Button [SLPB] Jul 12 00:08:34.270362 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 12 00:08:34.270407 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Jul 12 00:08:34.270638 kernel: serial 0000:00:01.0: enabling device (0010 -> 0012) Jul 12 00:08:34.270666 kernel: printk: console [ttyS0] disabled Jul 12 00:08:34.270686 kernel: 0000:00:01.0: ttyS0 at MMIO 0x80118000 (irq = 14, base_baud = 115200) is a 16550A Jul 12 00:08:34.270704 kernel: printk: console [ttyS0] enabled Jul 12 00:08:34.270723 kernel: printk: bootconsole [uart0] disabled Jul 12 00:08:34.270741 kernel: thunder_xcv, ver 1.0 Jul 12 00:08:34.270759 kernel: thunder_bgx, ver 1.0 Jul 12 00:08:34.270777 kernel: nicpf, ver 1.0 Jul 12 00:08:34.270801 kernel: nicvf, ver 1.0 Jul 12 00:08:34.271011 kernel: rtc-efi rtc-efi.0: registered as rtc0 Jul 12 00:08:34.271210 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-07-12T00:08:33 UTC (1752278913) Jul 12 00:08:34.271236 kernel: hid: raw HID events driver (C) Jiri Kosina Jul 12 00:08:34.271255 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 3 counters available Jul 12 00:08:34.271274 kernel: watchdog: Delayed init of the lockup detector failed: -19 Jul 12 00:08:34.271292 kernel: watchdog: Hard watchdog permanently disabled Jul 12 00:08:34.271310 kernel: NET: Registered PF_INET6 protocol family Jul 12 00:08:34.271334 kernel: Segment Routing with IPv6 Jul 12 00:08:34.271353 kernel: In-situ OAM (IOAM) with IPv6 Jul 12 00:08:34.271371 kernel: NET: Registered PF_PACKET protocol family Jul 12 00:08:34.271389 kernel: Key type dns_resolver registered Jul 12 00:08:34.271407 kernel: registered taskstats version 1 Jul 12 00:08:34.273026 kernel: Loading compiled-in X.509 certificates Jul 12 00:08:34.273079 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.96-flatcar: ed6b382df707adbd5942eaa048a1031fe26cbf15' Jul 12 00:08:34.273130 kernel: Key type .fscrypt registered Jul 12 00:08:34.273153 kernel: Key type fscrypt-provisioning registered Jul 12 00:08:34.273179 kernel: ima: No TPM chip found, activating TPM-bypass! Jul 12 00:08:34.273209 kernel: ima: Allocated hash algorithm: sha1 Jul 12 00:08:34.273227 kernel: ima: No architecture policies found Jul 12 00:08:34.273246 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Jul 12 00:08:34.273265 kernel: clk: Disabling unused clocks Jul 12 00:08:34.273283 kernel: Freeing unused kernel memory: 39424K Jul 12 00:08:34.273301 kernel: Run /init as init process Jul 12 00:08:34.273319 kernel: with arguments: Jul 12 00:08:34.273337 kernel: /init Jul 12 00:08:34.273354 kernel: with environment: Jul 12 00:08:34.273393 kernel: HOME=/ Jul 12 00:08:34.273413 kernel: TERM=linux Jul 12 00:08:34.273464 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 12 00:08:34.273490 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jul 12 00:08:34.273514 systemd[1]: Detected virtualization amazon. Jul 12 00:08:34.273534 systemd[1]: Detected architecture arm64. Jul 12 00:08:34.273553 systemd[1]: Running in initrd. Jul 12 00:08:34.273578 systemd[1]: No hostname configured, using default hostname. Jul 12 00:08:34.273598 systemd[1]: Hostname set to . Jul 12 00:08:34.273619 systemd[1]: Initializing machine ID from VM UUID. Jul 12 00:08:34.273638 systemd[1]: Queued start job for default target initrd.target. Jul 12 00:08:34.273658 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 12 00:08:34.273678 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 12 00:08:34.273699 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jul 12 00:08:34.273719 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 12 00:08:34.273744 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jul 12 00:08:34.273765 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jul 12 00:08:34.273788 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jul 12 00:08:34.273809 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jul 12 00:08:34.273829 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 12 00:08:34.273849 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 12 00:08:34.273869 systemd[1]: Reached target paths.target - Path Units. Jul 12 00:08:34.273893 systemd[1]: Reached target slices.target - Slice Units. Jul 12 00:08:34.273914 systemd[1]: Reached target swap.target - Swaps. Jul 12 00:08:34.273933 systemd[1]: Reached target timers.target - Timer Units. Jul 12 00:08:34.273953 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jul 12 00:08:34.273973 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 12 00:08:34.273993 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 12 00:08:34.274013 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jul 12 00:08:34.274033 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 12 00:08:34.274053 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 12 00:08:34.274077 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 12 00:08:34.274097 systemd[1]: Reached target sockets.target - Socket Units. Jul 12 00:08:34.274117 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jul 12 00:08:34.274137 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 12 00:08:34.274157 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jul 12 00:08:34.274177 systemd[1]: Starting systemd-fsck-usr.service... Jul 12 00:08:34.274197 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 12 00:08:34.274217 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 12 00:08:34.274241 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 00:08:34.274262 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jul 12 00:08:34.274282 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 12 00:08:34.274301 systemd[1]: Finished systemd-fsck-usr.service. Jul 12 00:08:34.274386 systemd-journald[251]: Collecting audit messages is disabled. Jul 12 00:08:34.275560 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 12 00:08:34.275611 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 00:08:34.275634 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 12 00:08:34.275655 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 12 00:08:34.275684 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 12 00:08:34.275707 systemd-journald[251]: Journal started Jul 12 00:08:34.275748 systemd-journald[251]: Runtime Journal (/run/log/journal/ec2a064f90693ef95d6132e4dfbd219a) is 8.0M, max 75.3M, 67.3M free. Jul 12 00:08:34.275826 kernel: Bridge firewalling registered Jul 12 00:08:34.233006 systemd-modules-load[252]: Inserted module 'overlay' Jul 12 00:08:34.276638 systemd-modules-load[252]: Inserted module 'br_netfilter' Jul 12 00:08:34.285450 systemd[1]: Started systemd-journald.service - Journal Service. Jul 12 00:08:34.288313 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 12 00:08:34.297674 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 12 00:08:34.301877 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 12 00:08:34.308861 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 12 00:08:34.336034 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 12 00:08:34.362655 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 12 00:08:34.374120 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 12 00:08:34.380881 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 12 00:08:34.390823 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jul 12 00:08:34.404758 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 12 00:08:34.433458 dracut-cmdline[287]: dracut-dracut-053 Jul 12 00:08:34.438280 dracut-cmdline[287]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=52e0eba0325ad9e58f7b221f0132165c94b480ebf93a398f4fe935660ba9e15c Jul 12 00:08:34.484599 systemd-resolved[289]: Positive Trust Anchors: Jul 12 00:08:34.484633 systemd-resolved[289]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 12 00:08:34.484696 systemd-resolved[289]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 12 00:08:34.603453 kernel: SCSI subsystem initialized Jul 12 00:08:34.609473 kernel: Loading iSCSI transport class v2.0-870. Jul 12 00:08:34.621471 kernel: iscsi: registered transport (tcp) Jul 12 00:08:34.644471 kernel: iscsi: registered transport (qla4xxx) Jul 12 00:08:34.644542 kernel: QLogic iSCSI HBA Driver Jul 12 00:08:34.728458 kernel: random: crng init done Jul 12 00:08:34.728700 systemd-resolved[289]: Defaulting to hostname 'linux'. Jul 12 00:08:34.732573 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 12 00:08:34.735878 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 12 00:08:34.762466 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jul 12 00:08:34.771745 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jul 12 00:08:34.817461 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 12 00:08:34.817554 kernel: device-mapper: uevent: version 1.0.3 Jul 12 00:08:34.817596 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jul 12 00:08:34.884476 kernel: raid6: neonx8 gen() 6703 MB/s Jul 12 00:08:34.901465 kernel: raid6: neonx4 gen() 6502 MB/s Jul 12 00:08:34.918465 kernel: raid6: neonx2 gen() 5425 MB/s Jul 12 00:08:34.935463 kernel: raid6: neonx1 gen() 3946 MB/s Jul 12 00:08:34.952462 kernel: raid6: int64x8 gen() 3801 MB/s Jul 12 00:08:34.969462 kernel: raid6: int64x4 gen() 3717 MB/s Jul 12 00:08:34.986462 kernel: raid6: int64x2 gen() 3593 MB/s Jul 12 00:08:35.004434 kernel: raid6: int64x1 gen() 2764 MB/s Jul 12 00:08:35.004469 kernel: raid6: using algorithm neonx8 gen() 6703 MB/s Jul 12 00:08:35.022424 kernel: raid6: .... xor() 4875 MB/s, rmw enabled Jul 12 00:08:35.022499 kernel: raid6: using neon recovery algorithm Jul 12 00:08:35.030466 kernel: xor: measuring software checksum speed Jul 12 00:08:35.030526 kernel: 8regs : 9911 MB/sec Jul 12 00:08:35.033793 kernel: 32regs : 10869 MB/sec Jul 12 00:08:35.033826 kernel: arm64_neon : 9568 MB/sec Jul 12 00:08:35.033850 kernel: xor: using function: 32regs (10869 MB/sec) Jul 12 00:08:35.119484 kernel: Btrfs loaded, zoned=no, fsverity=no Jul 12 00:08:35.138884 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jul 12 00:08:35.148761 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 12 00:08:35.194061 systemd-udevd[471]: Using default interface naming scheme 'v255'. Jul 12 00:08:35.203264 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 12 00:08:35.216306 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jul 12 00:08:35.258459 dracut-pre-trigger[476]: rd.md=0: removing MD RAID activation Jul 12 00:08:35.317850 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jul 12 00:08:35.326713 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 12 00:08:35.455745 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 12 00:08:35.474765 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jul 12 00:08:35.522302 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jul 12 00:08:35.527920 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jul 12 00:08:35.538636 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 12 00:08:35.541283 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 12 00:08:35.555760 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jul 12 00:08:35.600314 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jul 12 00:08:35.651751 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Jul 12 00:08:35.651814 kernel: ena 0000:00:05.0: enabling device (0010 -> 0012) Jul 12 00:08:35.662458 kernel: ena 0000:00:05.0: ENA device version: 0.10 Jul 12 00:08:35.662790 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Jul 12 00:08:35.674466 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80114000, mac addr 06:1d:9c:3a:25:b3 Jul 12 00:08:35.678870 (udev-worker)[543]: Network interface NamePolicy= disabled on kernel command line. Jul 12 00:08:35.687179 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Jul 12 00:08:35.687254 kernel: nvme nvme0: pci function 0000:00:04.0 Jul 12 00:08:35.700318 kernel: nvme nvme0: 2/0/0 default/read/poll queues Jul 12 00:08:35.697376 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 12 00:08:35.697521 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 12 00:08:35.705647 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 12 00:08:35.705761 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 12 00:08:35.710898 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 00:08:35.721402 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jul 12 00:08:35.721467 kernel: GPT:9289727 != 16777215 Jul 12 00:08:35.721496 kernel: GPT:Alternate GPT header not at the end of the disk. Jul 12 00:08:35.713519 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 00:08:35.731853 kernel: GPT:9289727 != 16777215 Jul 12 00:08:35.731897 kernel: GPT: Use GNU Parted to correct GPT errors. Jul 12 00:08:35.731938 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jul 12 00:08:35.739802 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 00:08:35.769484 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 00:08:35.780820 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 12 00:08:35.829028 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 12 00:08:35.851474 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/nvme0n1p6 scanned by (udev-worker) (525) Jul 12 00:08:35.871459 kernel: BTRFS: device fsid 394cecf3-1fd4-438a-991e-dc2b4121da0c devid 1 transid 39 /dev/nvme0n1p3 scanned by (udev-worker) (521) Jul 12 00:08:35.951529 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Jul 12 00:08:35.996206 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Jul 12 00:08:36.015096 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Jul 12 00:08:36.032314 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Jul 12 00:08:36.038642 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Jul 12 00:08:36.054791 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jul 12 00:08:36.068917 disk-uuid[662]: Primary Header is updated. Jul 12 00:08:36.068917 disk-uuid[662]: Secondary Entries is updated. Jul 12 00:08:36.068917 disk-uuid[662]: Secondary Header is updated. Jul 12 00:08:36.077570 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jul 12 00:08:37.102455 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jul 12 00:08:37.104164 disk-uuid[663]: The operation has completed successfully. Jul 12 00:08:37.281863 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 12 00:08:37.284156 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jul 12 00:08:37.332738 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jul 12 00:08:37.345005 sh[1008]: Success Jul 12 00:08:37.370458 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Jul 12 00:08:37.487982 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jul 12 00:08:37.494242 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jul 12 00:08:37.506650 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jul 12 00:08:37.535684 kernel: BTRFS info (device dm-0): first mount of filesystem 394cecf3-1fd4-438a-991e-dc2b4121da0c Jul 12 00:08:37.535748 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Jul 12 00:08:37.537669 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jul 12 00:08:37.537705 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jul 12 00:08:37.538992 kernel: BTRFS info (device dm-0): using free space tree Jul 12 00:08:37.655469 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jul 12 00:08:37.679587 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jul 12 00:08:37.683995 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jul 12 00:08:37.695718 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jul 12 00:08:37.702506 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jul 12 00:08:37.736711 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 2ba3179f-4493-4560-9191-8e514f82bd95 Jul 12 00:08:37.737528 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Jul 12 00:08:37.737557 kernel: BTRFS info (device nvme0n1p6): using free space tree Jul 12 00:08:37.754476 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Jul 12 00:08:37.774573 systemd[1]: mnt-oem.mount: Deactivated successfully. Jul 12 00:08:37.777638 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 2ba3179f-4493-4560-9191-8e514f82bd95 Jul 12 00:08:37.784182 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jul 12 00:08:37.798717 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jul 12 00:08:37.889497 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 12 00:08:37.906709 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 12 00:08:37.953104 systemd-networkd[1201]: lo: Link UP Jul 12 00:08:37.953126 systemd-networkd[1201]: lo: Gained carrier Jul 12 00:08:37.958361 systemd-networkd[1201]: Enumeration completed Jul 12 00:08:37.958699 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 12 00:08:37.959603 systemd-networkd[1201]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 00:08:37.959610 systemd-networkd[1201]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 12 00:08:37.961175 systemd[1]: Reached target network.target - Network. Jul 12 00:08:37.970973 systemd-networkd[1201]: eth0: Link UP Jul 12 00:08:37.970981 systemd-networkd[1201]: eth0: Gained carrier Jul 12 00:08:37.970999 systemd-networkd[1201]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 00:08:37.995532 systemd-networkd[1201]: eth0: DHCPv4 address 172.31.29.62/20, gateway 172.31.16.1 acquired from 172.31.16.1 Jul 12 00:08:38.246741 ignition[1127]: Ignition 2.19.0 Jul 12 00:08:38.246770 ignition[1127]: Stage: fetch-offline Jul 12 00:08:38.250883 ignition[1127]: no configs at "/usr/lib/ignition/base.d" Jul 12 00:08:38.250930 ignition[1127]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 12 00:08:38.255684 ignition[1127]: Ignition finished successfully Jul 12 00:08:38.259938 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jul 12 00:08:38.270758 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jul 12 00:08:38.303036 ignition[1211]: Ignition 2.19.0 Jul 12 00:08:38.303660 ignition[1211]: Stage: fetch Jul 12 00:08:38.304351 ignition[1211]: no configs at "/usr/lib/ignition/base.d" Jul 12 00:08:38.304377 ignition[1211]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 12 00:08:38.304573 ignition[1211]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 12 00:08:38.333252 ignition[1211]: PUT result: OK Jul 12 00:08:38.337861 ignition[1211]: parsed url from cmdline: "" Jul 12 00:08:38.337884 ignition[1211]: no config URL provided Jul 12 00:08:38.337902 ignition[1211]: reading system config file "/usr/lib/ignition/user.ign" Jul 12 00:08:38.337930 ignition[1211]: no config at "/usr/lib/ignition/user.ign" Jul 12 00:08:38.337966 ignition[1211]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 12 00:08:38.342488 ignition[1211]: PUT result: OK Jul 12 00:08:38.342745 ignition[1211]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Jul 12 00:08:38.347601 ignition[1211]: GET result: OK Jul 12 00:08:38.348143 ignition[1211]: parsing config with SHA512: ae923c3f5731db2189a57ef775cd932153b02dd1bd640467ee7eb249ecabbbeb4fc8b1144c609d5cbf111f6dcd659fc5dcb05262c4d97752a1e227b3e4125ede Jul 12 00:08:38.361417 unknown[1211]: fetched base config from "system" Jul 12 00:08:38.361485 unknown[1211]: fetched base config from "system" Jul 12 00:08:38.361502 unknown[1211]: fetched user config from "aws" Jul 12 00:08:38.368118 ignition[1211]: fetch: fetch complete Jul 12 00:08:38.368149 ignition[1211]: fetch: fetch passed Jul 12 00:08:38.368263 ignition[1211]: Ignition finished successfully Jul 12 00:08:38.372968 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jul 12 00:08:38.385790 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jul 12 00:08:38.415049 ignition[1217]: Ignition 2.19.0 Jul 12 00:08:38.415073 ignition[1217]: Stage: kargs Jul 12 00:08:38.417193 ignition[1217]: no configs at "/usr/lib/ignition/base.d" Jul 12 00:08:38.417219 ignition[1217]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 12 00:08:38.417389 ignition[1217]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 12 00:08:38.422815 ignition[1217]: PUT result: OK Jul 12 00:08:38.430219 ignition[1217]: kargs: kargs passed Jul 12 00:08:38.430643 ignition[1217]: Ignition finished successfully Jul 12 00:08:38.436734 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jul 12 00:08:38.447823 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jul 12 00:08:38.482539 ignition[1223]: Ignition 2.19.0 Jul 12 00:08:38.482570 ignition[1223]: Stage: disks Jul 12 00:08:38.483853 ignition[1223]: no configs at "/usr/lib/ignition/base.d" Jul 12 00:08:38.483885 ignition[1223]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 12 00:08:38.484073 ignition[1223]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 12 00:08:38.486071 ignition[1223]: PUT result: OK Jul 12 00:08:38.497336 ignition[1223]: disks: disks passed Jul 12 00:08:38.497696 ignition[1223]: Ignition finished successfully Jul 12 00:08:38.504839 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jul 12 00:08:38.507516 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jul 12 00:08:38.514562 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 12 00:08:38.517804 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 12 00:08:38.523702 systemd[1]: Reached target sysinit.target - System Initialization. Jul 12 00:08:38.528170 systemd[1]: Reached target basic.target - Basic System. Jul 12 00:08:38.542450 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jul 12 00:08:38.587526 systemd-fsck[1231]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jul 12 00:08:38.594894 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jul 12 00:08:38.604624 systemd[1]: Mounting sysroot.mount - /sysroot... Jul 12 00:08:38.709549 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 44c8362f-9431-4909-bc9a-f90e514bd0e9 r/w with ordered data mode. Quota mode: none. Jul 12 00:08:38.710907 systemd[1]: Mounted sysroot.mount - /sysroot. Jul 12 00:08:38.715481 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jul 12 00:08:38.731682 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 12 00:08:38.736675 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jul 12 00:08:38.748828 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jul 12 00:08:38.748907 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 12 00:08:38.748956 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jul 12 00:08:38.760648 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jul 12 00:08:38.778469 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/nvme0n1p6 scanned by mount (1250) Jul 12 00:08:38.787480 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 2ba3179f-4493-4560-9191-8e514f82bd95 Jul 12 00:08:38.787560 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Jul 12 00:08:38.787406 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jul 12 00:08:38.793460 kernel: BTRFS info (device nvme0n1p6): using free space tree Jul 12 00:08:38.809468 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Jul 12 00:08:38.812706 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 12 00:08:39.267681 initrd-setup-root[1274]: cut: /sysroot/etc/passwd: No such file or directory Jul 12 00:08:39.287087 initrd-setup-root[1281]: cut: /sysroot/etc/group: No such file or directory Jul 12 00:08:39.297045 initrd-setup-root[1288]: cut: /sysroot/etc/shadow: No such file or directory Jul 12 00:08:39.304868 initrd-setup-root[1295]: cut: /sysroot/etc/gshadow: No such file or directory Jul 12 00:08:39.394579 systemd-networkd[1201]: eth0: Gained IPv6LL Jul 12 00:08:39.688517 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jul 12 00:08:39.700644 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jul 12 00:08:39.705013 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jul 12 00:08:39.727478 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 2ba3179f-4493-4560-9191-8e514f82bd95 Jul 12 00:08:39.728354 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jul 12 00:08:39.766494 ignition[1362]: INFO : Ignition 2.19.0 Jul 12 00:08:39.768599 ignition[1362]: INFO : Stage: mount Jul 12 00:08:39.768599 ignition[1362]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 12 00:08:39.768599 ignition[1362]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 12 00:08:39.776850 ignition[1362]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 12 00:08:39.779925 ignition[1362]: INFO : PUT result: OK Jul 12 00:08:39.787962 ignition[1362]: INFO : mount: mount passed Jul 12 00:08:39.790166 ignition[1362]: INFO : Ignition finished successfully Jul 12 00:08:39.796516 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jul 12 00:08:39.807786 systemd[1]: Starting ignition-files.service - Ignition (files)... Jul 12 00:08:39.815472 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jul 12 00:08:39.836806 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 12 00:08:39.870477 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 scanned by mount (1375) Jul 12 00:08:39.874213 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 2ba3179f-4493-4560-9191-8e514f82bd95 Jul 12 00:08:39.874263 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Jul 12 00:08:39.874290 kernel: BTRFS info (device nvme0n1p6): using free space tree Jul 12 00:08:39.880466 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Jul 12 00:08:39.884200 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 12 00:08:39.923584 ignition[1392]: INFO : Ignition 2.19.0 Jul 12 00:08:39.923584 ignition[1392]: INFO : Stage: files Jul 12 00:08:39.927586 ignition[1392]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 12 00:08:39.927586 ignition[1392]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 12 00:08:39.927586 ignition[1392]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 12 00:08:39.935015 ignition[1392]: INFO : PUT result: OK Jul 12 00:08:39.939688 ignition[1392]: DEBUG : files: compiled without relabeling support, skipping Jul 12 00:08:39.947765 ignition[1392]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 12 00:08:39.947765 ignition[1392]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 12 00:08:39.974082 ignition[1392]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 12 00:08:39.977293 ignition[1392]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 12 00:08:39.980811 unknown[1392]: wrote ssh authorized keys file for user: core Jul 12 00:08:39.983295 ignition[1392]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 12 00:08:39.986177 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Jul 12 00:08:39.986177 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Jul 12 00:08:39.986177 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jul 12 00:08:39.986177 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Jul 12 00:08:40.075975 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Jul 12 00:08:40.210583 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jul 12 00:08:40.210583 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 12 00:08:40.210583 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Jul 12 00:08:40.556077 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Jul 12 00:08:40.708669 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 12 00:08:40.714557 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 12 00:08:40.745035 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 12 00:08:40.745035 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 12 00:08:40.745035 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 00:08:40.745035 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 00:08:40.745035 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 00:08:40.745035 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-arm64.raw: attempt #1 Jul 12 00:08:41.021172 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK Jul 12 00:08:41.389542 ignition[1392]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Jul 12 00:08:41.389542 ignition[1392]: INFO : files: op(d): [started] processing unit "containerd.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(d): [finished] processing unit "containerd.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(f): [started] processing unit "prepare-helm.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 12 00:08:41.397127 ignition[1392]: INFO : files: files passed Jul 12 00:08:41.397127 ignition[1392]: INFO : Ignition finished successfully Jul 12 00:08:41.429241 systemd[1]: Finished ignition-files.service - Ignition (files). Jul 12 00:08:41.450784 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jul 12 00:08:41.458630 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jul 12 00:08:41.470022 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 12 00:08:41.471521 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jul 12 00:08:41.496339 initrd-setup-root-after-ignition[1420]: grep: Jul 12 00:08:41.496339 initrd-setup-root-after-ignition[1424]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 12 00:08:41.501623 initrd-setup-root-after-ignition[1420]: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 12 00:08:41.501623 initrd-setup-root-after-ignition[1420]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jul 12 00:08:41.508075 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 12 00:08:41.511192 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jul 12 00:08:41.523797 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jul 12 00:08:41.576073 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 12 00:08:41.576307 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jul 12 00:08:41.583962 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jul 12 00:08:41.588116 systemd[1]: Reached target initrd.target - Initrd Default Target. Jul 12 00:08:41.588956 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jul 12 00:08:41.597780 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jul 12 00:08:41.633321 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 12 00:08:41.650487 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jul 12 00:08:41.673451 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jul 12 00:08:41.676355 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 12 00:08:41.683768 systemd[1]: Stopped target timers.target - Timer Units. Jul 12 00:08:41.687805 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 12 00:08:41.688073 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 12 00:08:41.695510 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jul 12 00:08:41.695824 systemd[1]: Stopped target basic.target - Basic System. Jul 12 00:08:41.701994 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jul 12 00:08:41.706672 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jul 12 00:08:41.711403 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jul 12 00:08:41.718718 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jul 12 00:08:41.721087 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jul 12 00:08:41.724295 systemd[1]: Stopped target sysinit.target - System Initialization. Jul 12 00:08:41.727151 systemd[1]: Stopped target local-fs.target - Local File Systems. Jul 12 00:08:41.731851 systemd[1]: Stopped target swap.target - Swaps. Jul 12 00:08:41.736391 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 12 00:08:41.736666 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jul 12 00:08:41.744937 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jul 12 00:08:41.754347 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 12 00:08:41.757029 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jul 12 00:08:41.759571 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 12 00:08:41.767669 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 12 00:08:41.768459 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jul 12 00:08:41.774773 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 12 00:08:41.775132 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 12 00:08:41.782843 systemd[1]: ignition-files.service: Deactivated successfully. Jul 12 00:08:41.783055 systemd[1]: Stopped ignition-files.service - Ignition (files). Jul 12 00:08:41.794849 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jul 12 00:08:41.797150 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 12 00:08:41.797537 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jul 12 00:08:41.812923 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jul 12 00:08:41.819213 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 12 00:08:41.821286 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jul 12 00:08:41.827303 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 12 00:08:41.830114 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jul 12 00:08:41.849486 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 12 00:08:41.854711 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jul 12 00:08:41.863641 ignition[1444]: INFO : Ignition 2.19.0 Jul 12 00:08:41.868623 ignition[1444]: INFO : Stage: umount Jul 12 00:08:41.868623 ignition[1444]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 12 00:08:41.868623 ignition[1444]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jul 12 00:08:41.868623 ignition[1444]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jul 12 00:08:41.887719 ignition[1444]: INFO : PUT result: OK Jul 12 00:08:41.890424 ignition[1444]: INFO : umount: umount passed Jul 12 00:08:41.893844 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 12 00:08:41.895939 ignition[1444]: INFO : Ignition finished successfully Jul 12 00:08:41.899902 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 12 00:08:41.900733 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jul 12 00:08:41.912479 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 12 00:08:41.913513 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jul 12 00:08:41.919468 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 12 00:08:41.919571 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jul 12 00:08:41.925845 systemd[1]: ignition-fetch.service: Deactivated successfully. Jul 12 00:08:41.925959 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jul 12 00:08:41.928216 systemd[1]: Stopped target network.target - Network. Jul 12 00:08:41.930062 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 12 00:08:41.930157 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jul 12 00:08:41.933374 systemd[1]: Stopped target paths.target - Path Units. Jul 12 00:08:41.936972 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 12 00:08:41.937078 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 12 00:08:41.941796 systemd[1]: Stopped target slices.target - Slice Units. Jul 12 00:08:41.945581 systemd[1]: Stopped target sockets.target - Socket Units. Jul 12 00:08:41.947857 systemd[1]: iscsid.socket: Deactivated successfully. Jul 12 00:08:41.947937 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jul 12 00:08:41.954639 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 12 00:08:41.954713 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 12 00:08:41.958280 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 12 00:08:41.958384 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jul 12 00:08:41.960641 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jul 12 00:08:41.960721 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jul 12 00:08:41.963309 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jul 12 00:08:41.965689 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jul 12 00:08:41.969083 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 12 00:08:41.969260 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jul 12 00:08:41.972934 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 12 00:08:41.973095 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jul 12 00:08:41.981573 systemd-networkd[1201]: eth0: DHCPv6 lease lost Jul 12 00:08:41.987132 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 12 00:08:41.987384 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jul 12 00:08:41.993977 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 12 00:08:41.994054 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jul 12 00:08:42.013210 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jul 12 00:08:42.019223 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 12 00:08:42.024912 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 12 00:08:42.031011 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 12 00:08:42.034705 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 12 00:08:42.035093 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jul 12 00:08:42.078740 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 12 00:08:42.081002 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 12 00:08:42.085756 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 12 00:08:42.085862 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jul 12 00:08:42.089812 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jul 12 00:08:42.089918 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 12 00:08:42.107134 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 12 00:08:42.107973 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 12 00:08:42.116383 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 12 00:08:42.116593 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jul 12 00:08:42.119624 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 12 00:08:42.119697 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jul 12 00:08:42.122516 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 12 00:08:42.122613 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jul 12 00:08:42.125495 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 12 00:08:42.125611 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jul 12 00:08:42.126102 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 12 00:08:42.126185 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 12 00:08:42.153803 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jul 12 00:08:42.156487 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jul 12 00:08:42.156617 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 12 00:08:42.159417 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 12 00:08:42.159522 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 00:08:42.162790 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 12 00:08:42.163015 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jul 12 00:08:42.203208 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 12 00:08:42.203629 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jul 12 00:08:42.211222 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jul 12 00:08:42.226283 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jul 12 00:08:42.249537 systemd[1]: Switching root. Jul 12 00:08:42.290151 systemd-journald[251]: Journal stopped Jul 12 00:08:44.428745 systemd-journald[251]: Received SIGTERM from PID 1 (systemd). Jul 12 00:08:44.428881 kernel: SELinux: policy capability network_peer_controls=1 Jul 12 00:08:44.428926 kernel: SELinux: policy capability open_perms=1 Jul 12 00:08:44.428959 kernel: SELinux: policy capability extended_socket_class=1 Jul 12 00:08:44.428989 kernel: SELinux: policy capability always_check_network=0 Jul 12 00:08:44.429020 kernel: SELinux: policy capability cgroup_seclabel=1 Jul 12 00:08:44.429050 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jul 12 00:08:44.429087 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jul 12 00:08:44.429117 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jul 12 00:08:44.429144 kernel: audit: type=1403 audit(1752278922.814:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jul 12 00:08:44.429183 systemd[1]: Successfully loaded SELinux policy in 49.191ms. Jul 12 00:08:44.429234 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 22.820ms. Jul 12 00:08:44.429269 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jul 12 00:08:44.429300 systemd[1]: Detected virtualization amazon. Jul 12 00:08:44.429331 systemd[1]: Detected architecture arm64. Jul 12 00:08:44.429364 systemd[1]: Detected first boot. Jul 12 00:08:44.429399 systemd[1]: Initializing machine ID from VM UUID. Jul 12 00:08:44.429448 zram_generator::config[1503]: No configuration found. Jul 12 00:08:44.429491 systemd[1]: Populated /etc with preset unit settings. Jul 12 00:08:44.429525 systemd[1]: Queued start job for default target multi-user.target. Jul 12 00:08:44.429557 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Jul 12 00:08:44.429590 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jul 12 00:08:44.429622 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jul 12 00:08:44.429651 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jul 12 00:08:44.429687 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jul 12 00:08:44.429719 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jul 12 00:08:44.429749 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jul 12 00:08:44.429779 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jul 12 00:08:44.429810 systemd[1]: Created slice user.slice - User and Session Slice. Jul 12 00:08:44.429841 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 12 00:08:44.429871 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 12 00:08:44.429900 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jul 12 00:08:44.429934 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jul 12 00:08:44.429964 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jul 12 00:08:44.429996 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 12 00:08:44.430027 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jul 12 00:08:44.430059 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 12 00:08:44.430091 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jul 12 00:08:44.430122 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 12 00:08:44.430152 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 12 00:08:44.430182 systemd[1]: Reached target slices.target - Slice Units. Jul 12 00:08:44.430217 systemd[1]: Reached target swap.target - Swaps. Jul 12 00:08:44.430248 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jul 12 00:08:44.430283 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jul 12 00:08:44.430364 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 12 00:08:44.430396 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jul 12 00:08:44.432448 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 12 00:08:44.432500 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 12 00:08:44.432535 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 12 00:08:44.432565 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jul 12 00:08:44.432602 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jul 12 00:08:44.432634 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jul 12 00:08:44.432666 systemd[1]: Mounting media.mount - External Media Directory... Jul 12 00:08:44.432698 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jul 12 00:08:44.432729 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jul 12 00:08:44.432762 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jul 12 00:08:44.432791 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jul 12 00:08:44.432826 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 00:08:44.432861 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 12 00:08:44.432891 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jul 12 00:08:44.432922 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 00:08:44.432951 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 12 00:08:44.432981 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 00:08:44.433013 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jul 12 00:08:44.433046 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 00:08:44.433079 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jul 12 00:08:44.433111 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Jul 12 00:08:44.433147 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Jul 12 00:08:44.433176 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 12 00:08:44.433206 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 12 00:08:44.433234 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 12 00:08:44.433263 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jul 12 00:08:44.433291 kernel: loop: module loaded Jul 12 00:08:44.433322 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 12 00:08:44.433356 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jul 12 00:08:44.433387 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jul 12 00:08:44.433421 systemd[1]: Mounted media.mount - External Media Directory. Jul 12 00:08:44.433477 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jul 12 00:08:44.433510 kernel: fuse: init (API version 7.39) Jul 12 00:08:44.433539 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jul 12 00:08:44.433568 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jul 12 00:08:44.433607 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 12 00:08:44.433637 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jul 12 00:08:44.433666 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jul 12 00:08:44.433694 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 00:08:44.433733 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 00:08:44.433763 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 00:08:44.433792 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 00:08:44.433870 systemd-journald[1599]: Collecting audit messages is disabled. Jul 12 00:08:44.433932 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jul 12 00:08:44.433964 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jul 12 00:08:44.433995 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 00:08:44.434030 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 00:08:44.434058 systemd-journald[1599]: Journal started Jul 12 00:08:44.434105 systemd-journald[1599]: Runtime Journal (/run/log/journal/ec2a064f90693ef95d6132e4dfbd219a) is 8.0M, max 75.3M, 67.3M free. Jul 12 00:08:44.440477 systemd[1]: Started systemd-journald.service - Journal Service. Jul 12 00:08:44.452854 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jul 12 00:08:44.456577 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 12 00:08:44.467351 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 12 00:08:44.500760 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 12 00:08:44.517388 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jul 12 00:08:44.528494 kernel: ACPI: bus type drm_connector registered Jul 12 00:08:44.527656 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jul 12 00:08:44.531709 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jul 12 00:08:44.552677 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jul 12 00:08:44.565636 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jul 12 00:08:44.570616 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 12 00:08:44.582752 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jul 12 00:08:44.585666 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 12 00:08:44.604690 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 12 00:08:44.617806 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 12 00:08:44.628996 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jul 12 00:08:44.633240 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 12 00:08:44.636814 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 12 00:08:44.640577 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jul 12 00:08:44.655892 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jul 12 00:08:44.659887 systemd-journald[1599]: Time spent on flushing to /var/log/journal/ec2a064f90693ef95d6132e4dfbd219a is 75.310ms for 897 entries. Jul 12 00:08:44.659887 systemd-journald[1599]: System Journal (/var/log/journal/ec2a064f90693ef95d6132e4dfbd219a) is 8.0M, max 195.6M, 187.6M free. Jul 12 00:08:44.743519 systemd-journald[1599]: Received client request to flush runtime journal. Jul 12 00:08:44.672277 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jul 12 00:08:44.678605 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jul 12 00:08:44.731270 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 12 00:08:44.751416 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jul 12 00:08:44.781472 systemd-tmpfiles[1653]: ACLs are not supported, ignoring. Jul 12 00:08:44.781506 systemd-tmpfiles[1653]: ACLs are not supported, ignoring. Jul 12 00:08:44.791192 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 12 00:08:44.805883 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jul 12 00:08:44.809557 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 12 00:08:44.822718 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jul 12 00:08:44.852762 udevadm[1671]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jul 12 00:08:44.901939 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jul 12 00:08:44.914843 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 12 00:08:44.953723 systemd-tmpfiles[1678]: ACLs are not supported, ignoring. Jul 12 00:08:44.954273 systemd-tmpfiles[1678]: ACLs are not supported, ignoring. Jul 12 00:08:44.963382 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 12 00:08:45.629544 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jul 12 00:08:45.642818 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 12 00:08:45.695128 systemd-udevd[1684]: Using default interface naming scheme 'v255'. Jul 12 00:08:45.734267 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 12 00:08:45.770808 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 12 00:08:45.797714 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jul 12 00:08:45.871822 (udev-worker)[1687]: Network interface NamePolicy= disabled on kernel command line. Jul 12 00:08:45.898563 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Jul 12 00:08:45.980113 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jul 12 00:08:46.135281 systemd-networkd[1694]: lo: Link UP Jul 12 00:08:46.135301 systemd-networkd[1694]: lo: Gained carrier Jul 12 00:08:46.138952 systemd-networkd[1694]: Enumeration completed Jul 12 00:08:46.139182 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 12 00:08:46.143068 systemd-networkd[1694]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 00:08:46.143092 systemd-networkd[1694]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 12 00:08:46.155947 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 39 scanned by (udev-worker) (1685) Jul 12 00:08:46.156612 systemd-networkd[1694]: eth0: Link UP Jul 12 00:08:46.157199 systemd-networkd[1694]: eth0: Gained carrier Jul 12 00:08:46.157252 systemd-networkd[1694]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jul 12 00:08:46.162988 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jul 12 00:08:46.208939 systemd-networkd[1694]: eth0: DHCPv4 address 172.31.29.62/20, gateway 172.31.16.1 acquired from 172.31.16.1 Jul 12 00:08:46.390218 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Jul 12 00:08:46.406766 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 12 00:08:46.410170 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jul 12 00:08:46.429786 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jul 12 00:08:46.467746 lvm[1809]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 12 00:08:46.509202 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jul 12 00:08:46.515043 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 12 00:08:46.525994 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jul 12 00:08:46.538311 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 12 00:08:46.546514 lvm[1814]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 12 00:08:46.581451 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jul 12 00:08:46.585461 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 12 00:08:46.588303 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jul 12 00:08:46.588366 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 12 00:08:46.590705 systemd[1]: Reached target machines.target - Containers. Jul 12 00:08:46.595542 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jul 12 00:08:46.607758 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jul 12 00:08:46.612710 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jul 12 00:08:46.615174 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 00:08:46.618082 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jul 12 00:08:46.636694 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jul 12 00:08:46.649726 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jul 12 00:08:46.657217 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jul 12 00:08:46.684042 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jul 12 00:08:46.685567 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jul 12 00:08:46.692797 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jul 12 00:08:46.729515 kernel: loop0: detected capacity change from 0 to 114432 Jul 12 00:08:46.759485 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jul 12 00:08:46.789836 kernel: loop1: detected capacity change from 0 to 114328 Jul 12 00:08:46.832547 kernel: loop2: detected capacity change from 0 to 203944 Jul 12 00:08:46.976022 kernel: loop3: detected capacity change from 0 to 52536 Jul 12 00:08:47.091506 kernel: loop4: detected capacity change from 0 to 114432 Jul 12 00:08:47.114529 kernel: loop5: detected capacity change from 0 to 114328 Jul 12 00:08:47.137482 kernel: loop6: detected capacity change from 0 to 203944 Jul 12 00:08:47.170526 kernel: loop7: detected capacity change from 0 to 52536 Jul 12 00:08:47.183304 (sd-merge)[1838]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Jul 12 00:08:47.184331 (sd-merge)[1838]: Merged extensions into '/usr'. Jul 12 00:08:47.191336 systemd[1]: Reloading requested from client PID 1825 ('systemd-sysext') (unit systemd-sysext.service)... Jul 12 00:08:47.191369 systemd[1]: Reloading... Jul 12 00:08:47.301894 ldconfig[1821]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jul 12 00:08:47.342463 zram_generator::config[1866]: No configuration found. Jul 12 00:08:47.594550 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 00:08:47.747059 systemd[1]: Reloading finished in 554 ms. Jul 12 00:08:47.770790 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jul 12 00:08:47.779158 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jul 12 00:08:47.795867 systemd[1]: Starting ensure-sysext.service... Jul 12 00:08:47.803168 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 12 00:08:47.817788 systemd[1]: Reloading requested from client PID 1925 ('systemctl') (unit ensure-sysext.service)... Jul 12 00:08:47.817815 systemd[1]: Reloading... Jul 12 00:08:47.843171 systemd-networkd[1694]: eth0: Gained IPv6LL Jul 12 00:08:47.870525 systemd-tmpfiles[1926]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jul 12 00:08:47.871189 systemd-tmpfiles[1926]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jul 12 00:08:47.873923 systemd-tmpfiles[1926]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jul 12 00:08:47.874582 systemd-tmpfiles[1926]: ACLs are not supported, ignoring. Jul 12 00:08:47.874716 systemd-tmpfiles[1926]: ACLs are not supported, ignoring. Jul 12 00:08:47.881125 systemd-tmpfiles[1926]: Detected autofs mount point /boot during canonicalization of boot. Jul 12 00:08:47.881152 systemd-tmpfiles[1926]: Skipping /boot Jul 12 00:08:47.904589 systemd-tmpfiles[1926]: Detected autofs mount point /boot during canonicalization of boot. Jul 12 00:08:47.904615 systemd-tmpfiles[1926]: Skipping /boot Jul 12 00:08:48.004491 zram_generator::config[1958]: No configuration found. Jul 12 00:08:48.234411 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 00:08:48.388719 systemd[1]: Reloading finished in 569 ms. Jul 12 00:08:48.416167 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jul 12 00:08:48.426732 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 12 00:08:48.447004 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jul 12 00:08:48.452472 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jul 12 00:08:48.461735 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jul 12 00:08:48.479808 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 12 00:08:48.494743 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jul 12 00:08:48.520212 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 00:08:48.533707 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 00:08:48.546318 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 00:08:48.566051 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 00:08:48.569989 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 00:08:48.575058 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 00:08:48.578685 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 00:08:48.592879 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 00:08:48.593258 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 00:08:48.615245 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 00:08:48.615799 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 00:08:48.628270 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jul 12 00:08:48.637009 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jul 12 00:08:48.650057 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 00:08:48.658029 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 00:08:48.667910 augenrules[2052]: No rules Jul 12 00:08:48.668532 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 00:08:48.687902 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 00:08:48.694757 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 00:08:48.703307 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jul 12 00:08:48.717933 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jul 12 00:08:48.725254 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 00:08:48.729926 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 00:08:48.734688 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 00:08:48.743625 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 00:08:48.747692 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 00:08:48.749089 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 00:08:48.775036 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jul 12 00:08:48.784099 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 12 00:08:48.798014 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 12 00:08:48.810626 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 12 00:08:48.826901 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 12 00:08:48.836962 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 12 00:08:48.837339 systemd[1]: Reached target time-set.target - System Time Set. Jul 12 00:08:48.851036 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jul 12 00:08:48.857392 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 12 00:08:48.858924 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 12 00:08:48.864765 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 12 00:08:48.865130 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 12 00:08:48.872827 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 12 00:08:48.873190 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 12 00:08:48.877018 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 12 00:08:48.877376 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 12 00:08:48.884842 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jul 12 00:08:48.906077 systemd[1]: Finished ensure-sysext.service. Jul 12 00:08:48.920914 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 12 00:08:48.921058 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 12 00:08:48.921104 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 12 00:08:48.939717 systemd-resolved[2025]: Positive Trust Anchors: Jul 12 00:08:48.939756 systemd-resolved[2025]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 12 00:08:48.939819 systemd-resolved[2025]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 12 00:08:48.953105 systemd-resolved[2025]: Defaulting to hostname 'linux'. Jul 12 00:08:48.956667 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 12 00:08:48.959250 systemd[1]: Reached target network.target - Network. Jul 12 00:08:48.961254 systemd[1]: Reached target network-online.target - Network is Online. Jul 12 00:08:48.963654 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 12 00:08:48.966194 systemd[1]: Reached target sysinit.target - System Initialization. Jul 12 00:08:48.968621 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jul 12 00:08:48.971282 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jul 12 00:08:48.974286 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jul 12 00:08:48.976819 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jul 12 00:08:48.979544 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jul 12 00:08:48.982262 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jul 12 00:08:48.982333 systemd[1]: Reached target paths.target - Path Units. Jul 12 00:08:48.984457 systemd[1]: Reached target timers.target - Timer Units. Jul 12 00:08:48.987309 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jul 12 00:08:48.992730 systemd[1]: Starting docker.socket - Docker Socket for the API... Jul 12 00:08:48.998006 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jul 12 00:08:49.003984 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jul 12 00:08:49.006799 systemd[1]: Reached target sockets.target - Socket Units. Jul 12 00:08:49.009614 systemd[1]: Reached target basic.target - Basic System. Jul 12 00:08:49.012205 systemd[1]: System is tainted: cgroupsv1 Jul 12 00:08:49.012316 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jul 12 00:08:49.012372 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jul 12 00:08:49.021642 systemd[1]: Starting containerd.service - containerd container runtime... Jul 12 00:08:49.033570 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jul 12 00:08:49.040589 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jul 12 00:08:49.046696 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jul 12 00:08:49.073986 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jul 12 00:08:49.083637 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jul 12 00:08:49.090792 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:08:49.101779 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jul 12 00:08:49.128518 systemd[1]: Started ntpd.service - Network Time Service. Jul 12 00:08:49.137153 jq[2094]: false Jul 12 00:08:49.158703 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jul 12 00:08:49.175046 extend-filesystems[2095]: Found loop4 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found loop5 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found loop6 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found loop7 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1p1 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1p2 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1p3 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found usr Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1p4 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1p6 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1p7 Jul 12 00:08:49.186573 extend-filesystems[2095]: Found nvme0n1p9 Jul 12 00:08:49.186573 extend-filesystems[2095]: Checking size of /dev/nvme0n1p9 Jul 12 00:08:49.176611 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jul 12 00:08:49.192648 dbus-daemon[2092]: [system] SELinux support is enabled Jul 12 00:08:49.193629 systemd[1]: Starting setup-oem.service - Setup OEM... Jul 12 00:08:49.226076 dbus-daemon[2092]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1694 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Jul 12 00:08:49.226794 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jul 12 00:08:49.257168 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jul 12 00:08:49.286819 systemd[1]: Starting systemd-logind.service - User Login Management... Jul 12 00:08:49.290859 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jul 12 00:08:49.300685 systemd[1]: Starting update-engine.service - Update Engine... Jul 12 00:08:49.311345 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jul 12 00:08:49.316055 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jul 12 00:08:49.343479 extend-filesystems[2095]: Resized partition /dev/nvme0n1p9 Jul 12 00:08:49.351931 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jul 12 00:08:49.352494 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jul 12 00:08:49.371547 extend-filesystems[2129]: resize2fs 1.47.1 (20-May-2024) Jul 12 00:08:49.447733 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Jul 12 00:08:49.447926 coreos-metadata[2091]: Jul 12 00:08:49.444 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: ntpd 4.2.8p17@1.4004-o Fri Jul 11 22:05:17 UTC 2025 (1): Starting Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: ---------------------------------------------------- Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: ntp-4 is maintained by Network Time Foundation, Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: corporation. Support and training for ntp-4 are Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: available at https://www.nwtime.org/support Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: ---------------------------------------------------- Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: proto: precision = 0.108 usec (-23) Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: basedate set to 2025-06-29 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: gps base set to 2025-06-29 (week 2373) Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Listen and drop on 0 v6wildcard [::]:123 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Listen normally on 2 lo 127.0.0.1:123 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Listen normally on 3 eth0 172.31.29.62:123 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Listen normally on 4 lo [::1]:123 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Listen normally on 5 eth0 [fe80::41d:9cff:fe3a:25b3%2]:123 Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: Listening on routing socket on fd #22 for interface updates Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 12 00:08:49.486165 ntpd[2101]: 12 Jul 00:08:49 ntpd[2101]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 12 00:08:49.498012 update_engine[2118]: I20250712 00:08:49.403631 2118 main.cc:92] Flatcar Update Engine starting Jul 12 00:08:49.498012 update_engine[2118]: I20250712 00:08:49.409384 2118 update_check_scheduler.cc:74] Next update check in 9m16s Jul 12 00:08:49.424162 ntpd[2101]: ntpd 4.2.8p17@1.4004-o Fri Jul 11 22:05:17 UTC 2025 (1): Starting Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.448 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.452 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.453 INFO Fetch failed with 404: resource not found Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.453 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetch successful Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Jul 12 00:08:49.524180 coreos-metadata[2091]: Jul 12 00:08:49.458 INFO Fetch successful Jul 12 00:08:49.465402 systemd[1]: motdgen.service: Deactivated successfully. Jul 12 00:08:49.424219 ntpd[2101]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Jul 12 00:08:49.471015 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jul 12 00:08:49.424243 ntpd[2101]: ---------------------------------------------------- Jul 12 00:08:49.481723 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jul 12 00:08:49.529937 jq[2122]: true Jul 12 00:08:49.424263 ntpd[2101]: ntp-4 is maintained by Network Time Foundation, Jul 12 00:08:49.482247 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jul 12 00:08:49.424283 ntpd[2101]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Jul 12 00:08:49.424303 ntpd[2101]: corporation. Support and training for ntp-4 are Jul 12 00:08:49.424327 ntpd[2101]: available at https://www.nwtime.org/support Jul 12 00:08:49.424346 ntpd[2101]: ---------------------------------------------------- Jul 12 00:08:49.429366 ntpd[2101]: proto: precision = 0.108 usec (-23) Jul 12 00:08:49.430344 ntpd[2101]: basedate set to 2025-06-29 Jul 12 00:08:49.430378 ntpd[2101]: gps base set to 2025-06-29 (week 2373) Jul 12 00:08:49.433129 ntpd[2101]: Listen and drop on 0 v6wildcard [::]:123 Jul 12 00:08:49.433204 ntpd[2101]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Jul 12 00:08:49.433498 ntpd[2101]: Listen normally on 2 lo 127.0.0.1:123 Jul 12 00:08:49.433565 ntpd[2101]: Listen normally on 3 eth0 172.31.29.62:123 Jul 12 00:08:49.433632 ntpd[2101]: Listen normally on 4 lo [::1]:123 Jul 12 00:08:49.433705 ntpd[2101]: Listen normally on 5 eth0 [fe80::41d:9cff:fe3a:25b3%2]:123 Jul 12 00:08:49.433768 ntpd[2101]: Listening on routing socket on fd #22 for interface updates Jul 12 00:08:49.441927 ntpd[2101]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 12 00:08:49.576046 systemd[1]: Started update-engine.service - Update Engine. Jul 12 00:08:49.442007 ntpd[2101]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jul 12 00:08:49.585573 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jul 12 00:08:49.581883 dbus-daemon[2092]: [system] Successfully activated service 'org.freedesktop.systemd1' Jul 12 00:08:49.585666 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jul 12 00:08:49.588765 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jul 12 00:08:49.588809 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jul 12 00:08:49.594640 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jul 12 00:08:49.613953 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jul 12 00:08:49.620244 (ntainerd)[2160]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jul 12 00:08:49.620699 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jul 12 00:08:49.642509 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Jul 12 00:08:49.658763 jq[2159]: true Jul 12 00:08:49.679535 extend-filesystems[2129]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Jul 12 00:08:49.679535 extend-filesystems[2129]: old_desc_blocks = 1, new_desc_blocks = 1 Jul 12 00:08:49.679535 extend-filesystems[2129]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Jul 12 00:08:49.695771 extend-filesystems[2095]: Resized filesystem in /dev/nvme0n1p9 Jul 12 00:08:49.710087 systemd[1]: extend-filesystems.service: Deactivated successfully. Jul 12 00:08:49.710683 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jul 12 00:08:49.714602 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jul 12 00:08:49.757035 tar[2135]: linux-arm64/helm Jul 12 00:08:49.772530 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jul 12 00:08:49.789555 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Jul 12 00:08:49.856618 systemd[1]: Finished setup-oem.service - Setup OEM. Jul 12 00:08:49.867798 systemd-logind[2117]: Watching system buttons on /dev/input/event0 (Power Button) Jul 12 00:08:49.867842 systemd-logind[2117]: Watching system buttons on /dev/input/event1 (Sleep Button) Jul 12 00:08:49.869113 systemd-logind[2117]: New seat seat0. Jul 12 00:08:49.943602 locksmithd[2169]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jul 12 00:08:49.964822 bash[2235]: Updated "/home/core/.ssh/authorized_keys" Jul 12 00:08:50.095921 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Jul 12 00:08:50.167273 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 39 scanned by (udev-worker) (2188) Jul 12 00:08:50.162959 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jul 12 00:08:50.168951 systemd[1]: Started systemd-logind.service - User Login Management. Jul 12 00:08:50.217879 systemd[1]: Starting sshkeys.service... Jul 12 00:08:50.363092 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jul 12 00:08:50.378136 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jul 12 00:08:50.507467 amazon-ssm-agent[2247]: Initializing new seelog logger Jul 12 00:08:50.507467 amazon-ssm-agent[2247]: New Seelog Logger Creation Complete Jul 12 00:08:50.507467 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.507467 amazon-ssm-agent[2247]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.512027 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 processing appconfig overrides Jul 12 00:08:50.517460 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.517460 amazon-ssm-agent[2247]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.517460 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 processing appconfig overrides Jul 12 00:08:50.519122 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.519122 amazon-ssm-agent[2247]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.519262 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 processing appconfig overrides Jul 12 00:08:50.525455 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO Proxy environment variables: Jul 12 00:08:50.538471 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.538471 amazon-ssm-agent[2247]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jul 12 00:08:50.538665 amazon-ssm-agent[2247]: 2025/07/12 00:08:50 processing appconfig overrides Jul 12 00:08:50.628708 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO https_proxy: Jul 12 00:08:50.657088 dbus-daemon[2092]: [system] Successfully activated service 'org.freedesktop.hostname1' Jul 12 00:08:50.659517 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Jul 12 00:08:50.672805 dbus-daemon[2092]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.7' (uid=0 pid=2189 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Jul 12 00:08:50.691220 systemd[1]: Starting polkit.service - Authorization Manager... Jul 12 00:08:50.730395 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO http_proxy: Jul 12 00:08:50.756649 polkitd[2324]: Started polkitd version 121 Jul 12 00:08:50.788860 coreos-metadata[2297]: Jul 12 00:08:50.787 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Jul 12 00:08:50.791319 coreos-metadata[2297]: Jul 12 00:08:50.789 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Jul 12 00:08:50.793756 coreos-metadata[2297]: Jul 12 00:08:50.793 INFO Fetch successful Jul 12 00:08:50.793899 coreos-metadata[2297]: Jul 12 00:08:50.793 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Jul 12 00:08:50.794581 polkitd[2324]: Loading rules from directory /etc/polkit-1/rules.d Jul 12 00:08:50.794811 polkitd[2324]: Loading rules from directory /usr/share/polkit-1/rules.d Jul 12 00:08:50.797461 coreos-metadata[2297]: Jul 12 00:08:50.797 INFO Fetch successful Jul 12 00:08:50.800064 polkitd[2324]: Finished loading, compiling and executing 2 rules Jul 12 00:08:50.805989 unknown[2297]: wrote ssh authorized keys file for user: core Jul 12 00:08:50.816358 dbus-daemon[2092]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Jul 12 00:08:50.817847 systemd[1]: Started polkit.service - Authorization Manager. Jul 12 00:08:50.824422 polkitd[2324]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Jul 12 00:08:50.838685 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO no_proxy: Jul 12 00:08:50.873469 update-ssh-keys[2341]: Updated "/home/core/.ssh/authorized_keys" Jul 12 00:08:50.876110 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jul 12 00:08:50.882570 containerd[2160]: time="2025-07-12T00:08:50.880495058Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Jul 12 00:08:50.897196 systemd[1]: Finished sshkeys.service. Jul 12 00:08:50.937053 systemd-hostnamed[2189]: Hostname set to (transient) Jul 12 00:08:50.937766 systemd-resolved[2025]: System hostname changed to 'ip-172-31-29-62'. Jul 12 00:08:50.940757 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO Checking if agent identity type OnPrem can be assumed Jul 12 00:08:51.043906 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO Checking if agent identity type EC2 can be assumed Jul 12 00:08:51.057136 containerd[2160]: time="2025-07-12T00:08:51.056783243Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jul 12 00:08:51.065287 containerd[2160]: time="2025-07-12T00:08:51.064580471Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.96-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jul 12 00:08:51.065287 containerd[2160]: time="2025-07-12T00:08:51.064645811Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jul 12 00:08:51.065287 containerd[2160]: time="2025-07-12T00:08:51.064679495Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jul 12 00:08:51.065287 containerd[2160]: time="2025-07-12T00:08:51.064974623Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jul 12 00:08:51.065287 containerd[2160]: time="2025-07-12T00:08:51.065007479Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jul 12 00:08:51.065287 containerd[2160]: time="2025-07-12T00:08:51.065123363Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jul 12 00:08:51.065287 containerd[2160]: time="2025-07-12T00:08:51.065152895Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jul 12 00:08:51.067022 containerd[2160]: time="2025-07-12T00:08:51.066966755Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 12 00:08:51.069992 containerd[2160]: time="2025-07-12T00:08:51.068466911Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jul 12 00:08:51.069992 containerd[2160]: time="2025-07-12T00:08:51.068512535Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jul 12 00:08:51.069992 containerd[2160]: time="2025-07-12T00:08:51.068540219Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jul 12 00:08:51.069992 containerd[2160]: time="2025-07-12T00:08:51.068751287Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jul 12 00:08:51.069992 containerd[2160]: time="2025-07-12T00:08:51.069151871Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jul 12 00:08:51.072053 containerd[2160]: time="2025-07-12T00:08:51.071606123Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 12 00:08:51.072053 containerd[2160]: time="2025-07-12T00:08:51.071659115Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jul 12 00:08:51.072053 containerd[2160]: time="2025-07-12T00:08:51.071892467Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jul 12 00:08:51.072053 containerd[2160]: time="2025-07-12T00:08:51.071992199Z" level=info msg="metadata content store policy set" policy=shared Jul 12 00:08:51.080695 containerd[2160]: time="2025-07-12T00:08:51.080294579Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jul 12 00:08:51.080695 containerd[2160]: time="2025-07-12T00:08:51.080396615Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jul 12 00:08:51.080695 containerd[2160]: time="2025-07-12T00:08:51.080530895Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jul 12 00:08:51.080695 containerd[2160]: time="2025-07-12T00:08:51.080569847Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jul 12 00:08:51.080695 containerd[2160]: time="2025-07-12T00:08:51.080629043Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jul 12 00:08:51.084241 containerd[2160]: time="2025-07-12T00:08:51.083584127Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085344239Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085659983Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085707587Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085751651Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085787387Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085830935Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085870931Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085913471Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.085974635Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.086016239Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.086056655Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.086090771Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.086142731Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.087482 containerd[2160]: time="2025-07-12T00:08:51.086184539Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.088108 containerd[2160]: time="2025-07-12T00:08:51.086224967Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.088108 containerd[2160]: time="2025-07-12T00:08:51.086288195Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.088108 containerd[2160]: time="2025-07-12T00:08:51.086332943Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.088108 containerd[2160]: time="2025-07-12T00:08:51.086379791Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.101453 containerd[2160]: time="2025-07-12T00:08:51.086411963Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.103973 containerd[2160]: time="2025-07-12T00:08:51.101391287Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.103973 containerd[2160]: time="2025-07-12T00:08:51.101747723Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.103973 containerd[2160]: time="2025-07-12T00:08:51.103719323Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.103973 containerd[2160]: time="2025-07-12T00:08:51.103767827Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.103973 containerd[2160]: time="2025-07-12T00:08:51.103922291Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.105315 containerd[2160]: time="2025-07-12T00:08:51.104997419Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.105315 containerd[2160]: time="2025-07-12T00:08:51.105177551Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jul 12 00:08:51.106054 containerd[2160]: time="2025-07-12T00:08:51.105913091Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.106203 containerd[2160]: time="2025-07-12T00:08:51.106175963Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.109501967Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.111767543Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.112068899Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.112100951Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.112132787Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.112157975Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.112190315Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.112216211Z" level=info msg="NRI interface is disabled by configuration." Jul 12 00:08:51.113454 containerd[2160]: time="2025-07-12T00:08:51.112245479Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jul 12 00:08:51.113934 containerd[2160]: time="2025-07-12T00:08:51.112754387Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jul 12 00:08:51.113934 containerd[2160]: time="2025-07-12T00:08:51.112863323Z" level=info msg="Connect containerd service" Jul 12 00:08:51.113934 containerd[2160]: time="2025-07-12T00:08:51.113068835Z" level=info msg="using legacy CRI server" Jul 12 00:08:51.113934 containerd[2160]: time="2025-07-12T00:08:51.113100791Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jul 12 00:08:51.113934 containerd[2160]: time="2025-07-12T00:08:51.113300879Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jul 12 00:08:51.124490 containerd[2160]: time="2025-07-12T00:08:51.122995379Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 12 00:08:51.124490 containerd[2160]: time="2025-07-12T00:08:51.123307607Z" level=info msg="Start subscribing containerd event" Jul 12 00:08:51.124490 containerd[2160]: time="2025-07-12T00:08:51.123401327Z" level=info msg="Start recovering state" Jul 12 00:08:51.124490 containerd[2160]: time="2025-07-12T00:08:51.123564383Z" level=info msg="Start event monitor" Jul 12 00:08:51.124490 containerd[2160]: time="2025-07-12T00:08:51.123589979Z" level=info msg="Start snapshots syncer" Jul 12 00:08:51.124490 containerd[2160]: time="2025-07-12T00:08:51.123611315Z" level=info msg="Start cni network conf syncer for default" Jul 12 00:08:51.124490 containerd[2160]: time="2025-07-12T00:08:51.123629795Z" level=info msg="Start streaming server" Jul 12 00:08:51.129336 containerd[2160]: time="2025-07-12T00:08:51.129111119Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jul 12 00:08:51.130477 containerd[2160]: time="2025-07-12T00:08:51.129308591Z" level=info msg=serving... address=/run/containerd/containerd.sock Jul 12 00:08:51.133791 systemd[1]: Started containerd.service - containerd container runtime. Jul 12 00:08:51.143243 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO Agent will take identity from EC2 Jul 12 00:08:51.143303 containerd[2160]: time="2025-07-12T00:08:51.133757927Z" level=info msg="containerd successfully booted in 0.260933s" Jul 12 00:08:51.241283 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [amazon-ssm-agent] using named pipe channel for IPC Jul 12 00:08:51.343862 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [amazon-ssm-agent] using named pipe channel for IPC Jul 12 00:08:51.441055 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [amazon-ssm-agent] using named pipe channel for IPC Jul 12 00:08:51.542329 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Jul 12 00:08:51.615253 sshd_keygen[2149]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jul 12 00:08:51.642655 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [amazon-ssm-agent] OS: linux, Arch: arm64 Jul 12 00:08:51.666370 tar[2135]: linux-arm64/LICENSE Jul 12 00:08:51.666370 tar[2135]: linux-arm64/README.md Jul 12 00:08:51.687849 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [amazon-ssm-agent] Starting Core Agent Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [amazon-ssm-agent] registrar detected. Attempting registration Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [Registrar] Starting registrar module Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:50 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:51 INFO [EC2Identity] EC2 registration was successful. Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:51 INFO [CredentialRefresher] credentialRefresher has started Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:51 INFO [CredentialRefresher] Starting credentials refresher loop Jul 12 00:08:51.690839 amazon-ssm-agent[2247]: 2025-07-12 00:08:51 INFO EC2RoleProvider Successfully connected with instance profile role credentials Jul 12 00:08:51.703673 systemd[1]: Starting issuegen.service - Generate /run/issue... Jul 12 00:08:51.718178 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jul 12 00:08:51.737188 systemd[1]: issuegen.service: Deactivated successfully. Jul 12 00:08:51.737770 systemd[1]: Finished issuegen.service - Generate /run/issue. Jul 12 00:08:51.741753 amazon-ssm-agent[2247]: 2025-07-12 00:08:51 INFO [CredentialRefresher] Next credential rotation will be in 30.783242125133334 minutes Jul 12 00:08:51.750039 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jul 12 00:08:51.783243 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jul 12 00:08:51.798201 systemd[1]: Started getty@tty1.service - Getty on tty1. Jul 12 00:08:51.813024 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jul 12 00:08:51.816167 systemd[1]: Reached target getty.target - Login Prompts. Jul 12 00:08:52.718879 amazon-ssm-agent[2247]: 2025-07-12 00:08:52 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Jul 12 00:08:52.822486 amazon-ssm-agent[2247]: 2025-07-12 00:08:52 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2381) started Jul 12 00:08:52.920631 amazon-ssm-agent[2247]: 2025-07-12 00:08:52 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Jul 12 00:08:53.033775 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:08:53.038510 systemd[1]: Reached target multi-user.target - Multi-User System. Jul 12 00:08:53.046538 systemd[1]: Startup finished in 10.185s (kernel) + 10.281s (userspace) = 20.467s. Jul 12 00:08:53.051783 (kubelet)[2398]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 12 00:08:54.232212 kubelet[2398]: E0712 00:08:54.232092 2398 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 12 00:08:54.237089 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 12 00:08:54.237979 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 12 00:08:56.047208 systemd-resolved[2025]: Clock change detected. Flushing caches. Jul 12 00:08:58.019848 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jul 12 00:08:58.028598 systemd[1]: Started sshd@0-172.31.29.62:22-139.178.89.65:58252.service - OpenSSH per-connection server daemon (139.178.89.65:58252). Jul 12 00:08:58.213316 sshd[2411]: Accepted publickey for core from 139.178.89.65 port 58252 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:08:58.216518 sshd[2411]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:08:58.232789 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jul 12 00:08:58.241625 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jul 12 00:08:58.246261 systemd-logind[2117]: New session 1 of user core. Jul 12 00:08:58.271360 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jul 12 00:08:58.282693 systemd[1]: Starting user@500.service - User Manager for UID 500... Jul 12 00:08:58.304167 (systemd)[2417]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jul 12 00:08:58.530630 systemd[2417]: Queued start job for default target default.target. Jul 12 00:08:58.532172 systemd[2417]: Created slice app.slice - User Application Slice. Jul 12 00:08:58.532228 systemd[2417]: Reached target paths.target - Paths. Jul 12 00:08:58.532261 systemd[2417]: Reached target timers.target - Timers. Jul 12 00:08:58.540292 systemd[2417]: Starting dbus.socket - D-Bus User Message Bus Socket... Jul 12 00:08:58.556215 systemd[2417]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jul 12 00:08:58.556347 systemd[2417]: Reached target sockets.target - Sockets. Jul 12 00:08:58.556379 systemd[2417]: Reached target basic.target - Basic System. Jul 12 00:08:58.556477 systemd[2417]: Reached target default.target - Main User Target. Jul 12 00:08:58.556539 systemd[2417]: Startup finished in 240ms. Jul 12 00:08:58.557333 systemd[1]: Started user@500.service - User Manager for UID 500. Jul 12 00:08:58.563976 systemd[1]: Started session-1.scope - Session 1 of User core. Jul 12 00:08:58.717726 systemd[1]: Started sshd@1-172.31.29.62:22-139.178.89.65:58268.service - OpenSSH per-connection server daemon (139.178.89.65:58268). Jul 12 00:08:58.899772 sshd[2429]: Accepted publickey for core from 139.178.89.65 port 58268 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:08:58.902265 sshd[2429]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:08:58.911058 systemd-logind[2117]: New session 2 of user core. Jul 12 00:08:58.916690 systemd[1]: Started session-2.scope - Session 2 of User core. Jul 12 00:08:59.047126 sshd[2429]: pam_unix(sshd:session): session closed for user core Jul 12 00:08:59.054093 systemd-logind[2117]: Session 2 logged out. Waiting for processes to exit. Jul 12 00:08:59.054967 systemd[1]: sshd@1-172.31.29.62:22-139.178.89.65:58268.service: Deactivated successfully. Jul 12 00:08:59.059738 systemd[1]: session-2.scope: Deactivated successfully. Jul 12 00:08:59.061717 systemd-logind[2117]: Removed session 2. Jul 12 00:08:59.079645 systemd[1]: Started sshd@2-172.31.29.62:22-139.178.89.65:58272.service - OpenSSH per-connection server daemon (139.178.89.65:58272). Jul 12 00:08:59.246241 sshd[2437]: Accepted publickey for core from 139.178.89.65 port 58272 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:08:59.249408 sshd[2437]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:08:59.258195 systemd-logind[2117]: New session 3 of user core. Jul 12 00:08:59.266585 systemd[1]: Started session-3.scope - Session 3 of User core. Jul 12 00:08:59.387777 sshd[2437]: pam_unix(sshd:session): session closed for user core Jul 12 00:08:59.395536 systemd[1]: sshd@2-172.31.29.62:22-139.178.89.65:58272.service: Deactivated successfully. Jul 12 00:08:59.396633 systemd-logind[2117]: Session 3 logged out. Waiting for processes to exit. Jul 12 00:08:59.401359 systemd[1]: session-3.scope: Deactivated successfully. Jul 12 00:08:59.403153 systemd-logind[2117]: Removed session 3. Jul 12 00:08:59.420584 systemd[1]: Started sshd@3-172.31.29.62:22-139.178.89.65:58278.service - OpenSSH per-connection server daemon (139.178.89.65:58278). Jul 12 00:08:59.584992 sshd[2445]: Accepted publickey for core from 139.178.89.65 port 58278 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:08:59.588130 sshd[2445]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:08:59.596690 systemd-logind[2117]: New session 4 of user core. Jul 12 00:08:59.607654 systemd[1]: Started session-4.scope - Session 4 of User core. Jul 12 00:08:59.734426 sshd[2445]: pam_unix(sshd:session): session closed for user core Jul 12 00:08:59.741556 systemd-logind[2117]: Session 4 logged out. Waiting for processes to exit. Jul 12 00:08:59.742536 systemd[1]: sshd@3-172.31.29.62:22-139.178.89.65:58278.service: Deactivated successfully. Jul 12 00:08:59.748081 systemd[1]: session-4.scope: Deactivated successfully. Jul 12 00:08:59.750276 systemd-logind[2117]: Removed session 4. Jul 12 00:08:59.765671 systemd[1]: Started sshd@4-172.31.29.62:22-139.178.89.65:48532.service - OpenSSH per-connection server daemon (139.178.89.65:48532). Jul 12 00:08:59.940746 sshd[2453]: Accepted publickey for core from 139.178.89.65 port 48532 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:08:59.943383 sshd[2453]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:08:59.952513 systemd-logind[2117]: New session 5 of user core. Jul 12 00:08:59.962674 systemd[1]: Started session-5.scope - Session 5 of User core. Jul 12 00:09:00.083046 sudo[2457]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jul 12 00:09:00.083729 sudo[2457]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 00:09:00.103873 sudo[2457]: pam_unix(sudo:session): session closed for user root Jul 12 00:09:00.127589 sshd[2453]: pam_unix(sshd:session): session closed for user core Jul 12 00:09:00.135311 systemd[1]: sshd@4-172.31.29.62:22-139.178.89.65:48532.service: Deactivated successfully. Jul 12 00:09:00.140510 systemd-logind[2117]: Session 5 logged out. Waiting for processes to exit. Jul 12 00:09:00.141792 systemd[1]: session-5.scope: Deactivated successfully. Jul 12 00:09:00.143583 systemd-logind[2117]: Removed session 5. Jul 12 00:09:00.159596 systemd[1]: Started sshd@5-172.31.29.62:22-139.178.89.65:48538.service - OpenSSH per-connection server daemon (139.178.89.65:48538). Jul 12 00:09:00.338866 sshd[2462]: Accepted publickey for core from 139.178.89.65 port 48538 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:09:00.341873 sshd[2462]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:09:00.350267 systemd-logind[2117]: New session 6 of user core. Jul 12 00:09:00.357668 systemd[1]: Started session-6.scope - Session 6 of User core. Jul 12 00:09:00.465080 sudo[2467]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jul 12 00:09:00.466441 sudo[2467]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 00:09:00.473028 sudo[2467]: pam_unix(sudo:session): session closed for user root Jul 12 00:09:00.483080 sudo[2466]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Jul 12 00:09:00.483734 sudo[2466]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 00:09:00.505604 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Jul 12 00:09:00.520276 auditctl[2470]: No rules Jul 12 00:09:00.523275 systemd[1]: audit-rules.service: Deactivated successfully. Jul 12 00:09:00.523830 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Jul 12 00:09:00.531947 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jul 12 00:09:00.587250 augenrules[2489]: No rules Jul 12 00:09:00.590725 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jul 12 00:09:00.594588 sudo[2466]: pam_unix(sudo:session): session closed for user root Jul 12 00:09:00.619889 sshd[2462]: pam_unix(sshd:session): session closed for user core Jul 12 00:09:00.626550 systemd[1]: sshd@5-172.31.29.62:22-139.178.89.65:48538.service: Deactivated successfully. Jul 12 00:09:00.626820 systemd-logind[2117]: Session 6 logged out. Waiting for processes to exit. Jul 12 00:09:00.633664 systemd[1]: session-6.scope: Deactivated successfully. Jul 12 00:09:00.635424 systemd-logind[2117]: Removed session 6. Jul 12 00:09:00.651631 systemd[1]: Started sshd@6-172.31.29.62:22-139.178.89.65:48546.service - OpenSSH per-connection server daemon (139.178.89.65:48546). Jul 12 00:09:00.821335 sshd[2498]: Accepted publickey for core from 139.178.89.65 port 48546 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:09:00.823901 sshd[2498]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:09:00.832168 systemd-logind[2117]: New session 7 of user core. Jul 12 00:09:00.839602 systemd[1]: Started session-7.scope - Session 7 of User core. Jul 12 00:09:00.944563 sudo[2502]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jul 12 00:09:00.945439 sudo[2502]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 12 00:09:01.436517 systemd[1]: Starting docker.service - Docker Application Container Engine... Jul 12 00:09:01.436874 (dockerd)[2517]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jul 12 00:09:01.849212 dockerd[2517]: time="2025-07-12T00:09:01.848802720Z" level=info msg="Starting up" Jul 12 00:09:01.968753 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport93951858-merged.mount: Deactivated successfully. Jul 12 00:09:02.181589 systemd[1]: var-lib-docker-metacopy\x2dcheck1070996216-merged.mount: Deactivated successfully. Jul 12 00:09:02.194636 dockerd[2517]: time="2025-07-12T00:09:02.194541622Z" level=info msg="Loading containers: start." Jul 12 00:09:02.357150 kernel: Initializing XFRM netlink socket Jul 12 00:09:02.389988 (udev-worker)[2539]: Network interface NamePolicy= disabled on kernel command line. Jul 12 00:09:02.478495 systemd-networkd[1694]: docker0: Link UP Jul 12 00:09:02.502924 dockerd[2517]: time="2025-07-12T00:09:02.501565620Z" level=info msg="Loading containers: done." Jul 12 00:09:02.526833 dockerd[2517]: time="2025-07-12T00:09:02.526655820Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jul 12 00:09:02.526833 dockerd[2517]: time="2025-07-12T00:09:02.526803216Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Jul 12 00:09:02.527130 dockerd[2517]: time="2025-07-12T00:09:02.526995444Z" level=info msg="Daemon has completed initialization" Jul 12 00:09:02.577921 dockerd[2517]: time="2025-07-12T00:09:02.577694100Z" level=info msg="API listen on /run/docker.sock" Jul 12 00:09:02.578347 systemd[1]: Started docker.service - Docker Application Container Engine. Jul 12 00:09:02.958716 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck442496653-merged.mount: Deactivated successfully. Jul 12 00:09:03.675731 containerd[2160]: time="2025-07-12T00:09:03.675644137Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\"" Jul 12 00:09:04.109589 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jul 12 00:09:04.121422 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:09:04.329329 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2087679378.mount: Deactivated successfully. Jul 12 00:09:04.511993 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:04.539805 (kubelet)[2682]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 12 00:09:04.642132 kubelet[2682]: E0712 00:09:04.642021 2682 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 12 00:09:04.656518 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 12 00:09:04.660200 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 12 00:09:05.740938 containerd[2160]: time="2025-07-12T00:09:05.740856928Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:05.742996 containerd[2160]: time="2025-07-12T00:09:05.742940260Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.10: active requests=0, bytes read=25651793" Jul 12 00:09:05.744958 containerd[2160]: time="2025-07-12T00:09:05.744882664Z" level=info msg="ImageCreate event name:\"sha256:8907c2d36348551c1038e24ef688f6830681069380376707e55518007a20a86c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:05.751140 containerd[2160]: time="2025-07-12T00:09:05.751017916Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:05.753839 containerd[2160]: time="2025-07-12T00:09:05.753416200Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.10\" with image id \"sha256:8907c2d36348551c1038e24ef688f6830681069380376707e55518007a20a86c\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\", size \"25648593\" in 2.077706243s" Jul 12 00:09:05.753839 containerd[2160]: time="2025-07-12T00:09:05.753488320Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\" returns image reference \"sha256:8907c2d36348551c1038e24ef688f6830681069380376707e55518007a20a86c\"" Jul 12 00:09:05.756519 containerd[2160]: time="2025-07-12T00:09:05.756212164Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\"" Jul 12 00:09:07.097243 containerd[2160]: time="2025-07-12T00:09:07.096298970Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:07.098562 containerd[2160]: time="2025-07-12T00:09:07.098491634Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.10: active requests=0, bytes read=22459677" Jul 12 00:09:07.100000 containerd[2160]: time="2025-07-12T00:09:07.099930554Z" level=info msg="ImageCreate event name:\"sha256:0f640d6889416d515a0ac4de1c26f4d80134c47641ff464abc831560a951175f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:07.107004 containerd[2160]: time="2025-07-12T00:09:07.106916702Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:07.109720 containerd[2160]: time="2025-07-12T00:09:07.109263050Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.10\" with image id \"sha256:0f640d6889416d515a0ac4de1c26f4d80134c47641ff464abc831560a951175f\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\", size \"23995467\" in 1.35298551s" Jul 12 00:09:07.109720 containerd[2160]: time="2025-07-12T00:09:07.109320326Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\" returns image reference \"sha256:0f640d6889416d515a0ac4de1c26f4d80134c47641ff464abc831560a951175f\"" Jul 12 00:09:07.110587 containerd[2160]: time="2025-07-12T00:09:07.110300498Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\"" Jul 12 00:09:08.332114 containerd[2160]: time="2025-07-12T00:09:08.332031605Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:08.334156 containerd[2160]: time="2025-07-12T00:09:08.334072865Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.10: active requests=0, bytes read=17125066" Jul 12 00:09:08.336139 containerd[2160]: time="2025-07-12T00:09:08.335026373Z" level=info msg="ImageCreate event name:\"sha256:23d79b83d912e2633bcb4f9f7b8b46024893e11d492a4249d8f1f8c9a26b7b2c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:08.340866 containerd[2160]: time="2025-07-12T00:09:08.340772633Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:08.343405 containerd[2160]: time="2025-07-12T00:09:08.343177253Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.10\" with image id \"sha256:23d79b83d912e2633bcb4f9f7b8b46024893e11d492a4249d8f1f8c9a26b7b2c\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\", size \"18660874\" in 1.232794939s" Jul 12 00:09:08.343405 containerd[2160]: time="2025-07-12T00:09:08.343241585Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\" returns image reference \"sha256:23d79b83d912e2633bcb4f9f7b8b46024893e11d492a4249d8f1f8c9a26b7b2c\"" Jul 12 00:09:08.344454 containerd[2160]: time="2025-07-12T00:09:08.344056265Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\"" Jul 12 00:09:09.562414 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1250239365.mount: Deactivated successfully. Jul 12 00:09:10.084888 containerd[2160]: time="2025-07-12T00:09:10.084829109Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:10.086929 containerd[2160]: time="2025-07-12T00:09:10.086872085Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.10: active requests=0, bytes read=26915957" Jul 12 00:09:10.087662 containerd[2160]: time="2025-07-12T00:09:10.087593177Z" level=info msg="ImageCreate event name:\"sha256:dde5ff0da443b455e81aefc7bf6a216fdd659d1cbe13b8e8ac8129c3ecd27f89\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:10.091558 containerd[2160]: time="2025-07-12T00:09:10.091482665Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:10.093079 containerd[2160]: time="2025-07-12T00:09:10.092849117Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.10\" with image id \"sha256:dde5ff0da443b455e81aefc7bf6a216fdd659d1cbe13b8e8ac8129c3ecd27f89\", repo tag \"registry.k8s.io/kube-proxy:v1.31.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\", size \"26914976\" in 1.748741456s" Jul 12 00:09:10.093079 containerd[2160]: time="2025-07-12T00:09:10.092919101Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\" returns image reference \"sha256:dde5ff0da443b455e81aefc7bf6a216fdd659d1cbe13b8e8ac8129c3ecd27f89\"" Jul 12 00:09:10.094092 containerd[2160]: time="2025-07-12T00:09:10.093892985Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Jul 12 00:09:10.720933 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1584531111.mount: Deactivated successfully. Jul 12 00:09:12.003778 containerd[2160]: time="2025-07-12T00:09:12.003672847Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:12.006425 containerd[2160]: time="2025-07-12T00:09:12.006353167Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951622" Jul 12 00:09:12.008525 containerd[2160]: time="2025-07-12T00:09:12.008431675Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:12.020157 containerd[2160]: time="2025-07-12T00:09:12.018501379Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:12.023035 containerd[2160]: time="2025-07-12T00:09:12.022969579Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.92902221s" Jul 12 00:09:12.023283 containerd[2160]: time="2025-07-12T00:09:12.023249131Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Jul 12 00:09:12.024656 containerd[2160]: time="2025-07-12T00:09:12.024594079Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jul 12 00:09:12.534643 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3799862704.mount: Deactivated successfully. Jul 12 00:09:12.546203 containerd[2160]: time="2025-07-12T00:09:12.546082941Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:12.548051 containerd[2160]: time="2025-07-12T00:09:12.547980765Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Jul 12 00:09:12.550592 containerd[2160]: time="2025-07-12T00:09:12.550517433Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:12.557060 containerd[2160]: time="2025-07-12T00:09:12.556985806Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:12.559721 containerd[2160]: time="2025-07-12T00:09:12.559630690Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 534.971715ms" Jul 12 00:09:12.559721 containerd[2160]: time="2025-07-12T00:09:12.559715974Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Jul 12 00:09:12.560589 containerd[2160]: time="2025-07-12T00:09:12.560545546Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Jul 12 00:09:13.156358 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1104530015.mount: Deactivated successfully. Jul 12 00:09:14.907315 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jul 12 00:09:14.914529 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:09:15.717359 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:15.731872 (kubelet)[2869]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 12 00:09:15.819565 kubelet[2869]: E0712 00:09:15.819187 2869 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 12 00:09:15.826845 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 12 00:09:15.827463 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 12 00:09:16.799818 containerd[2160]: time="2025-07-12T00:09:16.799758903Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:16.802833 containerd[2160]: time="2025-07-12T00:09:16.802792083Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66406465" Jul 12 00:09:16.804355 containerd[2160]: time="2025-07-12T00:09:16.804311079Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:16.811540 containerd[2160]: time="2025-07-12T00:09:16.811468179Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:16.814312 containerd[2160]: time="2025-07-12T00:09:16.814247571Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 4.253645721s" Jul 12 00:09:16.814778 containerd[2160]: time="2025-07-12T00:09:16.814309959Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Jul 12 00:09:20.597590 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Jul 12 00:09:24.193364 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:24.201602 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:09:24.257655 systemd[1]: Reloading requested from client PID 2911 ('systemctl') (unit session-7.scope)... Jul 12 00:09:24.257688 systemd[1]: Reloading... Jul 12 00:09:24.489158 zram_generator::config[2954]: No configuration found. Jul 12 00:09:24.749356 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 00:09:24.919258 systemd[1]: Reloading finished in 660 ms. Jul 12 00:09:25.017468 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:25.027256 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:09:25.030954 systemd[1]: kubelet.service: Deactivated successfully. Jul 12 00:09:25.031530 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:25.040574 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:09:25.361585 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:25.386810 (kubelet)[3029]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 12 00:09:25.457825 kubelet[3029]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 00:09:25.457825 kubelet[3029]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 12 00:09:25.457825 kubelet[3029]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 00:09:25.458514 kubelet[3029]: I0712 00:09:25.457930 3029 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 12 00:09:26.000333 kubelet[3029]: I0712 00:09:26.000265 3029 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 12 00:09:26.000333 kubelet[3029]: I0712 00:09:26.000316 3029 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 12 00:09:26.000805 kubelet[3029]: I0712 00:09:26.000757 3029 server.go:934] "Client rotation is on, will bootstrap in background" Jul 12 00:09:26.040250 kubelet[3029]: E0712 00:09:26.040147 3029 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.29.62:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:26.041797 kubelet[3029]: I0712 00:09:26.041737 3029 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 12 00:09:26.056916 kubelet[3029]: E0712 00:09:26.056721 3029 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jul 12 00:09:26.056916 kubelet[3029]: I0712 00:09:26.056783 3029 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jul 12 00:09:26.064369 kubelet[3029]: I0712 00:09:26.064049 3029 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 12 00:09:26.065154 kubelet[3029]: I0712 00:09:26.065086 3029 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 12 00:09:26.065408 kubelet[3029]: I0712 00:09:26.065350 3029 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 12 00:09:26.065688 kubelet[3029]: I0712 00:09:26.065409 3029 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-29-62","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Jul 12 00:09:26.065872 kubelet[3029]: I0712 00:09:26.065840 3029 topology_manager.go:138] "Creating topology manager with none policy" Jul 12 00:09:26.065937 kubelet[3029]: I0712 00:09:26.065871 3029 container_manager_linux.go:300] "Creating device plugin manager" Jul 12 00:09:26.066230 kubelet[3029]: I0712 00:09:26.066201 3029 state_mem.go:36] "Initialized new in-memory state store" Jul 12 00:09:26.071456 kubelet[3029]: I0712 00:09:26.070899 3029 kubelet.go:408] "Attempting to sync node with API server" Jul 12 00:09:26.071456 kubelet[3029]: I0712 00:09:26.070947 3029 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 12 00:09:26.071456 kubelet[3029]: I0712 00:09:26.070982 3029 kubelet.go:314] "Adding apiserver pod source" Jul 12 00:09:26.071456 kubelet[3029]: I0712 00:09:26.071194 3029 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 12 00:09:26.073344 kubelet[3029]: W0712 00:09:26.072679 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.29.62:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-29-62&limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:26.073344 kubelet[3029]: E0712 00:09:26.072802 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.29.62:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-29-62&limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:26.079846 kubelet[3029]: I0712 00:09:26.079809 3029 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jul 12 00:09:26.081236 kubelet[3029]: I0712 00:09:26.081202 3029 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 12 00:09:26.081701 kubelet[3029]: W0712 00:09:26.081680 3029 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jul 12 00:09:26.083608 kubelet[3029]: I0712 00:09:26.083573 3029 server.go:1274] "Started kubelet" Jul 12 00:09:26.085267 kubelet[3029]: W0712 00:09:26.083961 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.29.62:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:26.085267 kubelet[3029]: E0712 00:09:26.084048 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.29.62:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:26.087298 kubelet[3029]: I0712 00:09:26.087205 3029 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 12 00:09:26.087858 kubelet[3029]: I0712 00:09:26.087814 3029 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 12 00:09:26.088120 kubelet[3029]: I0712 00:09:26.088053 3029 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 12 00:09:26.089810 kubelet[3029]: I0712 00:09:26.089769 3029 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 12 00:09:26.090361 kubelet[3029]: I0712 00:09:26.090317 3029 server.go:449] "Adding debug handlers to kubelet server" Jul 12 00:09:26.097704 kubelet[3029]: I0712 00:09:26.097672 3029 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 12 00:09:26.098629 kubelet[3029]: E0712 00:09:26.098597 3029 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-29-62\" not found" Jul 12 00:09:26.101243 kubelet[3029]: I0712 00:09:26.099335 3029 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 12 00:09:26.102059 kubelet[3029]: I0712 00:09:26.099763 3029 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 12 00:09:26.104380 kubelet[3029]: I0712 00:09:26.099844 3029 reconciler.go:26] "Reconciler: start to sync state" Jul 12 00:09:26.104728 kubelet[3029]: I0712 00:09:26.104690 3029 factory.go:221] Registration of the systemd container factory successfully Jul 12 00:09:26.106593 kubelet[3029]: I0712 00:09:26.106548 3029 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 12 00:09:26.108130 kubelet[3029]: E0712 00:09:26.104538 3029 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.29.62:6443/api/v1/namespaces/default/events\": dial tcp 172.31.29.62:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-29-62.1851586f88cc5295 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-29-62,UID:ip-172-31-29-62,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-29-62,},FirstTimestamp:2025-07-12 00:09:26.083539605 +0000 UTC m=+0.690713645,LastTimestamp:2025-07-12 00:09:26.083539605 +0000 UTC m=+0.690713645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-29-62,}" Jul 12 00:09:26.108552 kubelet[3029]: W0712 00:09:26.108417 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.29.62:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:26.108552 kubelet[3029]: E0712 00:09:26.108507 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.29.62:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:26.108694 kubelet[3029]: E0712 00:09:26.108615 3029 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.62:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-62?timeout=10s\": dial tcp 172.31.29.62:6443: connect: connection refused" interval="200ms" Jul 12 00:09:26.111346 kubelet[3029]: I0712 00:09:26.111221 3029 factory.go:221] Registration of the containerd container factory successfully Jul 12 00:09:26.112305 kubelet[3029]: E0712 00:09:26.112160 3029 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 12 00:09:26.164146 kubelet[3029]: I0712 00:09:26.163863 3029 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 12 00:09:26.166772 kubelet[3029]: I0712 00:09:26.166732 3029 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 12 00:09:26.167371 kubelet[3029]: I0712 00:09:26.167347 3029 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 12 00:09:26.168043 kubelet[3029]: I0712 00:09:26.168003 3029 kubelet.go:2321] "Starting kubelet main sync loop" Jul 12 00:09:26.168826 kubelet[3029]: E0712 00:09:26.168784 3029 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 12 00:09:26.171200 kubelet[3029]: I0712 00:09:26.170451 3029 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 12 00:09:26.171200 kubelet[3029]: I0712 00:09:26.170776 3029 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 12 00:09:26.171200 kubelet[3029]: I0712 00:09:26.170811 3029 state_mem.go:36] "Initialized new in-memory state store" Jul 12 00:09:26.171438 kubelet[3029]: W0712 00:09:26.171413 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.29.62:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:26.171510 kubelet[3029]: E0712 00:09:26.171473 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.29.62:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:26.177615 kubelet[3029]: I0712 00:09:26.177579 3029 policy_none.go:49] "None policy: Start" Jul 12 00:09:26.179282 kubelet[3029]: I0712 00:09:26.179198 3029 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 12 00:09:26.179282 kubelet[3029]: I0712 00:09:26.179242 3029 state_mem.go:35] "Initializing new in-memory state store" Jul 12 00:09:26.191144 kubelet[3029]: I0712 00:09:26.190284 3029 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 12 00:09:26.191144 kubelet[3029]: I0712 00:09:26.190576 3029 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 12 00:09:26.191144 kubelet[3029]: I0712 00:09:26.190596 3029 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 12 00:09:26.194483 kubelet[3029]: I0712 00:09:26.194454 3029 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 12 00:09:26.197958 kubelet[3029]: E0712 00:09:26.197911 3029 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-29-62\" not found" Jul 12 00:09:26.294888 kubelet[3029]: I0712 00:09:26.294647 3029 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-29-62" Jul 12 00:09:26.296663 kubelet[3029]: E0712 00:09:26.296546 3029 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.29.62:6443/api/v1/nodes\": dial tcp 172.31.29.62:6443: connect: connection refused" node="ip-172-31-29-62" Jul 12 00:09:26.305896 kubelet[3029]: I0712 00:09:26.305463 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-ca-certs\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:26.305896 kubelet[3029]: I0712 00:09:26.305518 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:26.305896 kubelet[3029]: I0712 00:09:26.305559 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:26.305896 kubelet[3029]: I0712 00:09:26.305596 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/830d67704643f783ed5af670e9b72fca-ca-certs\") pod \"kube-apiserver-ip-172-31-29-62\" (UID: \"830d67704643f783ed5af670e9b72fca\") " pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:26.305896 kubelet[3029]: I0712 00:09:26.305632 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/830d67704643f783ed5af670e9b72fca-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-29-62\" (UID: \"830d67704643f783ed5af670e9b72fca\") " pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:26.306286 kubelet[3029]: I0712 00:09:26.305671 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-k8s-certs\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:26.306286 kubelet[3029]: I0712 00:09:26.305707 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-kubeconfig\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:26.306286 kubelet[3029]: I0712 00:09:26.305742 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a4fbd316186c0850039629540f818287-kubeconfig\") pod \"kube-scheduler-ip-172-31-29-62\" (UID: \"a4fbd316186c0850039629540f818287\") " pod="kube-system/kube-scheduler-ip-172-31-29-62" Jul 12 00:09:26.306286 kubelet[3029]: I0712 00:09:26.305776 3029 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/830d67704643f783ed5af670e9b72fca-k8s-certs\") pod \"kube-apiserver-ip-172-31-29-62\" (UID: \"830d67704643f783ed5af670e9b72fca\") " pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:26.310275 kubelet[3029]: E0712 00:09:26.310220 3029 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.62:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-62?timeout=10s\": dial tcp 172.31.29.62:6443: connect: connection refused" interval="400ms" Jul 12 00:09:26.499279 kubelet[3029]: I0712 00:09:26.499218 3029 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-29-62" Jul 12 00:09:26.499901 kubelet[3029]: E0712 00:09:26.499716 3029 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.29.62:6443/api/v1/nodes\": dial tcp 172.31.29.62:6443: connect: connection refused" node="ip-172-31-29-62" Jul 12 00:09:26.544562 kubelet[3029]: E0712 00:09:26.544401 3029 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.29.62:6443/api/v1/namespaces/default/events\": dial tcp 172.31.29.62:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-29-62.1851586f88cc5295 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-29-62,UID:ip-172-31-29-62,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-29-62,},FirstTimestamp:2025-07-12 00:09:26.083539605 +0000 UTC m=+0.690713645,LastTimestamp:2025-07-12 00:09:26.083539605 +0000 UTC m=+0.690713645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-29-62,}" Jul 12 00:09:26.582087 containerd[2160]: time="2025-07-12T00:09:26.581641871Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-29-62,Uid:a4fbd316186c0850039629540f818287,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:26.589528 containerd[2160]: time="2025-07-12T00:09:26.589134023Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-29-62,Uid:494476933b0c0f1b79178bb9c91c0455,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:26.590263 containerd[2160]: time="2025-07-12T00:09:26.590198219Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-29-62,Uid:830d67704643f783ed5af670e9b72fca,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:26.711579 kubelet[3029]: E0712 00:09:26.711514 3029 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.62:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-62?timeout=10s\": dial tcp 172.31.29.62:6443: connect: connection refused" interval="800ms" Jul 12 00:09:26.902616 kubelet[3029]: I0712 00:09:26.902114 3029 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-29-62" Jul 12 00:09:26.902616 kubelet[3029]: E0712 00:09:26.902568 3029 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.29.62:6443/api/v1/nodes\": dial tcp 172.31.29.62:6443: connect: connection refused" node="ip-172-31-29-62" Jul 12 00:09:26.973733 kubelet[3029]: W0712 00:09:26.973653 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.29.62:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:26.973898 kubelet[3029]: E0712 00:09:26.973750 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.29.62:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:27.088416 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1314198675.mount: Deactivated successfully. Jul 12 00:09:27.101937 containerd[2160]: time="2025-07-12T00:09:27.101657470Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 00:09:27.108634 containerd[2160]: time="2025-07-12T00:09:27.108557434Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Jul 12 00:09:27.110518 containerd[2160]: time="2025-07-12T00:09:27.110453518Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 00:09:27.113555 containerd[2160]: time="2025-07-12T00:09:27.113293798Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 00:09:27.115711 containerd[2160]: time="2025-07-12T00:09:27.115637026Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 00:09:27.117559 containerd[2160]: time="2025-07-12T00:09:27.117492310Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jul 12 00:09:27.119012 containerd[2160]: time="2025-07-12T00:09:27.118903042Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jul 12 00:09:27.123850 containerd[2160]: time="2025-07-12T00:09:27.123762766Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 12 00:09:27.126838 containerd[2160]: time="2025-07-12T00:09:27.126435214Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 544.686471ms" Jul 12 00:09:27.129046 containerd[2160]: time="2025-07-12T00:09:27.128685766Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 538.375875ms" Jul 12 00:09:27.137829 containerd[2160]: time="2025-07-12T00:09:27.137484670Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 548.238459ms" Jul 12 00:09:27.171344 kubelet[3029]: W0712 00:09:27.170030 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.29.62:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:27.171344 kubelet[3029]: E0712 00:09:27.170149 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.29.62:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:27.345179 kubelet[3029]: W0712 00:09:27.344987 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.29.62:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-29-62&limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:27.345179 kubelet[3029]: E0712 00:09:27.345083 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.29.62:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-29-62&limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:27.380132 containerd[2160]: time="2025-07-12T00:09:27.379958519Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:09:27.380383 containerd[2160]: time="2025-07-12T00:09:27.380068727Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:09:27.380383 containerd[2160]: time="2025-07-12T00:09:27.380143307Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:27.380604 containerd[2160]: time="2025-07-12T00:09:27.380336939Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:27.385191 containerd[2160]: time="2025-07-12T00:09:27.384986003Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:09:27.387026 containerd[2160]: time="2025-07-12T00:09:27.385117727Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:09:27.387427 containerd[2160]: time="2025-07-12T00:09:27.386716679Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:27.387770 containerd[2160]: time="2025-07-12T00:09:27.387640175Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:09:27.388447 containerd[2160]: time="2025-07-12T00:09:27.388300391Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:27.388632 containerd[2160]: time="2025-07-12T00:09:27.388463603Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:09:27.388632 containerd[2160]: time="2025-07-12T00:09:27.388546523Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:27.388919 containerd[2160]: time="2025-07-12T00:09:27.388735883Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:27.494211 kubelet[3029]: W0712 00:09:27.494018 3029 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.29.62:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.29.62:6443: connect: connection refused Jul 12 00:09:27.496580 kubelet[3029]: E0712 00:09:27.494091 3029 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.29.62:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.29.62:6443: connect: connection refused" logger="UnhandledError" Jul 12 00:09:27.512255 kubelet[3029]: E0712 00:09:27.512004 3029 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.62:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-62?timeout=10s\": dial tcp 172.31.29.62:6443: connect: connection refused" interval="1.6s" Jul 12 00:09:27.543696 containerd[2160]: time="2025-07-12T00:09:27.542559120Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-29-62,Uid:830d67704643f783ed5af670e9b72fca,Namespace:kube-system,Attempt:0,} returns sandbox id \"1d5027fdfc65953b6f731dfe31982bfa21a31f232f7353705aff96995efc82e0\"" Jul 12 00:09:27.556701 containerd[2160]: time="2025-07-12T00:09:27.556517124Z" level=info msg="CreateContainer within sandbox \"1d5027fdfc65953b6f731dfe31982bfa21a31f232f7353705aff96995efc82e0\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jul 12 00:09:27.564527 containerd[2160]: time="2025-07-12T00:09:27.564324168Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-29-62,Uid:a4fbd316186c0850039629540f818287,Namespace:kube-system,Attempt:0,} returns sandbox id \"fba0ab38385978627822e35ac670b4acfe00d1905088a0c8e79e43dc973a8506\"" Jul 12 00:09:27.567803 containerd[2160]: time="2025-07-12T00:09:27.566672940Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-29-62,Uid:494476933b0c0f1b79178bb9c91c0455,Namespace:kube-system,Attempt:0,} returns sandbox id \"a0fc4c1fb46341a8a0fd3cd96f3a362516d006e986359c14a83f169dd2713fec\"" Jul 12 00:09:27.579035 containerd[2160]: time="2025-07-12T00:09:27.578900952Z" level=info msg="CreateContainer within sandbox \"fba0ab38385978627822e35ac670b4acfe00d1905088a0c8e79e43dc973a8506\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jul 12 00:09:27.579909 containerd[2160]: time="2025-07-12T00:09:27.579861468Z" level=info msg="CreateContainer within sandbox \"a0fc4c1fb46341a8a0fd3cd96f3a362516d006e986359c14a83f169dd2713fec\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jul 12 00:09:27.611632 containerd[2160]: time="2025-07-12T00:09:27.611573712Z" level=info msg="CreateContainer within sandbox \"1d5027fdfc65953b6f731dfe31982bfa21a31f232f7353705aff96995efc82e0\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"34b058ab58b5c0bef4e6adbb1b01abb88fa51ed9a1338663360496afd47a39ca\"" Jul 12 00:09:27.613849 containerd[2160]: time="2025-07-12T00:09:27.613391820Z" level=info msg="StartContainer for \"34b058ab58b5c0bef4e6adbb1b01abb88fa51ed9a1338663360496afd47a39ca\"" Jul 12 00:09:27.625279 containerd[2160]: time="2025-07-12T00:09:27.625209816Z" level=info msg="CreateContainer within sandbox \"a0fc4c1fb46341a8a0fd3cd96f3a362516d006e986359c14a83f169dd2713fec\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b5a4318fbac094b856dc52c3ba989532a56b47a22d4d6e5df5216a0383f187d4\"" Jul 12 00:09:27.626297 containerd[2160]: time="2025-07-12T00:09:27.626223336Z" level=info msg="StartContainer for \"b5a4318fbac094b856dc52c3ba989532a56b47a22d4d6e5df5216a0383f187d4\"" Jul 12 00:09:27.635424 containerd[2160]: time="2025-07-12T00:09:27.635362476Z" level=info msg="CreateContainer within sandbox \"fba0ab38385978627822e35ac670b4acfe00d1905088a0c8e79e43dc973a8506\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ee89a1192f85ce65d9f50ba839b8838599dc4ab658e08fdd97c748d1e6df178b\"" Jul 12 00:09:27.639126 containerd[2160]: time="2025-07-12T00:09:27.637630560Z" level=info msg="StartContainer for \"ee89a1192f85ce65d9f50ba839b8838599dc4ab658e08fdd97c748d1e6df178b\"" Jul 12 00:09:27.708797 kubelet[3029]: I0712 00:09:27.708759 3029 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-29-62" Jul 12 00:09:27.709855 kubelet[3029]: E0712 00:09:27.709807 3029 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.29.62:6443/api/v1/nodes\": dial tcp 172.31.29.62:6443: connect: connection refused" node="ip-172-31-29-62" Jul 12 00:09:27.851383 containerd[2160]: time="2025-07-12T00:09:27.849047053Z" level=info msg="StartContainer for \"34b058ab58b5c0bef4e6adbb1b01abb88fa51ed9a1338663360496afd47a39ca\" returns successfully" Jul 12 00:09:27.870210 containerd[2160]: time="2025-07-12T00:09:27.868798706Z" level=info msg="StartContainer for \"b5a4318fbac094b856dc52c3ba989532a56b47a22d4d6e5df5216a0383f187d4\" returns successfully" Jul 12 00:09:27.877599 containerd[2160]: time="2025-07-12T00:09:27.877476854Z" level=info msg="StartContainer for \"ee89a1192f85ce65d9f50ba839b8838599dc4ab658e08fdd97c748d1e6df178b\" returns successfully" Jul 12 00:09:29.319988 kubelet[3029]: I0712 00:09:29.319509 3029 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-29-62" Jul 12 00:09:31.038818 kubelet[3029]: E0712 00:09:31.038755 3029 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-29-62\" not found" node="ip-172-31-29-62" Jul 12 00:09:31.079130 kubelet[3029]: I0712 00:09:31.077902 3029 apiserver.go:52] "Watching apiserver" Jul 12 00:09:31.202743 kubelet[3029]: I0712 00:09:31.202687 3029 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 12 00:09:31.288135 kubelet[3029]: I0712 00:09:31.286145 3029 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-29-62" Jul 12 00:09:31.288135 kubelet[3029]: E0712 00:09:31.286203 3029 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ip-172-31-29-62\": node \"ip-172-31-29-62\" not found" Jul 12 00:09:33.357989 systemd[1]: Reloading requested from client PID 3305 ('systemctl') (unit session-7.scope)... Jul 12 00:09:33.358022 systemd[1]: Reloading... Jul 12 00:09:33.519143 zram_generator::config[3348]: No configuration found. Jul 12 00:09:33.779987 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 12 00:09:34.004323 systemd[1]: Reloading finished in 645 ms. Jul 12 00:09:34.070665 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:09:34.087950 systemd[1]: kubelet.service: Deactivated successfully. Jul 12 00:09:34.088894 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:34.098663 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 12 00:09:34.215165 update_engine[2118]: I20250712 00:09:34.214820 2118 update_attempter.cc:509] Updating boot flags... Jul 12 00:09:34.333191 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 39 scanned by (udev-worker) (3421) Jul 12 00:09:34.678450 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 12 00:09:34.704808 (kubelet)[3513]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 12 00:09:34.768170 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 39 scanned by (udev-worker) (3427) Jul 12 00:09:34.880515 sudo[3547]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Jul 12 00:09:34.884437 sudo[3547]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Jul 12 00:09:34.888936 kubelet[3513]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 00:09:34.893725 kubelet[3513]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 12 00:09:34.893725 kubelet[3513]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 12 00:09:34.893725 kubelet[3513]: I0712 00:09:34.892421 3513 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 12 00:09:34.929008 kubelet[3513]: I0712 00:09:34.928440 3513 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 12 00:09:34.929008 kubelet[3513]: I0712 00:09:34.928489 3513 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 12 00:09:34.929391 kubelet[3513]: I0712 00:09:34.929365 3513 server.go:934] "Client rotation is on, will bootstrap in background" Jul 12 00:09:34.936145 kubelet[3513]: I0712 00:09:34.935059 3513 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jul 12 00:09:34.945213 kubelet[3513]: I0712 00:09:34.944426 3513 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 12 00:09:34.955896 kubelet[3513]: E0712 00:09:34.955835 3513 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jul 12 00:09:34.957519 kubelet[3513]: I0712 00:09:34.956044 3513 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jul 12 00:09:34.965985 kubelet[3513]: I0712 00:09:34.965948 3513 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 12 00:09:34.967074 kubelet[3513]: I0712 00:09:34.967046 3513 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 12 00:09:34.967958 kubelet[3513]: I0712 00:09:34.967904 3513 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 12 00:09:34.968535 kubelet[3513]: I0712 00:09:34.968065 3513 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-29-62","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} Jul 12 00:09:34.969549 kubelet[3513]: I0712 00:09:34.969274 3513 topology_manager.go:138] "Creating topology manager with none policy" Jul 12 00:09:34.969549 kubelet[3513]: I0712 00:09:34.969310 3513 container_manager_linux.go:300] "Creating device plugin manager" Jul 12 00:09:34.969549 kubelet[3513]: I0712 00:09:34.969379 3513 state_mem.go:36] "Initialized new in-memory state store" Jul 12 00:09:34.969942 kubelet[3513]: I0712 00:09:34.969921 3513 kubelet.go:408] "Attempting to sync node with API server" Jul 12 00:09:34.973153 kubelet[3513]: I0712 00:09:34.971340 3513 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 12 00:09:34.973153 kubelet[3513]: I0712 00:09:34.971399 3513 kubelet.go:314] "Adding apiserver pod source" Jul 12 00:09:34.973153 kubelet[3513]: I0712 00:09:34.971431 3513 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 12 00:09:34.978923 kubelet[3513]: I0712 00:09:34.978886 3513 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jul 12 00:09:34.980092 kubelet[3513]: I0712 00:09:34.979925 3513 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 12 00:09:34.983009 kubelet[3513]: I0712 00:09:34.982977 3513 server.go:1274] "Started kubelet" Jul 12 00:09:35.033569 kubelet[3513]: I0712 00:09:35.033529 3513 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 12 00:09:35.044359 kubelet[3513]: I0712 00:09:35.044035 3513 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 12 00:09:35.089297 kubelet[3513]: I0712 00:09:35.088985 3513 server.go:449] "Adding debug handlers to kubelet server" Jul 12 00:09:35.107617 kubelet[3513]: I0712 00:09:35.050686 3513 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 12 00:09:35.128357 kubelet[3513]: I0712 00:09:35.045539 3513 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 12 00:09:35.129017 kubelet[3513]: I0712 00:09:35.128857 3513 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 12 00:09:35.153671 kubelet[3513]: E0712 00:09:35.058187 3513 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-29-62\" not found" Jul 12 00:09:35.153671 kubelet[3513]: I0712 00:09:35.058015 3513 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 12 00:09:35.166828 kubelet[3513]: I0712 00:09:35.057992 3513 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 12 00:09:35.195844 kubelet[3513]: I0712 00:09:35.152988 3513 factory.go:221] Registration of the systemd container factory successfully Jul 12 00:09:35.196087 kubelet[3513]: I0712 00:09:35.195895 3513 reconciler.go:26] "Reconciler: start to sync state" Jul 12 00:09:35.203817 kubelet[3513]: I0712 00:09:35.201498 3513 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 12 00:09:35.243904 kubelet[3513]: I0712 00:09:35.240752 3513 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 12 00:09:35.246995 kubelet[3513]: I0712 00:09:35.245365 3513 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 12 00:09:35.246995 kubelet[3513]: I0712 00:09:35.245406 3513 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 12 00:09:35.246995 kubelet[3513]: I0712 00:09:35.245435 3513 kubelet.go:2321] "Starting kubelet main sync loop" Jul 12 00:09:35.246995 kubelet[3513]: E0712 00:09:35.245508 3513 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 12 00:09:35.264241 kubelet[3513]: I0712 00:09:35.260689 3513 factory.go:221] Registration of the containerd container factory successfully Jul 12 00:09:35.303755 kubelet[3513]: E0712 00:09:35.302786 3513 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 12 00:09:35.353595 kubelet[3513]: E0712 00:09:35.352862 3513 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 12 00:09:35.502987 kubelet[3513]: I0712 00:09:35.502749 3513 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 12 00:09:35.502987 kubelet[3513]: I0712 00:09:35.502788 3513 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 12 00:09:35.502987 kubelet[3513]: I0712 00:09:35.502825 3513 state_mem.go:36] "Initialized new in-memory state store" Jul 12 00:09:35.503271 kubelet[3513]: I0712 00:09:35.503071 3513 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jul 12 00:09:35.503271 kubelet[3513]: I0712 00:09:35.503093 3513 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jul 12 00:09:35.503271 kubelet[3513]: I0712 00:09:35.503175 3513 policy_none.go:49] "None policy: Start" Jul 12 00:09:35.504897 kubelet[3513]: I0712 00:09:35.504523 3513 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 12 00:09:35.504897 kubelet[3513]: I0712 00:09:35.504572 3513 state_mem.go:35] "Initializing new in-memory state store" Jul 12 00:09:35.504897 kubelet[3513]: I0712 00:09:35.504821 3513 state_mem.go:75] "Updated machine memory state" Jul 12 00:09:35.508758 kubelet[3513]: I0712 00:09:35.508373 3513 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 12 00:09:35.508758 kubelet[3513]: I0712 00:09:35.508692 3513 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 12 00:09:35.508758 kubelet[3513]: I0712 00:09:35.508717 3513 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 12 00:09:35.520261 kubelet[3513]: I0712 00:09:35.520215 3513 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 12 00:09:35.579531 kubelet[3513]: E0712 00:09:35.579306 3513 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ip-172-31-29-62\" already exists" pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:35.582878 kubelet[3513]: E0712 00:09:35.582834 3513 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ip-172-31-29-62\" already exists" pod="kube-system/kube-scheduler-ip-172-31-29-62" Jul 12 00:09:35.637420 kubelet[3513]: I0712 00:09:35.636466 3513 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-29-62" Jul 12 00:09:35.652377 kubelet[3513]: I0712 00:09:35.652210 3513 kubelet_node_status.go:111] "Node was previously registered" node="ip-172-31-29-62" Jul 12 00:09:35.653814 kubelet[3513]: I0712 00:09:35.652837 3513 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-29-62" Jul 12 00:09:35.703799 kubelet[3513]: I0712 00:09:35.703735 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/830d67704643f783ed5af670e9b72fca-k8s-certs\") pod \"kube-apiserver-ip-172-31-29-62\" (UID: \"830d67704643f783ed5af670e9b72fca\") " pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:35.703799 kubelet[3513]: I0712 00:09:35.703803 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-kubeconfig\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:35.703982 kubelet[3513]: I0712 00:09:35.703850 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:35.703982 kubelet[3513]: I0712 00:09:35.703893 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a4fbd316186c0850039629540f818287-kubeconfig\") pod \"kube-scheduler-ip-172-31-29-62\" (UID: \"a4fbd316186c0850039629540f818287\") " pod="kube-system/kube-scheduler-ip-172-31-29-62" Jul 12 00:09:35.703982 kubelet[3513]: I0712 00:09:35.703932 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/830d67704643f783ed5af670e9b72fca-ca-certs\") pod \"kube-apiserver-ip-172-31-29-62\" (UID: \"830d67704643f783ed5af670e9b72fca\") " pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:35.703982 kubelet[3513]: I0712 00:09:35.703968 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/830d67704643f783ed5af670e9b72fca-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-29-62\" (UID: \"830d67704643f783ed5af670e9b72fca\") " pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:35.704233 kubelet[3513]: I0712 00:09:35.704008 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-ca-certs\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:35.704233 kubelet[3513]: I0712 00:09:35.704043 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:35.704233 kubelet[3513]: I0712 00:09:35.704078 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/494476933b0c0f1b79178bb9c91c0455-k8s-certs\") pod \"kube-controller-manager-ip-172-31-29-62\" (UID: \"494476933b0c0f1b79178bb9c91c0455\") " pod="kube-system/kube-controller-manager-ip-172-31-29-62" Jul 12 00:09:35.976908 kubelet[3513]: I0712 00:09:35.976517 3513 apiserver.go:52] "Watching apiserver" Jul 12 00:09:36.015577 sudo[3547]: pam_unix(sudo:session): session closed for user root Jul 12 00:09:36.053671 kubelet[3513]: I0712 00:09:36.053586 3513 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 12 00:09:36.220217 kubelet[3513]: I0712 00:09:36.220029 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-29-62" podStartSLOduration=4.220005739 podStartE2EDuration="4.220005739s" podCreationTimestamp="2025-07-12 00:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 00:09:36.200506327 +0000 UTC m=+1.483507460" watchObservedRunningTime="2025-07-12 00:09:36.220005739 +0000 UTC m=+1.503006848" Jul 12 00:09:36.242158 kubelet[3513]: I0712 00:09:36.239767 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-29-62" podStartSLOduration=1.239745283 podStartE2EDuration="1.239745283s" podCreationTimestamp="2025-07-12 00:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 00:09:36.239261887 +0000 UTC m=+1.522263020" watchObservedRunningTime="2025-07-12 00:09:36.239745283 +0000 UTC m=+1.522746548" Jul 12 00:09:36.242158 kubelet[3513]: I0712 00:09:36.239894 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-29-62" podStartSLOduration=3.239885215 podStartE2EDuration="3.239885215s" podCreationTimestamp="2025-07-12 00:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 00:09:36.221323471 +0000 UTC m=+1.504324616" watchObservedRunningTime="2025-07-12 00:09:36.239885215 +0000 UTC m=+1.522886348" Jul 12 00:09:36.400251 kubelet[3513]: E0712 00:09:36.399922 3513 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ip-172-31-29-62\" already exists" pod="kube-system/kube-apiserver-ip-172-31-29-62" Jul 12 00:09:38.538388 kubelet[3513]: I0712 00:09:38.538211 3513 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jul 12 00:09:38.539549 containerd[2160]: time="2025-07-12T00:09:38.539490851Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jul 12 00:09:38.540130 kubelet[3513]: I0712 00:09:38.539853 3513 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jul 12 00:09:39.400977 sudo[2502]: pam_unix(sudo:session): session closed for user root Jul 12 00:09:39.425014 sshd[2498]: pam_unix(sshd:session): session closed for user core Jul 12 00:09:39.448184 systemd[1]: sshd@6-172.31.29.62:22-139.178.89.65:48546.service: Deactivated successfully. Jul 12 00:09:39.455222 systemd-logind[2117]: Session 7 logged out. Waiting for processes to exit. Jul 12 00:09:39.467166 systemd[1]: session-7.scope: Deactivated successfully. Jul 12 00:09:39.481510 systemd-logind[2117]: Removed session 7. Jul 12 00:09:39.536491 kubelet[3513]: I0712 00:09:39.536413 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-config-path\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536491 kubelet[3513]: I0712 00:09:39.536491 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-net\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536670 kubelet[3513]: I0712 00:09:39.536537 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-hubble-tls\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536670 kubelet[3513]: I0712 00:09:39.536573 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d79acfae-9177-49b3-b7be-36aa9a525506-lib-modules\") pod \"kube-proxy-wpwxx\" (UID: \"d79acfae-9177-49b3-b7be-36aa9a525506\") " pod="kube-system/kube-proxy-wpwxx" Jul 12 00:09:39.536670 kubelet[3513]: I0712 00:09:39.536610 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-lib-modules\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536670 kubelet[3513]: I0712 00:09:39.536643 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-xtables-lock\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536922 kubelet[3513]: I0712 00:09:39.536676 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d79acfae-9177-49b3-b7be-36aa9a525506-xtables-lock\") pod \"kube-proxy-wpwxx\" (UID: \"d79acfae-9177-49b3-b7be-36aa9a525506\") " pod="kube-system/kube-proxy-wpwxx" Jul 12 00:09:39.536922 kubelet[3513]: I0712 00:09:39.536710 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcdx2\" (UniqueName: \"kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-kube-api-access-fcdx2\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536922 kubelet[3513]: I0712 00:09:39.536746 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-run\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536922 kubelet[3513]: I0712 00:09:39.536797 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-cgroup\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.536922 kubelet[3513]: I0712 00:09:39.536831 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cni-path\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.537941 kubelet[3513]: I0712 00:09:39.536864 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8t9x\" (UniqueName: \"kubernetes.io/projected/d79acfae-9177-49b3-b7be-36aa9a525506-kube-api-access-p8t9x\") pod \"kube-proxy-wpwxx\" (UID: \"d79acfae-9177-49b3-b7be-36aa9a525506\") " pod="kube-system/kube-proxy-wpwxx" Jul 12 00:09:39.537941 kubelet[3513]: I0712 00:09:39.536900 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-bpf-maps\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.537941 kubelet[3513]: I0712 00:09:39.536966 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-hostproc\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.537941 kubelet[3513]: I0712 00:09:39.537002 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/69c79161-e4af-41d8-a2b5-6c833216baa7-clustermesh-secrets\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.537941 kubelet[3513]: I0712 00:09:39.537038 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/d79acfae-9177-49b3-b7be-36aa9a525506-kube-proxy\") pod \"kube-proxy-wpwxx\" (UID: \"d79acfae-9177-49b3-b7be-36aa9a525506\") " pod="kube-system/kube-proxy-wpwxx" Jul 12 00:09:39.537941 kubelet[3513]: I0712 00:09:39.537071 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-etc-cni-netd\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.538361 kubelet[3513]: I0712 00:09:39.537141 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-kernel\") pod \"cilium-wgxbq\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " pod="kube-system/cilium-wgxbq" Jul 12 00:09:39.743472 kubelet[3513]: I0712 00:09:39.742073 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0ab6073-b03c-47f9-a2b2-b287a5841f36-cilium-config-path\") pod \"cilium-operator-5d85765b45-trhq6\" (UID: \"c0ab6073-b03c-47f9-a2b2-b287a5841f36\") " pod="kube-system/cilium-operator-5d85765b45-trhq6" Jul 12 00:09:39.746965 kubelet[3513]: I0712 00:09:39.743263 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkbq8\" (UniqueName: \"kubernetes.io/projected/c0ab6073-b03c-47f9-a2b2-b287a5841f36-kube-api-access-pkbq8\") pod \"cilium-operator-5d85765b45-trhq6\" (UID: \"c0ab6073-b03c-47f9-a2b2-b287a5841f36\") " pod="kube-system/cilium-operator-5d85765b45-trhq6" Jul 12 00:09:39.795643 containerd[2160]: time="2025-07-12T00:09:39.795562801Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wpwxx,Uid:d79acfae-9177-49b3-b7be-36aa9a525506,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:39.801791 containerd[2160]: time="2025-07-12T00:09:39.801707509Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wgxbq,Uid:69c79161-e4af-41d8-a2b5-6c833216baa7,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:39.862908 containerd[2160]: time="2025-07-12T00:09:39.860302717Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:09:39.862908 containerd[2160]: time="2025-07-12T00:09:39.860403097Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:09:39.862908 containerd[2160]: time="2025-07-12T00:09:39.860439829Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:39.862908 containerd[2160]: time="2025-07-12T00:09:39.860677573Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:39.900312 containerd[2160]: time="2025-07-12T00:09:39.899826589Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:09:39.900312 containerd[2160]: time="2025-07-12T00:09:39.899926249Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:09:39.900312 containerd[2160]: time="2025-07-12T00:09:39.899953105Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:39.900312 containerd[2160]: time="2025-07-12T00:09:39.900183913Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:39.950669 containerd[2160]: time="2025-07-12T00:09:39.950309798Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-trhq6,Uid:c0ab6073-b03c-47f9-a2b2-b287a5841f36,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:39.968090 containerd[2160]: time="2025-07-12T00:09:39.968032214Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wpwxx,Uid:d79acfae-9177-49b3-b7be-36aa9a525506,Namespace:kube-system,Attempt:0,} returns sandbox id \"0f279af7c3de134ecd784689c48b722eef47dfc21feee670f14224259aecd09b\"" Jul 12 00:09:39.980152 containerd[2160]: time="2025-07-12T00:09:39.979618958Z" level=info msg="CreateContainer within sandbox \"0f279af7c3de134ecd784689c48b722eef47dfc21feee670f14224259aecd09b\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jul 12 00:09:40.001087 containerd[2160]: time="2025-07-12T00:09:40.000084166Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wgxbq,Uid:69c79161-e4af-41d8-a2b5-6c833216baa7,Namespace:kube-system,Attempt:0,} returns sandbox id \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\"" Jul 12 00:09:40.006338 containerd[2160]: time="2025-07-12T00:09:40.005677870Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Jul 12 00:09:40.021484 containerd[2160]: time="2025-07-12T00:09:40.021220966Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:09:40.021484 containerd[2160]: time="2025-07-12T00:09:40.021304138Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:09:40.021484 containerd[2160]: time="2025-07-12T00:09:40.021329074Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:40.021915 containerd[2160]: time="2025-07-12T00:09:40.021821806Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:09:40.022703 containerd[2160]: time="2025-07-12T00:09:40.022632658Z" level=info msg="CreateContainer within sandbox \"0f279af7c3de134ecd784689c48b722eef47dfc21feee670f14224259aecd09b\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8c4bc6b523c794938d5fb7d87c0ea08601ff98e45c13e45b97f34f29f706efb7\"" Jul 12 00:09:40.025697 containerd[2160]: time="2025-07-12T00:09:40.023957254Z" level=info msg="StartContainer for \"8c4bc6b523c794938d5fb7d87c0ea08601ff98e45c13e45b97f34f29f706efb7\"" Jul 12 00:09:40.157222 containerd[2160]: time="2025-07-12T00:09:40.157052291Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-trhq6,Uid:c0ab6073-b03c-47f9-a2b2-b287a5841f36,Namespace:kube-system,Attempt:0,} returns sandbox id \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\"" Jul 12 00:09:40.172500 containerd[2160]: time="2025-07-12T00:09:40.172351655Z" level=info msg="StartContainer for \"8c4bc6b523c794938d5fb7d87c0ea08601ff98e45c13e45b97f34f29f706efb7\" returns successfully" Jul 12 00:09:43.554756 kubelet[3513]: I0712 00:09:43.554583 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-wpwxx" podStartSLOduration=4.554559231 podStartE2EDuration="4.554559231s" podCreationTimestamp="2025-07-12 00:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 00:09:40.431229804 +0000 UTC m=+5.714230961" watchObservedRunningTime="2025-07-12 00:09:43.554559231 +0000 UTC m=+8.837560460" Jul 12 00:09:45.032189 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2991233206.mount: Deactivated successfully. Jul 12 00:09:47.744561 containerd[2160]: time="2025-07-12T00:09:47.744474608Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:47.746474 containerd[2160]: time="2025-07-12T00:09:47.746370116Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" Jul 12 00:09:47.749470 containerd[2160]: time="2025-07-12T00:09:47.749333084Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:47.752632 containerd[2160]: time="2025-07-12T00:09:47.752316716Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 7.746450794s" Jul 12 00:09:47.752632 containerd[2160]: time="2025-07-12T00:09:47.752387180Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Jul 12 00:09:47.755886 containerd[2160]: time="2025-07-12T00:09:47.755819696Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Jul 12 00:09:47.757920 containerd[2160]: time="2025-07-12T00:09:47.757834076Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 12 00:09:47.784064 containerd[2160]: time="2025-07-12T00:09:47.783987356Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\"" Jul 12 00:09:47.786840 containerd[2160]: time="2025-07-12T00:09:47.785775560Z" level=info msg="StartContainer for \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\"" Jul 12 00:09:47.886785 containerd[2160]: time="2025-07-12T00:09:47.886716597Z" level=info msg="StartContainer for \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\" returns successfully" Jul 12 00:09:48.641761 containerd[2160]: time="2025-07-12T00:09:48.641601141Z" level=info msg="shim disconnected" id=49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6 namespace=k8s.io Jul 12 00:09:48.641761 containerd[2160]: time="2025-07-12T00:09:48.641675529Z" level=warning msg="cleaning up after shim disconnected" id=49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6 namespace=k8s.io Jul 12 00:09:48.641761 containerd[2160]: time="2025-07-12T00:09:48.641696301Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:09:48.774420 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6-rootfs.mount: Deactivated successfully. Jul 12 00:09:49.250373 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2408467089.mount: Deactivated successfully. Jul 12 00:09:49.460947 containerd[2160]: time="2025-07-12T00:09:49.458307429Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 12 00:09:49.521154 containerd[2160]: time="2025-07-12T00:09:49.519950289Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\"" Jul 12 00:09:49.525029 containerd[2160]: time="2025-07-12T00:09:49.520995909Z" level=info msg="StartContainer for \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\"" Jul 12 00:09:49.652219 containerd[2160]: time="2025-07-12T00:09:49.652076734Z" level=info msg="StartContainer for \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\" returns successfully" Jul 12 00:09:49.681392 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 12 00:09:49.682027 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 12 00:09:49.684011 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Jul 12 00:09:49.695301 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 12 00:09:49.757247 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 12 00:09:49.794495 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90-rootfs.mount: Deactivated successfully. Jul 12 00:09:49.847721 containerd[2160]: time="2025-07-12T00:09:49.847254299Z" level=info msg="shim disconnected" id=e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90 namespace=k8s.io Jul 12 00:09:49.847721 containerd[2160]: time="2025-07-12T00:09:49.847335887Z" level=warning msg="cleaning up after shim disconnected" id=e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90 namespace=k8s.io Jul 12 00:09:49.847721 containerd[2160]: time="2025-07-12T00:09:49.847358387Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:09:50.165016 containerd[2160]: time="2025-07-12T00:09:50.164671352Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:50.167064 containerd[2160]: time="2025-07-12T00:09:50.166968416Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" Jul 12 00:09:50.169189 containerd[2160]: time="2025-07-12T00:09:50.169128896Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 12 00:09:50.173402 containerd[2160]: time="2025-07-12T00:09:50.172412120Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 2.416525932s" Jul 12 00:09:50.173402 containerd[2160]: time="2025-07-12T00:09:50.172495904Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Jul 12 00:09:50.177668 containerd[2160]: time="2025-07-12T00:09:50.177607676Z" level=info msg="CreateContainer within sandbox \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Jul 12 00:09:50.201455 containerd[2160]: time="2025-07-12T00:09:50.201374720Z" level=info msg="CreateContainer within sandbox \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\"" Jul 12 00:09:50.203058 containerd[2160]: time="2025-07-12T00:09:50.202075892Z" level=info msg="StartContainer for \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\"" Jul 12 00:09:50.291358 containerd[2160]: time="2025-07-12T00:09:50.291222021Z" level=info msg="StartContainer for \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\" returns successfully" Jul 12 00:09:50.479215 containerd[2160]: time="2025-07-12T00:09:50.478611598Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 12 00:09:50.493731 kubelet[3513]: I0712 00:09:50.491566 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-trhq6" podStartSLOduration=1.477919033 podStartE2EDuration="11.491539546s" podCreationTimestamp="2025-07-12 00:09:39 +0000 UTC" firstStartedPulling="2025-07-12 00:09:40.160016411 +0000 UTC m=+5.443017520" lastFinishedPulling="2025-07-12 00:09:50.173636924 +0000 UTC m=+15.456638033" observedRunningTime="2025-07-12 00:09:50.48877015 +0000 UTC m=+15.771771271" watchObservedRunningTime="2025-07-12 00:09:50.491539546 +0000 UTC m=+15.774540655" Jul 12 00:09:50.528441 containerd[2160]: time="2025-07-12T00:09:50.528382606Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\"" Jul 12 00:09:50.531410 containerd[2160]: time="2025-07-12T00:09:50.529531546Z" level=info msg="StartContainer for \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\"" Jul 12 00:09:50.840042 containerd[2160]: time="2025-07-12T00:09:50.839869128Z" level=info msg="StartContainer for \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\" returns successfully" Jul 12 00:09:50.947367 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a-rootfs.mount: Deactivated successfully. Jul 12 00:09:50.997527 containerd[2160]: time="2025-07-12T00:09:50.997427868Z" level=info msg="shim disconnected" id=b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a namespace=k8s.io Jul 12 00:09:50.997527 containerd[2160]: time="2025-07-12T00:09:50.997511760Z" level=warning msg="cleaning up after shim disconnected" id=b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a namespace=k8s.io Jul 12 00:09:50.997840 containerd[2160]: time="2025-07-12T00:09:50.997535244Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:09:51.497126 containerd[2160]: time="2025-07-12T00:09:51.495135947Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 12 00:09:51.539160 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount516941476.mount: Deactivated successfully. Jul 12 00:09:51.548753 containerd[2160]: time="2025-07-12T00:09:51.548458163Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\"" Jul 12 00:09:51.558858 containerd[2160]: time="2025-07-12T00:09:51.558472763Z" level=info msg="StartContainer for \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\"" Jul 12 00:09:51.879412 containerd[2160]: time="2025-07-12T00:09:51.879335125Z" level=info msg="StartContainer for \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\" returns successfully" Jul 12 00:09:51.968665 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977-rootfs.mount: Deactivated successfully. Jul 12 00:09:51.976861 containerd[2160]: time="2025-07-12T00:09:51.976642165Z" level=info msg="shim disconnected" id=953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977 namespace=k8s.io Jul 12 00:09:51.978381 containerd[2160]: time="2025-07-12T00:09:51.976851997Z" level=warning msg="cleaning up after shim disconnected" id=953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977 namespace=k8s.io Jul 12 00:09:51.978536 containerd[2160]: time="2025-07-12T00:09:51.978375337Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:09:52.502482 containerd[2160]: time="2025-07-12T00:09:52.502326000Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 12 00:09:52.537618 containerd[2160]: time="2025-07-12T00:09:52.537468864Z" level=info msg="CreateContainer within sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\"" Jul 12 00:09:52.538809 containerd[2160]: time="2025-07-12T00:09:52.538560276Z" level=info msg="StartContainer for \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\"" Jul 12 00:09:52.681167 containerd[2160]: time="2025-07-12T00:09:52.680962969Z" level=info msg="StartContainer for \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\" returns successfully" Jul 12 00:09:52.978444 kubelet[3513]: I0712 00:09:52.977513 3513 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Jul 12 00:09:53.163835 kubelet[3513]: I0712 00:09:53.163534 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5334433-385a-4c56-8a64-5f60f41ca052-config-volume\") pod \"coredns-7c65d6cfc9-l2fm7\" (UID: \"f5334433-385a-4c56-8a64-5f60f41ca052\") " pod="kube-system/coredns-7c65d6cfc9-l2fm7" Jul 12 00:09:53.163835 kubelet[3513]: I0712 00:09:53.163612 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5998j\" (UniqueName: \"kubernetes.io/projected/f5334433-385a-4c56-8a64-5f60f41ca052-kube-api-access-5998j\") pod \"coredns-7c65d6cfc9-l2fm7\" (UID: \"f5334433-385a-4c56-8a64-5f60f41ca052\") " pod="kube-system/coredns-7c65d6cfc9-l2fm7" Jul 12 00:09:53.163835 kubelet[3513]: I0712 00:09:53.163662 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/36a2150d-1e4a-4860-9b9b-8e0d28cde3c0-config-volume\") pod \"coredns-7c65d6cfc9-8frgg\" (UID: \"36a2150d-1e4a-4860-9b9b-8e0d28cde3c0\") " pod="kube-system/coredns-7c65d6cfc9-8frgg" Jul 12 00:09:53.163835 kubelet[3513]: I0712 00:09:53.163704 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdghb\" (UniqueName: \"kubernetes.io/projected/36a2150d-1e4a-4860-9b9b-8e0d28cde3c0-kube-api-access-bdghb\") pod \"coredns-7c65d6cfc9-8frgg\" (UID: \"36a2150d-1e4a-4860-9b9b-8e0d28cde3c0\") " pod="kube-system/coredns-7c65d6cfc9-8frgg" Jul 12 00:09:53.345584 containerd[2160]: time="2025-07-12T00:09:53.344874324Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-8frgg,Uid:36a2150d-1e4a-4860-9b9b-8e0d28cde3c0,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:53.376180 containerd[2160]: time="2025-07-12T00:09:53.375580920Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-l2fm7,Uid:f5334433-385a-4c56-8a64-5f60f41ca052,Namespace:kube-system,Attempt:0,}" Jul 12 00:09:53.604079 kubelet[3513]: I0712 00:09:53.601503 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-wgxbq" podStartSLOduration=6.850160143 podStartE2EDuration="14.601477165s" podCreationTimestamp="2025-07-12 00:09:39 +0000 UTC" firstStartedPulling="2025-07-12 00:09:40.003217606 +0000 UTC m=+5.286218727" lastFinishedPulling="2025-07-12 00:09:47.754534652 +0000 UTC m=+13.037535749" observedRunningTime="2025-07-12 00:09:53.597877621 +0000 UTC m=+18.880878814" watchObservedRunningTime="2025-07-12 00:09:53.601477165 +0000 UTC m=+18.884478262" Jul 12 00:09:55.770400 systemd-networkd[1694]: cilium_host: Link UP Jul 12 00:09:55.771857 systemd-networkd[1694]: cilium_net: Link UP Jul 12 00:09:55.772759 (udev-worker)[4390]: Network interface NamePolicy= disabled on kernel command line. Jul 12 00:09:55.773496 systemd-networkd[1694]: cilium_net: Gained carrier Jul 12 00:09:55.773902 systemd-networkd[1694]: cilium_host: Gained carrier Jul 12 00:09:55.777296 (udev-worker)[4391]: Network interface NamePolicy= disabled on kernel command line. Jul 12 00:09:55.960137 systemd-networkd[1694]: cilium_vxlan: Link UP Jul 12 00:09:55.960157 systemd-networkd[1694]: cilium_vxlan: Gained carrier Jul 12 00:09:56.208778 systemd-networkd[1694]: cilium_host: Gained IPv6LL Jul 12 00:09:56.456455 systemd-networkd[1694]: cilium_net: Gained IPv6LL Jul 12 00:09:56.527146 kernel: NET: Registered PF_ALG protocol family Jul 12 00:09:57.845952 systemd-networkd[1694]: lxc_health: Link UP Jul 12 00:09:57.858438 systemd-networkd[1694]: lxc_health: Gained carrier Jul 12 00:09:57.859013 (udev-worker)[4435]: Network interface NamePolicy= disabled on kernel command line. Jul 12 00:09:57.928385 systemd-networkd[1694]: cilium_vxlan: Gained IPv6LL Jul 12 00:09:58.536541 systemd-networkd[1694]: lxc8210f5d80788: Link UP Jul 12 00:09:58.552512 kernel: eth0: renamed from tmp50d20 Jul 12 00:09:58.566358 systemd-networkd[1694]: lxc8210f5d80788: Gained carrier Jul 12 00:09:58.589296 systemd-networkd[1694]: lxc09d65d24a424: Link UP Jul 12 00:09:58.616152 kernel: eth0: renamed from tmpeeae2 Jul 12 00:09:58.631478 systemd-networkd[1694]: lxc09d65d24a424: Gained carrier Jul 12 00:09:59.784294 systemd-networkd[1694]: lxc_health: Gained IPv6LL Jul 12 00:10:00.232385 systemd-networkd[1694]: lxc09d65d24a424: Gained IPv6LL Jul 12 00:10:00.296384 systemd-networkd[1694]: lxc8210f5d80788: Gained IPv6LL Jul 12 00:10:03.047047 ntpd[2101]: Listen normally on 6 cilium_host 192.168.0.244:123 Jul 12 00:10:03.048419 ntpd[2101]: 12 Jul 00:10:03 ntpd[2101]: Listen normally on 6 cilium_host 192.168.0.244:123 Jul 12 00:10:03.048419 ntpd[2101]: 12 Jul 00:10:03 ntpd[2101]: Listen normally on 7 cilium_net [fe80::c42c:c8ff:fe63:6dfc%4]:123 Jul 12 00:10:03.048419 ntpd[2101]: 12 Jul 00:10:03 ntpd[2101]: Listen normally on 8 cilium_host [fe80::b0b3:26ff:fedc:20aa%5]:123 Jul 12 00:10:03.048419 ntpd[2101]: 12 Jul 00:10:03 ntpd[2101]: Listen normally on 9 cilium_vxlan [fe80::ac6e:36ff:fef6:897b%6]:123 Jul 12 00:10:03.048419 ntpd[2101]: 12 Jul 00:10:03 ntpd[2101]: Listen normally on 10 lxc_health [fe80::44be:10ff:fe27:1e0c%8]:123 Jul 12 00:10:03.048419 ntpd[2101]: 12 Jul 00:10:03 ntpd[2101]: Listen normally on 11 lxc8210f5d80788 [fe80::cde:e0ff:fed4:8a58%10]:123 Jul 12 00:10:03.048419 ntpd[2101]: 12 Jul 00:10:03 ntpd[2101]: Listen normally on 12 lxc09d65d24a424 [fe80::482e:6fff:febe:1d6f%12]:123 Jul 12 00:10:03.047278 ntpd[2101]: Listen normally on 7 cilium_net [fe80::c42c:c8ff:fe63:6dfc%4]:123 Jul 12 00:10:03.047364 ntpd[2101]: Listen normally on 8 cilium_host [fe80::b0b3:26ff:fedc:20aa%5]:123 Jul 12 00:10:03.047459 ntpd[2101]: Listen normally on 9 cilium_vxlan [fe80::ac6e:36ff:fef6:897b%6]:123 Jul 12 00:10:03.047529 ntpd[2101]: Listen normally on 10 lxc_health [fe80::44be:10ff:fe27:1e0c%8]:123 Jul 12 00:10:03.047597 ntpd[2101]: Listen normally on 11 lxc8210f5d80788 [fe80::cde:e0ff:fed4:8a58%10]:123 Jul 12 00:10:03.047665 ntpd[2101]: Listen normally on 12 lxc09d65d24a424 [fe80::482e:6fff:febe:1d6f%12]:123 Jul 12 00:10:06.963009 containerd[2160]: time="2025-07-12T00:10:06.962283256Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:10:06.963009 containerd[2160]: time="2025-07-12T00:10:06.962402080Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:10:06.963009 containerd[2160]: time="2025-07-12T00:10:06.962438620Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:10:06.963009 containerd[2160]: time="2025-07-12T00:10:06.962608300Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:10:07.007676 containerd[2160]: time="2025-07-12T00:10:07.004514856Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:10:07.007676 containerd[2160]: time="2025-07-12T00:10:07.006555240Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:10:07.007676 containerd[2160]: time="2025-07-12T00:10:07.006628404Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:10:07.011737 containerd[2160]: time="2025-07-12T00:10:07.008473128Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:10:07.242398 containerd[2160]: time="2025-07-12T00:10:07.241956841Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-l2fm7,Uid:f5334433-385a-4c56-8a64-5f60f41ca052,Namespace:kube-system,Attempt:0,} returns sandbox id \"eeae28c1a302ac01c5b45958938b9afd18eb2505b4952929830c22e0e8510b9f\"" Jul 12 00:10:07.243220 containerd[2160]: time="2025-07-12T00:10:07.242932381Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-8frgg,Uid:36a2150d-1e4a-4860-9b9b-8e0d28cde3c0,Namespace:kube-system,Attempt:0,} returns sandbox id \"50d209fb7d47e612a63b2791a2f7ad5c63430e414cdb3166a2aef809a93fef75\"" Jul 12 00:10:07.271689 containerd[2160]: time="2025-07-12T00:10:07.270074053Z" level=info msg="CreateContainer within sandbox \"50d209fb7d47e612a63b2791a2f7ad5c63430e414cdb3166a2aef809a93fef75\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 12 00:10:07.289267 containerd[2160]: time="2025-07-12T00:10:07.289075765Z" level=info msg="CreateContainer within sandbox \"eeae28c1a302ac01c5b45958938b9afd18eb2505b4952929830c22e0e8510b9f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 12 00:10:07.343729 containerd[2160]: time="2025-07-12T00:10:07.343669430Z" level=info msg="CreateContainer within sandbox \"50d209fb7d47e612a63b2791a2f7ad5c63430e414cdb3166a2aef809a93fef75\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8fbb83fd2b17b2dd6a2c4817a57f7caf75c0560cea67c7de0e26b8eec7e092a5\"" Jul 12 00:10:07.346895 containerd[2160]: time="2025-07-12T00:10:07.345679262Z" level=info msg="StartContainer for \"8fbb83fd2b17b2dd6a2c4817a57f7caf75c0560cea67c7de0e26b8eec7e092a5\"" Jul 12 00:10:07.363698 containerd[2160]: time="2025-07-12T00:10:07.363340226Z" level=info msg="CreateContainer within sandbox \"eeae28c1a302ac01c5b45958938b9afd18eb2505b4952929830c22e0e8510b9f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"fe5fd32dce4ce4c49f919accb30ae60f6ec8c890495769abb57a5378a1ca8d12\"" Jul 12 00:10:07.365668 containerd[2160]: time="2025-07-12T00:10:07.364665074Z" level=info msg="StartContainer for \"fe5fd32dce4ce4c49f919accb30ae60f6ec8c890495769abb57a5378a1ca8d12\"" Jul 12 00:10:07.463659 containerd[2160]: time="2025-07-12T00:10:07.463341938Z" level=info msg="StartContainer for \"8fbb83fd2b17b2dd6a2c4817a57f7caf75c0560cea67c7de0e26b8eec7e092a5\" returns successfully" Jul 12 00:10:07.511382 containerd[2160]: time="2025-07-12T00:10:07.511052426Z" level=info msg="StartContainer for \"fe5fd32dce4ce4c49f919accb30ae60f6ec8c890495769abb57a5378a1ca8d12\" returns successfully" Jul 12 00:10:07.649390 kubelet[3513]: I0712 00:10:07.647984 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-l2fm7" podStartSLOduration=28.647961507 podStartE2EDuration="28.647961507s" podCreationTimestamp="2025-07-12 00:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 00:10:07.644866887 +0000 UTC m=+32.927868068" watchObservedRunningTime="2025-07-12 00:10:07.647961507 +0000 UTC m=+32.930962616" Jul 12 00:10:07.747238 kubelet[3513]: I0712 00:10:07.745630 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-8frgg" podStartSLOduration=28.745609324 podStartE2EDuration="28.745609324s" podCreationTimestamp="2025-07-12 00:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 00:10:07.695090367 +0000 UTC m=+32.978091500" watchObservedRunningTime="2025-07-12 00:10:07.745609324 +0000 UTC m=+33.028610457" Jul 12 00:10:25.151089 systemd[1]: Started sshd@7-172.31.29.62:22-139.178.89.65:56432.service - OpenSSH per-connection server daemon (139.178.89.65:56432). Jul 12 00:10:25.318359 sshd[4961]: Accepted publickey for core from 139.178.89.65 port 56432 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:25.321054 sshd[4961]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:25.328902 systemd-logind[2117]: New session 8 of user core. Jul 12 00:10:25.335601 systemd[1]: Started session-8.scope - Session 8 of User core. Jul 12 00:10:25.627877 sshd[4961]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:25.635714 systemd[1]: sshd@7-172.31.29.62:22-139.178.89.65:56432.service: Deactivated successfully. Jul 12 00:10:25.644067 systemd[1]: session-8.scope: Deactivated successfully. Jul 12 00:10:25.646432 systemd-logind[2117]: Session 8 logged out. Waiting for processes to exit. Jul 12 00:10:25.648901 systemd-logind[2117]: Removed session 8. Jul 12 00:10:30.661600 systemd[1]: Started sshd@8-172.31.29.62:22-139.178.89.65:60840.service - OpenSSH per-connection server daemon (139.178.89.65:60840). Jul 12 00:10:30.833255 sshd[4977]: Accepted publickey for core from 139.178.89.65 port 60840 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:30.836022 sshd[4977]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:30.843650 systemd-logind[2117]: New session 9 of user core. Jul 12 00:10:30.850653 systemd[1]: Started session-9.scope - Session 9 of User core. Jul 12 00:10:31.098507 sshd[4977]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:31.106353 systemd[1]: sshd@8-172.31.29.62:22-139.178.89.65:60840.service: Deactivated successfully. Jul 12 00:10:31.113220 systemd-logind[2117]: Session 9 logged out. Waiting for processes to exit. Jul 12 00:10:31.113426 systemd[1]: session-9.scope: Deactivated successfully. Jul 12 00:10:31.117543 systemd-logind[2117]: Removed session 9. Jul 12 00:10:36.128913 systemd[1]: Started sshd@9-172.31.29.62:22-139.178.89.65:60850.service - OpenSSH per-connection server daemon (139.178.89.65:60850). Jul 12 00:10:36.311733 sshd[4994]: Accepted publickey for core from 139.178.89.65 port 60850 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:36.314634 sshd[4994]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:36.323375 systemd-logind[2117]: New session 10 of user core. Jul 12 00:10:36.328739 systemd[1]: Started session-10.scope - Session 10 of User core. Jul 12 00:10:36.576943 sshd[4994]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:36.583906 systemd[1]: sshd@9-172.31.29.62:22-139.178.89.65:60850.service: Deactivated successfully. Jul 12 00:10:36.591827 systemd[1]: session-10.scope: Deactivated successfully. Jul 12 00:10:36.596764 systemd-logind[2117]: Session 10 logged out. Waiting for processes to exit. Jul 12 00:10:36.599059 systemd-logind[2117]: Removed session 10. Jul 12 00:10:41.607954 systemd[1]: Started sshd@10-172.31.29.62:22-139.178.89.65:51770.service - OpenSSH per-connection server daemon (139.178.89.65:51770). Jul 12 00:10:41.789267 sshd[5011]: Accepted publickey for core from 139.178.89.65 port 51770 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:41.792554 sshd[5011]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:41.802747 systemd-logind[2117]: New session 11 of user core. Jul 12 00:10:41.809700 systemd[1]: Started session-11.scope - Session 11 of User core. Jul 12 00:10:42.051088 sshd[5011]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:42.058544 systemd[1]: sshd@10-172.31.29.62:22-139.178.89.65:51770.service: Deactivated successfully. Jul 12 00:10:42.065695 systemd[1]: session-11.scope: Deactivated successfully. Jul 12 00:10:42.067908 systemd-logind[2117]: Session 11 logged out. Waiting for processes to exit. Jul 12 00:10:42.071545 systemd-logind[2117]: Removed session 11. Jul 12 00:10:42.082666 systemd[1]: Started sshd@11-172.31.29.62:22-139.178.89.65:51786.service - OpenSSH per-connection server daemon (139.178.89.65:51786). Jul 12 00:10:42.268843 sshd[5025]: Accepted publickey for core from 139.178.89.65 port 51786 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:42.271526 sshd[5025]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:42.279358 systemd-logind[2117]: New session 12 of user core. Jul 12 00:10:42.288750 systemd[1]: Started session-12.scope - Session 12 of User core. Jul 12 00:10:42.617396 sshd[5025]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:42.633997 systemd[1]: sshd@11-172.31.29.62:22-139.178.89.65:51786.service: Deactivated successfully. Jul 12 00:10:42.646856 systemd[1]: session-12.scope: Deactivated successfully. Jul 12 00:10:42.647369 systemd-logind[2117]: Session 12 logged out. Waiting for processes to exit. Jul 12 00:10:42.676313 systemd[1]: Started sshd@12-172.31.29.62:22-139.178.89.65:51790.service - OpenSSH per-connection server daemon (139.178.89.65:51790). Jul 12 00:10:42.678199 systemd-logind[2117]: Removed session 12. Jul 12 00:10:42.868161 sshd[5037]: Accepted publickey for core from 139.178.89.65 port 51790 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:42.871257 sshd[5037]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:42.879744 systemd-logind[2117]: New session 13 of user core. Jul 12 00:10:42.888764 systemd[1]: Started session-13.scope - Session 13 of User core. Jul 12 00:10:43.130044 sshd[5037]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:43.138436 systemd[1]: sshd@12-172.31.29.62:22-139.178.89.65:51790.service: Deactivated successfully. Jul 12 00:10:43.148158 systemd[1]: session-13.scope: Deactivated successfully. Jul 12 00:10:43.150042 systemd-logind[2117]: Session 13 logged out. Waiting for processes to exit. Jul 12 00:10:43.154537 systemd-logind[2117]: Removed session 13. Jul 12 00:10:48.164620 systemd[1]: Started sshd@13-172.31.29.62:22-139.178.89.65:51792.service - OpenSSH per-connection server daemon (139.178.89.65:51792). Jul 12 00:10:48.338092 sshd[5051]: Accepted publickey for core from 139.178.89.65 port 51792 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:48.340786 sshd[5051]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:48.350162 systemd-logind[2117]: New session 14 of user core. Jul 12 00:10:48.355814 systemd[1]: Started session-14.scope - Session 14 of User core. Jul 12 00:10:48.600319 sshd[5051]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:48.605252 systemd[1]: sshd@13-172.31.29.62:22-139.178.89.65:51792.service: Deactivated successfully. Jul 12 00:10:48.613916 systemd[1]: session-14.scope: Deactivated successfully. Jul 12 00:10:48.614265 systemd-logind[2117]: Session 14 logged out. Waiting for processes to exit. Jul 12 00:10:48.616908 systemd-logind[2117]: Removed session 14. Jul 12 00:10:53.629582 systemd[1]: Started sshd@14-172.31.29.62:22-139.178.89.65:59670.service - OpenSSH per-connection server daemon (139.178.89.65:59670). Jul 12 00:10:53.814496 sshd[5066]: Accepted publickey for core from 139.178.89.65 port 59670 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:53.817407 sshd[5066]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:53.825441 systemd-logind[2117]: New session 15 of user core. Jul 12 00:10:53.833738 systemd[1]: Started session-15.scope - Session 15 of User core. Jul 12 00:10:54.083359 sshd[5066]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:54.090121 systemd[1]: sshd@14-172.31.29.62:22-139.178.89.65:59670.service: Deactivated successfully. Jul 12 00:10:54.097347 systemd[1]: session-15.scope: Deactivated successfully. Jul 12 00:10:54.098300 systemd-logind[2117]: Session 15 logged out. Waiting for processes to exit. Jul 12 00:10:54.101596 systemd-logind[2117]: Removed session 15. Jul 12 00:10:59.113658 systemd[1]: Started sshd@15-172.31.29.62:22-139.178.89.65:59680.service - OpenSSH per-connection server daemon (139.178.89.65:59680). Jul 12 00:10:59.291570 sshd[5081]: Accepted publickey for core from 139.178.89.65 port 59680 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:10:59.294510 sshd[5081]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:10:59.303196 systemd-logind[2117]: New session 16 of user core. Jul 12 00:10:59.308678 systemd[1]: Started session-16.scope - Session 16 of User core. Jul 12 00:10:59.568592 sshd[5081]: pam_unix(sshd:session): session closed for user core Jul 12 00:10:59.578008 systemd[1]: sshd@15-172.31.29.62:22-139.178.89.65:59680.service: Deactivated successfully. Jul 12 00:10:59.586694 systemd[1]: session-16.scope: Deactivated successfully. Jul 12 00:10:59.589602 systemd-logind[2117]: Session 16 logged out. Waiting for processes to exit. Jul 12 00:10:59.591798 systemd-logind[2117]: Removed session 16. Jul 12 00:11:04.597650 systemd[1]: Started sshd@16-172.31.29.62:22-139.178.89.65:52980.service - OpenSSH per-connection server daemon (139.178.89.65:52980). Jul 12 00:11:04.775012 sshd[5095]: Accepted publickey for core from 139.178.89.65 port 52980 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:04.777828 sshd[5095]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:04.785547 systemd-logind[2117]: New session 17 of user core. Jul 12 00:11:04.792593 systemd[1]: Started session-17.scope - Session 17 of User core. Jul 12 00:11:05.038246 sshd[5095]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:05.045473 systemd[1]: sshd@16-172.31.29.62:22-139.178.89.65:52980.service: Deactivated successfully. Jul 12 00:11:05.052332 systemd[1]: session-17.scope: Deactivated successfully. Jul 12 00:11:05.053989 systemd-logind[2117]: Session 17 logged out. Waiting for processes to exit. Jul 12 00:11:05.055971 systemd-logind[2117]: Removed session 17. Jul 12 00:11:05.069693 systemd[1]: Started sshd@17-172.31.29.62:22-139.178.89.65:52992.service - OpenSSH per-connection server daemon (139.178.89.65:52992). Jul 12 00:11:05.244572 sshd[5109]: Accepted publickey for core from 139.178.89.65 port 52992 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:05.247812 sshd[5109]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:05.258293 systemd-logind[2117]: New session 18 of user core. Jul 12 00:11:05.261607 systemd[1]: Started session-18.scope - Session 18 of User core. Jul 12 00:11:05.607259 sshd[5109]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:05.614786 systemd-logind[2117]: Session 18 logged out. Waiting for processes to exit. Jul 12 00:11:05.615639 systemd[1]: sshd@17-172.31.29.62:22-139.178.89.65:52992.service: Deactivated successfully. Jul 12 00:11:05.620417 systemd[1]: session-18.scope: Deactivated successfully. Jul 12 00:11:05.622552 systemd-logind[2117]: Removed session 18. Jul 12 00:11:05.638838 systemd[1]: Started sshd@18-172.31.29.62:22-139.178.89.65:52998.service - OpenSSH per-connection server daemon (139.178.89.65:52998). Jul 12 00:11:05.817046 sshd[5121]: Accepted publickey for core from 139.178.89.65 port 52998 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:05.819745 sshd[5121]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:05.827852 systemd-logind[2117]: New session 19 of user core. Jul 12 00:11:05.841747 systemd[1]: Started session-19.scope - Session 19 of User core. Jul 12 00:11:08.550603 sshd[5121]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:08.570844 systemd[1]: sshd@18-172.31.29.62:22-139.178.89.65:52998.service: Deactivated successfully. Jul 12 00:11:08.579775 systemd[1]: session-19.scope: Deactivated successfully. Jul 12 00:11:08.583005 systemd-logind[2117]: Session 19 logged out. Waiting for processes to exit. Jul 12 00:11:08.596658 systemd[1]: Started sshd@19-172.31.29.62:22-139.178.89.65:53000.service - OpenSSH per-connection server daemon (139.178.89.65:53000). Jul 12 00:11:08.598501 systemd-logind[2117]: Removed session 19. Jul 12 00:11:08.775977 sshd[5140]: Accepted publickey for core from 139.178.89.65 port 53000 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:08.783484 sshd[5140]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:08.793682 systemd-logind[2117]: New session 20 of user core. Jul 12 00:11:08.803755 systemd[1]: Started session-20.scope - Session 20 of User core. Jul 12 00:11:09.293290 sshd[5140]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:09.298310 systemd[1]: sshd@19-172.31.29.62:22-139.178.89.65:53000.service: Deactivated successfully. Jul 12 00:11:09.307437 systemd[1]: session-20.scope: Deactivated successfully. Jul 12 00:11:09.311982 systemd-logind[2117]: Session 20 logged out. Waiting for processes to exit. Jul 12 00:11:09.314728 systemd-logind[2117]: Removed session 20. Jul 12 00:11:09.325610 systemd[1]: Started sshd@20-172.31.29.62:22-139.178.89.65:53012.service - OpenSSH per-connection server daemon (139.178.89.65:53012). Jul 12 00:11:09.510960 sshd[5152]: Accepted publickey for core from 139.178.89.65 port 53012 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:09.513664 sshd[5152]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:09.522839 systemd-logind[2117]: New session 21 of user core. Jul 12 00:11:09.532725 systemd[1]: Started session-21.scope - Session 21 of User core. Jul 12 00:11:09.776058 sshd[5152]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:09.781623 systemd[1]: sshd@20-172.31.29.62:22-139.178.89.65:53012.service: Deactivated successfully. Jul 12 00:11:09.787992 systemd-logind[2117]: Session 21 logged out. Waiting for processes to exit. Jul 12 00:11:09.792323 systemd[1]: session-21.scope: Deactivated successfully. Jul 12 00:11:09.796654 systemd-logind[2117]: Removed session 21. Jul 12 00:11:14.806645 systemd[1]: Started sshd@21-172.31.29.62:22-139.178.89.65:56512.service - OpenSSH per-connection server daemon (139.178.89.65:56512). Jul 12 00:11:14.987834 sshd[5168]: Accepted publickey for core from 139.178.89.65 port 56512 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:14.990522 sshd[5168]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:14.998989 systemd-logind[2117]: New session 22 of user core. Jul 12 00:11:15.004730 systemd[1]: Started session-22.scope - Session 22 of User core. Jul 12 00:11:15.242914 sshd[5168]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:15.252058 systemd[1]: sshd@21-172.31.29.62:22-139.178.89.65:56512.service: Deactivated successfully. Jul 12 00:11:15.259591 systemd[1]: session-22.scope: Deactivated successfully. Jul 12 00:11:15.261304 systemd-logind[2117]: Session 22 logged out. Waiting for processes to exit. Jul 12 00:11:15.263449 systemd-logind[2117]: Removed session 22. Jul 12 00:11:20.273560 systemd[1]: Started sshd@22-172.31.29.62:22-139.178.89.65:45192.service - OpenSSH per-connection server daemon (139.178.89.65:45192). Jul 12 00:11:20.457739 sshd[5185]: Accepted publickey for core from 139.178.89.65 port 45192 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:20.460349 sshd[5185]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:20.469245 systemd-logind[2117]: New session 23 of user core. Jul 12 00:11:20.479610 systemd[1]: Started session-23.scope - Session 23 of User core. Jul 12 00:11:20.717275 sshd[5185]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:20.722733 systemd[1]: sshd@22-172.31.29.62:22-139.178.89.65:45192.service: Deactivated successfully. Jul 12 00:11:20.730192 systemd[1]: session-23.scope: Deactivated successfully. Jul 12 00:11:20.732664 systemd-logind[2117]: Session 23 logged out. Waiting for processes to exit. Jul 12 00:11:20.735926 systemd-logind[2117]: Removed session 23. Jul 12 00:11:25.747617 systemd[1]: Started sshd@23-172.31.29.62:22-139.178.89.65:45198.service - OpenSSH per-connection server daemon (139.178.89.65:45198). Jul 12 00:11:25.930164 sshd[5199]: Accepted publickey for core from 139.178.89.65 port 45198 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:25.935043 sshd[5199]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:25.945871 systemd-logind[2117]: New session 24 of user core. Jul 12 00:11:25.953755 systemd[1]: Started session-24.scope - Session 24 of User core. Jul 12 00:11:26.219893 sshd[5199]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:26.225776 systemd-logind[2117]: Session 24 logged out. Waiting for processes to exit. Jul 12 00:11:26.226037 systemd[1]: sshd@23-172.31.29.62:22-139.178.89.65:45198.service: Deactivated successfully. Jul 12 00:11:26.232977 systemd[1]: session-24.scope: Deactivated successfully. Jul 12 00:11:26.240489 systemd-logind[2117]: Removed session 24. Jul 12 00:11:31.255013 systemd[1]: Started sshd@24-172.31.29.62:22-139.178.89.65:33244.service - OpenSSH per-connection server daemon (139.178.89.65:33244). Jul 12 00:11:31.433700 sshd[5213]: Accepted publickey for core from 139.178.89.65 port 33244 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:31.436448 sshd[5213]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:31.444377 systemd-logind[2117]: New session 25 of user core. Jul 12 00:11:31.458722 systemd[1]: Started session-25.scope - Session 25 of User core. Jul 12 00:11:31.696387 sshd[5213]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:31.703076 systemd[1]: sshd@24-172.31.29.62:22-139.178.89.65:33244.service: Deactivated successfully. Jul 12 00:11:31.710254 systemd[1]: session-25.scope: Deactivated successfully. Jul 12 00:11:31.712264 systemd-logind[2117]: Session 25 logged out. Waiting for processes to exit. Jul 12 00:11:31.714071 systemd-logind[2117]: Removed session 25. Jul 12 00:11:31.724642 systemd[1]: Started sshd@25-172.31.29.62:22-139.178.89.65:33260.service - OpenSSH per-connection server daemon (139.178.89.65:33260). Jul 12 00:11:31.900335 sshd[5227]: Accepted publickey for core from 139.178.89.65 port 33260 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:31.903122 sshd[5227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:31.913455 systemd-logind[2117]: New session 26 of user core. Jul 12 00:11:31.923634 systemd[1]: Started session-26.scope - Session 26 of User core. Jul 12 00:11:34.251906 containerd[2160]: time="2025-07-12T00:11:34.251260273Z" level=info msg="StopContainer for \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\" with timeout 30 (s)" Jul 12 00:11:34.255724 containerd[2160]: time="2025-07-12T00:11:34.255347065Z" level=info msg="Stop container \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\" with signal terminated" Jul 12 00:11:34.309841 containerd[2160]: time="2025-07-12T00:11:34.309766922Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 12 00:11:34.326217 containerd[2160]: time="2025-07-12T00:11:34.326067182Z" level=info msg="StopContainer for \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\" with timeout 2 (s)" Jul 12 00:11:34.327084 containerd[2160]: time="2025-07-12T00:11:34.326945810Z" level=info msg="Stop container \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\" with signal terminated" Jul 12 00:11:34.342576 systemd-networkd[1694]: lxc_health: Link DOWN Jul 12 00:11:34.342590 systemd-networkd[1694]: lxc_health: Lost carrier Jul 12 00:11:34.361453 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b-rootfs.mount: Deactivated successfully. Jul 12 00:11:34.384780 containerd[2160]: time="2025-07-12T00:11:34.384230498Z" level=info msg="shim disconnected" id=ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b namespace=k8s.io Jul 12 00:11:34.384780 containerd[2160]: time="2025-07-12T00:11:34.384723182Z" level=warning msg="cleaning up after shim disconnected" id=ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b namespace=k8s.io Jul 12 00:11:34.385695 containerd[2160]: time="2025-07-12T00:11:34.384751886Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:34.416317 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed-rootfs.mount: Deactivated successfully. Jul 12 00:11:34.425268 containerd[2160]: time="2025-07-12T00:11:34.425195798Z" level=info msg="shim disconnected" id=908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed namespace=k8s.io Jul 12 00:11:34.425898 containerd[2160]: time="2025-07-12T00:11:34.425648918Z" level=warning msg="cleaning up after shim disconnected" id=908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed namespace=k8s.io Jul 12 00:11:34.425898 containerd[2160]: time="2025-07-12T00:11:34.425695010Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:34.426563 containerd[2160]: time="2025-07-12T00:11:34.426322370Z" level=info msg="StopContainer for \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\" returns successfully" Jul 12 00:11:34.427916 containerd[2160]: time="2025-07-12T00:11:34.427584962Z" level=info msg="StopPodSandbox for \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\"" Jul 12 00:11:34.427916 containerd[2160]: time="2025-07-12T00:11:34.427762718Z" level=info msg="Container to stop \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 12 00:11:34.432341 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab-shm.mount: Deactivated successfully. Jul 12 00:11:34.468710 containerd[2160]: time="2025-07-12T00:11:34.468641498Z" level=info msg="StopContainer for \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\" returns successfully" Jul 12 00:11:34.469854 containerd[2160]: time="2025-07-12T00:11:34.469796654Z" level=info msg="StopPodSandbox for \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\"" Jul 12 00:11:34.470015 containerd[2160]: time="2025-07-12T00:11:34.469869314Z" level=info msg="Container to stop \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 12 00:11:34.470015 containerd[2160]: time="2025-07-12T00:11:34.469897886Z" level=info msg="Container to stop \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 12 00:11:34.470015 containerd[2160]: time="2025-07-12T00:11:34.469921274Z" level=info msg="Container to stop \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 12 00:11:34.470015 containerd[2160]: time="2025-07-12T00:11:34.469947974Z" level=info msg="Container to stop \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 12 00:11:34.470015 containerd[2160]: time="2025-07-12T00:11:34.469971758Z" level=info msg="Container to stop \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 12 00:11:34.476319 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783-shm.mount: Deactivated successfully. Jul 12 00:11:34.521325 containerd[2160]: time="2025-07-12T00:11:34.521157747Z" level=info msg="shim disconnected" id=358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab namespace=k8s.io Jul 12 00:11:34.521813 containerd[2160]: time="2025-07-12T00:11:34.521564763Z" level=warning msg="cleaning up after shim disconnected" id=358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab namespace=k8s.io Jul 12 00:11:34.521813 containerd[2160]: time="2025-07-12T00:11:34.521600559Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:34.551353 containerd[2160]: time="2025-07-12T00:11:34.551188347Z" level=info msg="shim disconnected" id=944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783 namespace=k8s.io Jul 12 00:11:34.551353 containerd[2160]: time="2025-07-12T00:11:34.551293731Z" level=warning msg="cleaning up after shim disconnected" id=944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783 namespace=k8s.io Jul 12 00:11:34.551353 containerd[2160]: time="2025-07-12T00:11:34.551319231Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:34.557478 containerd[2160]: time="2025-07-12T00:11:34.557067063Z" level=warning msg="cleanup warnings time=\"2025-07-12T00:11:34Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Jul 12 00:11:34.562055 containerd[2160]: time="2025-07-12T00:11:34.561865035Z" level=info msg="TearDown network for sandbox \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" successfully" Jul 12 00:11:34.562055 containerd[2160]: time="2025-07-12T00:11:34.561918363Z" level=info msg="StopPodSandbox for \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" returns successfully" Jul 12 00:11:34.587740 containerd[2160]: time="2025-07-12T00:11:34.587668827Z" level=warning msg="cleanup warnings time=\"2025-07-12T00:11:34Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Jul 12 00:11:34.591474 containerd[2160]: time="2025-07-12T00:11:34.590677071Z" level=info msg="TearDown network for sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" successfully" Jul 12 00:11:34.591474 containerd[2160]: time="2025-07-12T00:11:34.590728251Z" level=info msg="StopPodSandbox for \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" returns successfully" Jul 12 00:11:34.675496 kubelet[3513]: I0712 00:11:34.674917 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0ab6073-b03c-47f9-a2b2-b287a5841f36-cilium-config-path\") pod \"c0ab6073-b03c-47f9-a2b2-b287a5841f36\" (UID: \"c0ab6073-b03c-47f9-a2b2-b287a5841f36\") " Jul 12 00:11:34.675496 kubelet[3513]: I0712 00:11:34.674990 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkbq8\" (UniqueName: \"kubernetes.io/projected/c0ab6073-b03c-47f9-a2b2-b287a5841f36-kube-api-access-pkbq8\") pod \"c0ab6073-b03c-47f9-a2b2-b287a5841f36\" (UID: \"c0ab6073-b03c-47f9-a2b2-b287a5841f36\") " Jul 12 00:11:34.680177 kubelet[3513]: I0712 00:11:34.680081 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ab6073-b03c-47f9-a2b2-b287a5841f36-kube-api-access-pkbq8" (OuterVolumeSpecName: "kube-api-access-pkbq8") pod "c0ab6073-b03c-47f9-a2b2-b287a5841f36" (UID: "c0ab6073-b03c-47f9-a2b2-b287a5841f36"). InnerVolumeSpecName "kube-api-access-pkbq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 12 00:11:34.681953 kubelet[3513]: I0712 00:11:34.681885 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0ab6073-b03c-47f9-a2b2-b287a5841f36-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "c0ab6073-b03c-47f9-a2b2-b287a5841f36" (UID: "c0ab6073-b03c-47f9-a2b2-b287a5841f36"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 12 00:11:34.775995 kubelet[3513]: I0712 00:11:34.775832 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-hubble-tls\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.775995 kubelet[3513]: I0712 00:11:34.775902 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-lib-modules\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.775995 kubelet[3513]: I0712 00:11:34.775941 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-run\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.775995 kubelet[3513]: I0712 00:11:34.775972 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-etc-cni-netd\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776327 kubelet[3513]: I0712 00:11:34.776005 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-bpf-maps\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776327 kubelet[3513]: I0712 00:11:34.776036 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cni-path\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776327 kubelet[3513]: I0712 00:11:34.776082 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/69c79161-e4af-41d8-a2b5-6c833216baa7-clustermesh-secrets\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776327 kubelet[3513]: I0712 00:11:34.776142 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-hostproc\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776327 kubelet[3513]: I0712 00:11:34.776187 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-config-path\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776327 kubelet[3513]: I0712 00:11:34.776224 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-net\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776653 kubelet[3513]: I0712 00:11:34.776256 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-kernel\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776653 kubelet[3513]: I0712 00:11:34.776288 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-cgroup\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776653 kubelet[3513]: I0712 00:11:34.776321 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-xtables-lock\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776653 kubelet[3513]: I0712 00:11:34.776357 3513 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcdx2\" (UniqueName: \"kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-kube-api-access-fcdx2\") pod \"69c79161-e4af-41d8-a2b5-6c833216baa7\" (UID: \"69c79161-e4af-41d8-a2b5-6c833216baa7\") " Jul 12 00:11:34.776653 kubelet[3513]: I0712 00:11:34.776426 3513 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c0ab6073-b03c-47f9-a2b2-b287a5841f36-cilium-config-path\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.776653 kubelet[3513]: I0712 00:11:34.776453 3513 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkbq8\" (UniqueName: \"kubernetes.io/projected/c0ab6073-b03c-47f9-a2b2-b287a5841f36-kube-api-access-pkbq8\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.780793 kubelet[3513]: I0712 00:11:34.779767 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-hostproc" (OuterVolumeSpecName: "hostproc") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.782243 kubelet[3513]: I0712 00:11:34.782185 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.782869 kubelet[3513]: I0712 00:11:34.782419 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.785604 kubelet[3513]: I0712 00:11:34.782450 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.785789 kubelet[3513]: I0712 00:11:34.782474 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.785954 kubelet[3513]: I0712 00:11:34.782497 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cni-path" (OuterVolumeSpecName: "cni-path") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.785954 kubelet[3513]: I0712 00:11:34.782705 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.785954 kubelet[3513]: I0712 00:11:34.782749 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.785954 kubelet[3513]: I0712 00:11:34.782780 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.785954 kubelet[3513]: I0712 00:11:34.782812 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 12 00:11:34.786616 kubelet[3513]: I0712 00:11:34.786482 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-kube-api-access-fcdx2" (OuterVolumeSpecName: "kube-api-access-fcdx2") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "kube-api-access-fcdx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 12 00:11:34.789574 kubelet[3513]: I0712 00:11:34.789521 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 12 00:11:34.790266 kubelet[3513]: I0712 00:11:34.790198 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69c79161-e4af-41d8-a2b5-6c833216baa7-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jul 12 00:11:34.791456 kubelet[3513]: I0712 00:11:34.791378 3513 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "69c79161-e4af-41d8-a2b5-6c833216baa7" (UID: "69c79161-e4af-41d8-a2b5-6c833216baa7"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 12 00:11:34.853597 kubelet[3513]: I0712 00:11:34.853510 3513 scope.go:117] "RemoveContainer" containerID="ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b" Jul 12 00:11:34.859873 containerd[2160]: time="2025-07-12T00:11:34.859592092Z" level=info msg="RemoveContainer for \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\"" Jul 12 00:11:34.868370 containerd[2160]: time="2025-07-12T00:11:34.867940744Z" level=info msg="RemoveContainer for \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\" returns successfully" Jul 12 00:11:34.869377 kubelet[3513]: I0712 00:11:34.869171 3513 scope.go:117] "RemoveContainer" containerID="ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b" Jul 12 00:11:34.872065 containerd[2160]: time="2025-07-12T00:11:34.871966780Z" level=error msg="ContainerStatus for \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\": not found" Jul 12 00:11:34.872661 kubelet[3513]: E0712 00:11:34.872600 3513 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\": not found" containerID="ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b" Jul 12 00:11:34.872934 kubelet[3513]: I0712 00:11:34.872815 3513 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b"} err="failed to get container status \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\": rpc error: code = NotFound desc = an error occurred when try to find container \"ac04e40c5023007d1b57f57fcef0e84726f5ff92377bb103cfbdc479ec9cc31b\": not found" Jul 12 00:11:34.873050 kubelet[3513]: I0712 00:11:34.873029 3513 scope.go:117] "RemoveContainer" containerID="908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed" Jul 12 00:11:34.878240 kubelet[3513]: I0712 00:11:34.877869 3513 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-hubble-tls\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.879495 kubelet[3513]: I0712 00:11:34.879405 3513 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-lib-modules\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.880993 kubelet[3513]: I0712 00:11:34.879916 3513 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-run\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.881905 kubelet[3513]: I0712 00:11:34.880955 3513 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-etc-cni-netd\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882199 3513 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-config-path\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882234 3513 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-bpf-maps\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882257 3513 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cni-path\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882278 3513 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/69c79161-e4af-41d8-a2b5-6c833216baa7-clustermesh-secrets\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882299 3513 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-hostproc\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882319 3513 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-net\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882340 3513 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-host-proc-sys-kernel\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.882457 kubelet[3513]: I0712 00:11:34.882361 3513 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-xtables-lock\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.883013 kubelet[3513]: I0712 00:11:34.882382 3513 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcdx2\" (UniqueName: \"kubernetes.io/projected/69c79161-e4af-41d8-a2b5-6c833216baa7-kube-api-access-fcdx2\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.883013 kubelet[3513]: I0712 00:11:34.882403 3513 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/69c79161-e4af-41d8-a2b5-6c833216baa7-cilium-cgroup\") on node \"ip-172-31-29-62\" DevicePath \"\"" Jul 12 00:11:34.883464 containerd[2160]: time="2025-07-12T00:11:34.882703624Z" level=info msg="RemoveContainer for \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\"" Jul 12 00:11:34.888995 containerd[2160]: time="2025-07-12T00:11:34.888941920Z" level=info msg="RemoveContainer for \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\" returns successfully" Jul 12 00:11:34.889873 kubelet[3513]: I0712 00:11:34.889589 3513 scope.go:117] "RemoveContainer" containerID="953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977" Jul 12 00:11:34.893760 containerd[2160]: time="2025-07-12T00:11:34.893683120Z" level=info msg="RemoveContainer for \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\"" Jul 12 00:11:34.906990 containerd[2160]: time="2025-07-12T00:11:34.906442745Z" level=info msg="RemoveContainer for \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\" returns successfully" Jul 12 00:11:34.907227 kubelet[3513]: I0712 00:11:34.906730 3513 scope.go:117] "RemoveContainer" containerID="b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a" Jul 12 00:11:34.912665 containerd[2160]: time="2025-07-12T00:11:34.912609581Z" level=info msg="RemoveContainer for \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\"" Jul 12 00:11:34.929134 containerd[2160]: time="2025-07-12T00:11:34.928982009Z" level=info msg="RemoveContainer for \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\" returns successfully" Jul 12 00:11:34.929650 kubelet[3513]: I0712 00:11:34.929600 3513 scope.go:117] "RemoveContainer" containerID="e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90" Jul 12 00:11:34.932766 containerd[2160]: time="2025-07-12T00:11:34.932721137Z" level=info msg="RemoveContainer for \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\"" Jul 12 00:11:34.939139 containerd[2160]: time="2025-07-12T00:11:34.939071801Z" level=info msg="RemoveContainer for \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\" returns successfully" Jul 12 00:11:34.939725 kubelet[3513]: I0712 00:11:34.939588 3513 scope.go:117] "RemoveContainer" containerID="49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6" Jul 12 00:11:34.941802 containerd[2160]: time="2025-07-12T00:11:34.941426789Z" level=info msg="RemoveContainer for \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\"" Jul 12 00:11:34.947310 containerd[2160]: time="2025-07-12T00:11:34.947261645Z" level=info msg="RemoveContainer for \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\" returns successfully" Jul 12 00:11:34.947791 kubelet[3513]: I0712 00:11:34.947740 3513 scope.go:117] "RemoveContainer" containerID="908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed" Jul 12 00:11:34.948212 containerd[2160]: time="2025-07-12T00:11:34.948127253Z" level=error msg="ContainerStatus for \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\": not found" Jul 12 00:11:34.948688 kubelet[3513]: E0712 00:11:34.948472 3513 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\": not found" containerID="908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed" Jul 12 00:11:34.948688 kubelet[3513]: I0712 00:11:34.948524 3513 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed"} err="failed to get container status \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\": rpc error: code = NotFound desc = an error occurred when try to find container \"908767c613a1e7632694a1793a2db9653a9412e4cf8209b5bc3c6a4bfb0e52ed\": not found" Jul 12 00:11:34.948688 kubelet[3513]: I0712 00:11:34.948559 3513 scope.go:117] "RemoveContainer" containerID="953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977" Jul 12 00:11:34.949332 containerd[2160]: time="2025-07-12T00:11:34.949086509Z" level=error msg="ContainerStatus for \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\": not found" Jul 12 00:11:34.949578 kubelet[3513]: E0712 00:11:34.949537 3513 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\": not found" containerID="953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977" Jul 12 00:11:34.949678 kubelet[3513]: I0712 00:11:34.949591 3513 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977"} err="failed to get container status \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\": rpc error: code = NotFound desc = an error occurred when try to find container \"953b9b0fcf593af2e6981ce1703c71440af96c99b82362b2539b41e30ca76977\": not found" Jul 12 00:11:34.949678 kubelet[3513]: I0712 00:11:34.949624 3513 scope.go:117] "RemoveContainer" containerID="b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a" Jul 12 00:11:34.950142 containerd[2160]: time="2025-07-12T00:11:34.950010197Z" level=error msg="ContainerStatus for \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\": not found" Jul 12 00:11:34.950337 kubelet[3513]: E0712 00:11:34.950295 3513 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\": not found" containerID="b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a" Jul 12 00:11:34.950426 kubelet[3513]: I0712 00:11:34.950347 3513 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a"} err="failed to get container status \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\": rpc error: code = NotFound desc = an error occurred when try to find container \"b497c9fdfac811cd5792f574d3ec731a702de328316029607c356013d328795a\": not found" Jul 12 00:11:34.950426 kubelet[3513]: I0712 00:11:34.950402 3513 scope.go:117] "RemoveContainer" containerID="e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90" Jul 12 00:11:34.950839 containerd[2160]: time="2025-07-12T00:11:34.950789897Z" level=error msg="ContainerStatus for \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\": not found" Jul 12 00:11:34.951267 kubelet[3513]: E0712 00:11:34.951183 3513 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\": not found" containerID="e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90" Jul 12 00:11:34.951395 kubelet[3513]: I0712 00:11:34.951293 3513 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90"} err="failed to get container status \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\": rpc error: code = NotFound desc = an error occurred when try to find container \"e46dabda29889fe2cea170f40e90287527f2e822af47143abb7b504f30323e90\": not found" Jul 12 00:11:34.951395 kubelet[3513]: I0712 00:11:34.951327 3513 scope.go:117] "RemoveContainer" containerID="49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6" Jul 12 00:11:34.952841 containerd[2160]: time="2025-07-12T00:11:34.951751673Z" level=error msg="ContainerStatus for \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\": not found" Jul 12 00:11:34.953000 kubelet[3513]: E0712 00:11:34.952573 3513 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\": not found" containerID="49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6" Jul 12 00:11:34.953000 kubelet[3513]: I0712 00:11:34.952623 3513 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6"} err="failed to get container status \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\": rpc error: code = NotFound desc = an error occurred when try to find container \"49e4b1fea481ce26c696cb58865e4d5518e15f8b36011f7d65b3ec9e9289eeb6\": not found" Jul 12 00:11:35.250521 kubelet[3513]: I0712 00:11:35.250453 3513 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69c79161-e4af-41d8-a2b5-6c833216baa7" path="/var/lib/kubelet/pods/69c79161-e4af-41d8-a2b5-6c833216baa7/volumes" Jul 12 00:11:35.251981 kubelet[3513]: I0712 00:11:35.251922 3513 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ab6073-b03c-47f9-a2b2-b287a5841f36" path="/var/lib/kubelet/pods/c0ab6073-b03c-47f9-a2b2-b287a5841f36/volumes" Jul 12 00:11:35.279410 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab-rootfs.mount: Deactivated successfully. Jul 12 00:11:35.279692 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783-rootfs.mount: Deactivated successfully. Jul 12 00:11:35.279917 systemd[1]: var-lib-kubelet-pods-c0ab6073\x2db03c\x2d47f9\x2da2b2\x2db287a5841f36-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dpkbq8.mount: Deactivated successfully. Jul 12 00:11:35.280591 systemd[1]: var-lib-kubelet-pods-69c79161\x2de4af\x2d41d8\x2da2b5\x2d6c833216baa7-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dfcdx2.mount: Deactivated successfully. Jul 12 00:11:35.280958 systemd[1]: var-lib-kubelet-pods-69c79161\x2de4af\x2d41d8\x2da2b5\x2d6c833216baa7-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jul 12 00:11:35.281343 systemd[1]: var-lib-kubelet-pods-69c79161\x2de4af\x2d41d8\x2da2b5\x2d6c833216baa7-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jul 12 00:11:35.355177 containerd[2160]: time="2025-07-12T00:11:35.345179439Z" level=info msg="StopPodSandbox for \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\"" Jul 12 00:11:35.355177 containerd[2160]: time="2025-07-12T00:11:35.345334995Z" level=info msg="TearDown network for sandbox \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" successfully" Jul 12 00:11:35.355177 containerd[2160]: time="2025-07-12T00:11:35.345362127Z" level=info msg="StopPodSandbox for \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" returns successfully" Jul 12 00:11:35.355177 containerd[2160]: time="2025-07-12T00:11:35.346373739Z" level=info msg="RemovePodSandbox for \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\"" Jul 12 00:11:35.355177 containerd[2160]: time="2025-07-12T00:11:35.346420611Z" level=info msg="Forcibly stopping sandbox \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\"" Jul 12 00:11:35.355177 containerd[2160]: time="2025-07-12T00:11:35.346581591Z" level=info msg="TearDown network for sandbox \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" successfully" Jul 12 00:11:35.356410 containerd[2160]: time="2025-07-12T00:11:35.356032227Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 12 00:11:35.356581 containerd[2160]: time="2025-07-12T00:11:35.356540055Z" level=info msg="RemovePodSandbox \"358af116292789c3f440b0a1421196951d8e56ba7619dd345ff00ab8f4348eab\" returns successfully" Jul 12 00:11:35.357942 containerd[2160]: time="2025-07-12T00:11:35.357890991Z" level=info msg="StopPodSandbox for \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\"" Jul 12 00:11:35.358074 containerd[2160]: time="2025-07-12T00:11:35.358033491Z" level=info msg="TearDown network for sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" successfully" Jul 12 00:11:35.358180 containerd[2160]: time="2025-07-12T00:11:35.358069215Z" level=info msg="StopPodSandbox for \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" returns successfully" Jul 12 00:11:35.358940 containerd[2160]: time="2025-07-12T00:11:35.358650927Z" level=info msg="RemovePodSandbox for \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\"" Jul 12 00:11:35.358940 containerd[2160]: time="2025-07-12T00:11:35.358710555Z" level=info msg="Forcibly stopping sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\"" Jul 12 00:11:35.358940 containerd[2160]: time="2025-07-12T00:11:35.358823751Z" level=info msg="TearDown network for sandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" successfully" Jul 12 00:11:35.364873 containerd[2160]: time="2025-07-12T00:11:35.364793367Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jul 12 00:11:35.364873 containerd[2160]: time="2025-07-12T00:11:35.364867131Z" level=info msg="RemovePodSandbox \"944e6f268c0fd03bceb07a5e23e16b19db94f77b924876f7e3b048e3b8976783\" returns successfully" Jul 12 00:11:35.556806 kubelet[3513]: E0712 00:11:35.556650 3513 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 12 00:11:36.183051 sshd[5227]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:36.188673 systemd[1]: sshd@25-172.31.29.62:22-139.178.89.65:33260.service: Deactivated successfully. Jul 12 00:11:36.198264 systemd[1]: session-26.scope: Deactivated successfully. Jul 12 00:11:36.200441 systemd-logind[2117]: Session 26 logged out. Waiting for processes to exit. Jul 12 00:11:36.202554 systemd-logind[2117]: Removed session 26. Jul 12 00:11:36.220592 systemd[1]: Started sshd@26-172.31.29.62:22-139.178.89.65:33262.service - OpenSSH per-connection server daemon (139.178.89.65:33262). Jul 12 00:11:36.394517 sshd[5398]: Accepted publickey for core from 139.178.89.65 port 33262 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:36.397942 sshd[5398]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:36.406209 systemd-logind[2117]: New session 27 of user core. Jul 12 00:11:36.413649 systemd[1]: Started session-27.scope - Session 27 of User core. Jul 12 00:11:37.047052 ntpd[2101]: Deleting interface #10 lxc_health, fe80::44be:10ff:fe27:1e0c%8#123, interface stats: received=0, sent=0, dropped=0, active_time=94 secs Jul 12 00:11:37.047620 ntpd[2101]: 12 Jul 00:11:37 ntpd[2101]: Deleting interface #10 lxc_health, fe80::44be:10ff:fe27:1e0c%8#123, interface stats: received=0, sent=0, dropped=0, active_time=94 secs Jul 12 00:11:38.225019 kubelet[3513]: I0712 00:11:38.224955 3513 setters.go:600] "Node became not ready" node="ip-172-31-29-62" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-07-12T00:11:38Z","lastTransitionTime":"2025-07-12T00:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Jul 12 00:11:38.241498 sshd[5398]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:38.255057 kubelet[3513]: E0712 00:11:38.250951 3513 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c0ab6073-b03c-47f9-a2b2-b287a5841f36" containerName="cilium-operator" Jul 12 00:11:38.255057 kubelet[3513]: E0712 00:11:38.251001 3513 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="69c79161-e4af-41d8-a2b5-6c833216baa7" containerName="mount-bpf-fs" Jul 12 00:11:38.255057 kubelet[3513]: E0712 00:11:38.251017 3513 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="69c79161-e4af-41d8-a2b5-6c833216baa7" containerName="clean-cilium-state" Jul 12 00:11:38.255057 kubelet[3513]: E0712 00:11:38.251032 3513 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="69c79161-e4af-41d8-a2b5-6c833216baa7" containerName="cilium-agent" Jul 12 00:11:38.255057 kubelet[3513]: E0712 00:11:38.251048 3513 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="69c79161-e4af-41d8-a2b5-6c833216baa7" containerName="apply-sysctl-overwrites" Jul 12 00:11:38.255057 kubelet[3513]: E0712 00:11:38.251063 3513 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="69c79161-e4af-41d8-a2b5-6c833216baa7" containerName="mount-cgroup" Jul 12 00:11:38.255057 kubelet[3513]: I0712 00:11:38.251123 3513 memory_manager.go:354] "RemoveStaleState removing state" podUID="69c79161-e4af-41d8-a2b5-6c833216baa7" containerName="cilium-agent" Jul 12 00:11:38.255057 kubelet[3513]: I0712 00:11:38.251144 3513 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ab6073-b03c-47f9-a2b2-b287a5841f36" containerName="cilium-operator" Jul 12 00:11:38.255656 systemd[1]: sshd@26-172.31.29.62:22-139.178.89.65:33262.service: Deactivated successfully. Jul 12 00:11:38.278667 kubelet[3513]: W0712 00:11:38.278602 3513 reflector.go:561] object-"kube-system"/"hubble-server-certs": failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ip-172-31-29-62" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ip-172-31-29-62' and this object Jul 12 00:11:38.278855 kubelet[3513]: E0712 00:11:38.278674 3513 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"hubble-server-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"hubble-server-certs\" is forbidden: User \"system:node:ip-172-31-29-62\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ip-172-31-29-62' and this object" logger="UnhandledError" Jul 12 00:11:38.278855 kubelet[3513]: W0712 00:11:38.278798 3513 reflector.go:561] object-"kube-system"/"cilium-clustermesh": failed to list *v1.Secret: secrets "cilium-clustermesh" is forbidden: User "system:node:ip-172-31-29-62" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ip-172-31-29-62' and this object Jul 12 00:11:38.278855 kubelet[3513]: E0712 00:11:38.278828 3513 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-clustermesh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-clustermesh\" is forbidden: User \"system:node:ip-172-31-29-62\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ip-172-31-29-62' and this object" logger="UnhandledError" Jul 12 00:11:38.294398 systemd[1]: session-27.scope: Deactivated successfully. Jul 12 00:11:38.303501 systemd-logind[2117]: Session 27 logged out. Waiting for processes to exit. Jul 12 00:11:38.308716 kubelet[3513]: W0712 00:11:38.307094 3513 reflector.go:561] object-"kube-system"/"cilium-ipsec-keys": failed to list *v1.Secret: secrets "cilium-ipsec-keys" is forbidden: User "system:node:ip-172-31-29-62" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ip-172-31-29-62' and this object Jul 12 00:11:38.308716 kubelet[3513]: E0712 00:11:38.308255 3513 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-ipsec-keys\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-ipsec-keys\" is forbidden: User \"system:node:ip-172-31-29-62\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ip-172-31-29-62' and this object" logger="UnhandledError" Jul 12 00:11:38.318674 systemd[1]: Started sshd@27-172.31.29.62:22-139.178.89.65:33264.service - OpenSSH per-connection server daemon (139.178.89.65:33264). Jul 12 00:11:38.329612 systemd-logind[2117]: Removed session 27. Jul 12 00:11:38.417469 kubelet[3513]: I0712 00:11:38.417415 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-host-proc-sys-net\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418169 kubelet[3513]: I0712 00:11:38.417694 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-etc-cni-netd\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418169 kubelet[3513]: I0712 00:11:38.417740 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-lib-modules\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418169 kubelet[3513]: I0712 00:11:38.417777 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-cilium-ipsec-secrets\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418169 kubelet[3513]: I0712 00:11:38.417815 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-cilium-config-path\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418169 kubelet[3513]: I0712 00:11:38.417852 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-xtables-lock\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418169 kubelet[3513]: I0712 00:11:38.417907 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-clustermesh-secrets\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418551 kubelet[3513]: I0712 00:11:38.417947 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-host-proc-sys-kernel\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418551 kubelet[3513]: I0712 00:11:38.417986 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-bpf-maps\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418551 kubelet[3513]: I0712 00:11:38.418019 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-hubble-tls\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418551 kubelet[3513]: I0712 00:11:38.418061 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-cilium-run\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418551 kubelet[3513]: I0712 00:11:38.418125 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-hostproc\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418551 kubelet[3513]: I0712 00:11:38.418167 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-cilium-cgroup\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418871 kubelet[3513]: I0712 00:11:38.418204 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-cni-path\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.418871 kubelet[3513]: I0712 00:11:38.418242 3513 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4bjs\" (UniqueName: \"kubernetes.io/projected/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-kube-api-access-l4bjs\") pod \"cilium-m5qpq\" (UID: \"f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8\") " pod="kube-system/cilium-m5qpq" Jul 12 00:11:38.532422 sshd[5410]: Accepted publickey for core from 139.178.89.65 port 33264 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:38.538244 sshd[5410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:38.566373 systemd-logind[2117]: New session 28 of user core. Jul 12 00:11:38.572974 systemd[1]: Started session-28.scope - Session 28 of User core. Jul 12 00:11:38.717172 sshd[5410]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:38.722767 systemd-logind[2117]: Session 28 logged out. Waiting for processes to exit. Jul 12 00:11:38.726481 systemd[1]: sshd@27-172.31.29.62:22-139.178.89.65:33264.service: Deactivated successfully. Jul 12 00:11:38.731320 systemd[1]: session-28.scope: Deactivated successfully. Jul 12 00:11:38.732560 systemd-logind[2117]: Removed session 28. Jul 12 00:11:38.746676 systemd[1]: Started sshd@28-172.31.29.62:22-139.178.89.65:33266.service - OpenSSH per-connection server daemon (139.178.89.65:33266). Jul 12 00:11:38.927970 sshd[5422]: Accepted publickey for core from 139.178.89.65 port 33266 ssh2: RSA SHA256:rqVc07ZHJYS8k/+pkEfeFkMPPbocnthwPTDCiAXji4Q Jul 12 00:11:38.930768 sshd[5422]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 12 00:11:38.939356 systemd-logind[2117]: New session 29 of user core. Jul 12 00:11:38.947611 systemd[1]: Started session-29.scope - Session 29 of User core. Jul 12 00:11:39.522224 kubelet[3513]: E0712 00:11:39.521907 3513 secret.go:189] Couldn't get secret kube-system/cilium-ipsec-keys: failed to sync secret cache: timed out waiting for the condition Jul 12 00:11:39.522224 kubelet[3513]: E0712 00:11:39.522019 3513 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-cilium-ipsec-secrets podName:f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8 nodeName:}" failed. No retries permitted until 2025-07-12 00:11:40.021988115 +0000 UTC m=+125.304989224 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-ipsec-secrets" (UniqueName: "kubernetes.io/secret/f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8-cilium-ipsec-secrets") pod "cilium-m5qpq" (UID: "f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8") : failed to sync secret cache: timed out waiting for the condition Jul 12 00:11:40.098658 containerd[2160]: time="2025-07-12T00:11:40.098589630Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-m5qpq,Uid:f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8,Namespace:kube-system,Attempt:0,}" Jul 12 00:11:40.143240 containerd[2160]: time="2025-07-12T00:11:40.142293187Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 12 00:11:40.143240 containerd[2160]: time="2025-07-12T00:11:40.143142907Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 12 00:11:40.143240 containerd[2160]: time="2025-07-12T00:11:40.143189455Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:11:40.143672 containerd[2160]: time="2025-07-12T00:11:40.143372587Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 12 00:11:40.214748 containerd[2160]: time="2025-07-12T00:11:40.214670767Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-m5qpq,Uid:f10665d8-cf7b-4c29-b8c2-5ee7acbbacf8,Namespace:kube-system,Attempt:0,} returns sandbox id \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\"" Jul 12 00:11:40.221900 containerd[2160]: time="2025-07-12T00:11:40.221833039Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 12 00:11:40.246908 kubelet[3513]: E0712 00:11:40.246822 3513 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7c65d6cfc9-8frgg" podUID="36a2150d-1e4a-4860-9b9b-8e0d28cde3c0" Jul 12 00:11:40.247694 containerd[2160]: time="2025-07-12T00:11:40.247474855Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"ebcd4ceb62fa004a7f82434f56c33421d4c89811cbc87d5dc8f038c80a10be52\"" Jul 12 00:11:40.249362 containerd[2160]: time="2025-07-12T00:11:40.249304675Z" level=info msg="StartContainer for \"ebcd4ceb62fa004a7f82434f56c33421d4c89811cbc87d5dc8f038c80a10be52\"" Jul 12 00:11:40.338149 containerd[2160]: time="2025-07-12T00:11:40.336911888Z" level=info msg="StartContainer for \"ebcd4ceb62fa004a7f82434f56c33421d4c89811cbc87d5dc8f038c80a10be52\" returns successfully" Jul 12 00:11:40.411226 containerd[2160]: time="2025-07-12T00:11:40.411139340Z" level=info msg="shim disconnected" id=ebcd4ceb62fa004a7f82434f56c33421d4c89811cbc87d5dc8f038c80a10be52 namespace=k8s.io Jul 12 00:11:40.411226 containerd[2160]: time="2025-07-12T00:11:40.411219008Z" level=warning msg="cleaning up after shim disconnected" id=ebcd4ceb62fa004a7f82434f56c33421d4c89811cbc87d5dc8f038c80a10be52 namespace=k8s.io Jul 12 00:11:40.411758 containerd[2160]: time="2025-07-12T00:11:40.411241856Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:40.557955 kubelet[3513]: E0712 00:11:40.557888 3513 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 12 00:11:40.887595 containerd[2160]: time="2025-07-12T00:11:40.887383966Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 12 00:11:40.916250 containerd[2160]: time="2025-07-12T00:11:40.913852966Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"8807ce6158cdaeb822bc796f85f2b0388596290bbc4dfd771e10847b9b7c3955\"" Jul 12 00:11:40.922667 containerd[2160]: time="2025-07-12T00:11:40.919678906Z" level=info msg="StartContainer for \"8807ce6158cdaeb822bc796f85f2b0388596290bbc4dfd771e10847b9b7c3955\"" Jul 12 00:11:41.041488 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2220817202.mount: Deactivated successfully. Jul 12 00:11:41.058941 containerd[2160]: time="2025-07-12T00:11:41.058833271Z" level=info msg="StartContainer for \"8807ce6158cdaeb822bc796f85f2b0388596290bbc4dfd771e10847b9b7c3955\" returns successfully" Jul 12 00:11:41.101877 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8807ce6158cdaeb822bc796f85f2b0388596290bbc4dfd771e10847b9b7c3955-rootfs.mount: Deactivated successfully. Jul 12 00:11:41.110033 containerd[2160]: time="2025-07-12T00:11:41.109944151Z" level=info msg="shim disconnected" id=8807ce6158cdaeb822bc796f85f2b0388596290bbc4dfd771e10847b9b7c3955 namespace=k8s.io Jul 12 00:11:41.110033 containerd[2160]: time="2025-07-12T00:11:41.110023087Z" level=warning msg="cleaning up after shim disconnected" id=8807ce6158cdaeb822bc796f85f2b0388596290bbc4dfd771e10847b9b7c3955 namespace=k8s.io Jul 12 00:11:41.110033 containerd[2160]: time="2025-07-12T00:11:41.110045515Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:41.904626 containerd[2160]: time="2025-07-12T00:11:41.904573463Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 12 00:11:41.937672 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2534614904.mount: Deactivated successfully. Jul 12 00:11:41.943610 containerd[2160]: time="2025-07-12T00:11:41.943531044Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"f59cd7fedbbf41df3f916ea5fb200af80ea41be795808e203411da32fb8af00a\"" Jul 12 00:11:41.945117 containerd[2160]: time="2025-07-12T00:11:41.944901804Z" level=info msg="StartContainer for \"f59cd7fedbbf41df3f916ea5fb200af80ea41be795808e203411da32fb8af00a\"" Jul 12 00:11:42.074818 containerd[2160]: time="2025-07-12T00:11:42.074639816Z" level=info msg="StartContainer for \"f59cd7fedbbf41df3f916ea5fb200af80ea41be795808e203411da32fb8af00a\" returns successfully" Jul 12 00:11:42.115721 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f59cd7fedbbf41df3f916ea5fb200af80ea41be795808e203411da32fb8af00a-rootfs.mount: Deactivated successfully. Jul 12 00:11:42.123585 containerd[2160]: time="2025-07-12T00:11:42.123506168Z" level=info msg="shim disconnected" id=f59cd7fedbbf41df3f916ea5fb200af80ea41be795808e203411da32fb8af00a namespace=k8s.io Jul 12 00:11:42.124510 containerd[2160]: time="2025-07-12T00:11:42.123584384Z" level=warning msg="cleaning up after shim disconnected" id=f59cd7fedbbf41df3f916ea5fb200af80ea41be795808e203411da32fb8af00a namespace=k8s.io Jul 12 00:11:42.124510 containerd[2160]: time="2025-07-12T00:11:42.123608636Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:42.247270 kubelet[3513]: E0712 00:11:42.246434 3513 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7c65d6cfc9-8frgg" podUID="36a2150d-1e4a-4860-9b9b-8e0d28cde3c0" Jul 12 00:11:42.901538 containerd[2160]: time="2025-07-12T00:11:42.901310400Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 12 00:11:42.947630 containerd[2160]: time="2025-07-12T00:11:42.947543017Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"c786e4aab1e3390725d6369a3b297f90f5f182908d5a53dbede724313df55582\"" Jul 12 00:11:42.952413 containerd[2160]: time="2025-07-12T00:11:42.950864233Z" level=info msg="StartContainer for \"c786e4aab1e3390725d6369a3b297f90f5f182908d5a53dbede724313df55582\"" Jul 12 00:11:43.115258 containerd[2160]: time="2025-07-12T00:11:43.114714213Z" level=info msg="StartContainer for \"c786e4aab1e3390725d6369a3b297f90f5f182908d5a53dbede724313df55582\" returns successfully" Jul 12 00:11:43.176679 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c786e4aab1e3390725d6369a3b297f90f5f182908d5a53dbede724313df55582-rootfs.mount: Deactivated successfully. Jul 12 00:11:43.182127 containerd[2160]: time="2025-07-12T00:11:43.182025526Z" level=info msg="shim disconnected" id=c786e4aab1e3390725d6369a3b297f90f5f182908d5a53dbede724313df55582 namespace=k8s.io Jul 12 00:11:43.182127 containerd[2160]: time="2025-07-12T00:11:43.182123446Z" level=warning msg="cleaning up after shim disconnected" id=c786e4aab1e3390725d6369a3b297f90f5f182908d5a53dbede724313df55582 namespace=k8s.io Jul 12 00:11:43.183113 containerd[2160]: time="2025-07-12T00:11:43.182146402Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:11:43.910223 containerd[2160]: time="2025-07-12T00:11:43.909341857Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 12 00:11:43.944175 containerd[2160]: time="2025-07-12T00:11:43.943905073Z" level=info msg="CreateContainer within sandbox \"e2344b1a96111fc07403077ce2b77746e0a27399a98c30040e5107a9f3672df6\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"60e6f80abfe1d9b24ad28a0cee5579be2d6233372c474bfbbe27d7c4ae50a777\"" Jul 12 00:11:43.946552 containerd[2160]: time="2025-07-12T00:11:43.946237057Z" level=info msg="StartContainer for \"60e6f80abfe1d9b24ad28a0cee5579be2d6233372c474bfbbe27d7c4ae50a777\"" Jul 12 00:11:44.058905 containerd[2160]: time="2025-07-12T00:11:44.058774582Z" level=info msg="StartContainer for \"60e6f80abfe1d9b24ad28a0cee5579be2d6233372c474bfbbe27d7c4ae50a777\" returns successfully" Jul 12 00:11:44.246508 kubelet[3513]: E0712 00:11:44.246344 3513 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7c65d6cfc9-8frgg" podUID="36a2150d-1e4a-4860-9b9b-8e0d28cde3c0" Jul 12 00:11:44.810140 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Jul 12 00:11:44.951214 kubelet[3513]: I0712 00:11:44.950138 3513 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-m5qpq" podStartSLOduration=6.950055698 podStartE2EDuration="6.950055698s" podCreationTimestamp="2025-07-12 00:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-12 00:11:44.949476782 +0000 UTC m=+130.232477915" watchObservedRunningTime="2025-07-12 00:11:44.950055698 +0000 UTC m=+130.233056807" Jul 12 00:11:48.972317 systemd-networkd[1694]: lxc_health: Link UP Jul 12 00:11:48.987873 (udev-worker)[6273]: Network interface NamePolicy= disabled on kernel command line. Jul 12 00:11:48.995615 systemd-networkd[1694]: lxc_health: Gained carrier Jul 12 00:11:50.121340 systemd-networkd[1694]: lxc_health: Gained IPv6LL Jul 12 00:11:53.047134 ntpd[2101]: Listen normally on 13 lxc_health [fe80::7013:afff:fe5c:36ee%14]:123 Jul 12 00:11:53.047801 ntpd[2101]: 12 Jul 00:11:53 ntpd[2101]: Listen normally on 13 lxc_health [fe80::7013:afff:fe5c:36ee%14]:123 Jul 12 00:11:54.506756 systemd[1]: run-containerd-runc-k8s.io-60e6f80abfe1d9b24ad28a0cee5579be2d6233372c474bfbbe27d7c4ae50a777-runc.ejkciI.mount: Deactivated successfully. Jul 12 00:11:54.628431 sshd[5422]: pam_unix(sshd:session): session closed for user core Jul 12 00:11:54.637732 systemd[1]: sshd@28-172.31.29.62:22-139.178.89.65:33266.service: Deactivated successfully. Jul 12 00:11:54.649948 systemd[1]: session-29.scope: Deactivated successfully. Jul 12 00:11:54.653728 systemd-logind[2117]: Session 29 logged out. Waiting for processes to exit. Jul 12 00:11:54.657499 systemd-logind[2117]: Removed session 29. Jul 12 00:12:08.433548 kubelet[3513]: E0712 00:12:08.433273 3513 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.62:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-62?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jul 12 00:12:08.486236 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b5a4318fbac094b856dc52c3ba989532a56b47a22d4d6e5df5216a0383f187d4-rootfs.mount: Deactivated successfully. Jul 12 00:12:08.541801 containerd[2160]: time="2025-07-12T00:12:08.541544652Z" level=info msg="shim disconnected" id=b5a4318fbac094b856dc52c3ba989532a56b47a22d4d6e5df5216a0383f187d4 namespace=k8s.io Jul 12 00:12:08.541801 containerd[2160]: time="2025-07-12T00:12:08.541619784Z" level=warning msg="cleaning up after shim disconnected" id=b5a4318fbac094b856dc52c3ba989532a56b47a22d4d6e5df5216a0383f187d4 namespace=k8s.io Jul 12 00:12:08.541801 containerd[2160]: time="2025-07-12T00:12:08.541640736Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:12:08.987677 kubelet[3513]: I0712 00:12:08.987295 3513 scope.go:117] "RemoveContainer" containerID="b5a4318fbac094b856dc52c3ba989532a56b47a22d4d6e5df5216a0383f187d4" Jul 12 00:12:08.990070 containerd[2160]: time="2025-07-12T00:12:08.989994014Z" level=info msg="CreateContainer within sandbox \"a0fc4c1fb46341a8a0fd3cd96f3a362516d006e986359c14a83f169dd2713fec\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Jul 12 00:12:09.008737 containerd[2160]: time="2025-07-12T00:12:09.008618770Z" level=info msg="CreateContainer within sandbox \"a0fc4c1fb46341a8a0fd3cd96f3a362516d006e986359c14a83f169dd2713fec\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"bc74f759f66c74dd25987f2ca7240ccfcf3943899aec3d7acc5f2806c0637b82\"" Jul 12 00:12:09.009436 containerd[2160]: time="2025-07-12T00:12:09.009347374Z" level=info msg="StartContainer for \"bc74f759f66c74dd25987f2ca7240ccfcf3943899aec3d7acc5f2806c0637b82\"" Jul 12 00:12:09.127996 containerd[2160]: time="2025-07-12T00:12:09.127851191Z" level=info msg="StartContainer for \"bc74f759f66c74dd25987f2ca7240ccfcf3943899aec3d7acc5f2806c0637b82\" returns successfully" Jul 12 00:12:14.395307 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ee89a1192f85ce65d9f50ba839b8838599dc4ab658e08fdd97c748d1e6df178b-rootfs.mount: Deactivated successfully. Jul 12 00:12:14.399593 containerd[2160]: time="2025-07-12T00:12:14.397727453Z" level=info msg="shim disconnected" id=ee89a1192f85ce65d9f50ba839b8838599dc4ab658e08fdd97c748d1e6df178b namespace=k8s.io Jul 12 00:12:14.399593 containerd[2160]: time="2025-07-12T00:12:14.397806977Z" level=warning msg="cleaning up after shim disconnected" id=ee89a1192f85ce65d9f50ba839b8838599dc4ab658e08fdd97c748d1e6df178b namespace=k8s.io Jul 12 00:12:14.399593 containerd[2160]: time="2025-07-12T00:12:14.397829237Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 12 00:12:15.013901 kubelet[3513]: I0712 00:12:15.013846 3513 scope.go:117] "RemoveContainer" containerID="ee89a1192f85ce65d9f50ba839b8838599dc4ab658e08fdd97c748d1e6df178b" Jul 12 00:12:15.017316 containerd[2160]: time="2025-07-12T00:12:15.017132236Z" level=info msg="CreateContainer within sandbox \"fba0ab38385978627822e35ac670b4acfe00d1905088a0c8e79e43dc973a8506\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Jul 12 00:12:15.036665 containerd[2160]: time="2025-07-12T00:12:15.036408616Z" level=info msg="CreateContainer within sandbox \"fba0ab38385978627822e35ac670b4acfe00d1905088a0c8e79e43dc973a8506\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"4e7634e716dc8b062d2e6a3e153570c3d171c5e7ac053277077ea4885270084f\"" Jul 12 00:12:15.037612 containerd[2160]: time="2025-07-12T00:12:15.037160392Z" level=info msg="StartContainer for \"4e7634e716dc8b062d2e6a3e153570c3d171c5e7ac053277077ea4885270084f\"" Jul 12 00:12:15.147075 containerd[2160]: time="2025-07-12T00:12:15.146782576Z" level=info msg="StartContainer for \"4e7634e716dc8b062d2e6a3e153570c3d171c5e7ac053277077ea4885270084f\" returns successfully" Jul 12 00:12:18.434931 kubelet[3513]: E0712 00:12:18.434215 3513 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.62:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-62?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jul 12 00:12:28.434959 kubelet[3513]: E0712 00:12:28.434652 3513 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.62:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-62?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"