May 7 23:46:34.165149 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd083] May 7 23:46:34.165194 kernel: Linux version 6.6.88-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT Wed May 7 22:21:35 -00 2025 May 7 23:46:34.165220 kernel: KASLR disabled due to lack of seed May 7 23:46:34.165236 kernel: efi: EFI v2.7 by EDK II May 7 23:46:34.165252 kernel: efi: SMBIOS=0x7bed0000 SMBIOS 3.0=0x7beb0000 ACPI=0x786e0000 ACPI 2.0=0x786e0014 MEMATTR=0x7a733a98 MEMRESERVE=0x78557598 May 7 23:46:34.165267 kernel: secureboot: Secure boot disabled May 7 23:46:34.165284 kernel: ACPI: Early table checksum verification disabled May 7 23:46:34.165299 kernel: ACPI: RSDP 0x00000000786E0014 000024 (v02 AMAZON) May 7 23:46:34.165315 kernel: ACPI: XSDT 0x00000000786D00E8 000064 (v01 AMAZON AMZNFACP 00000001 01000013) May 7 23:46:34.165330 kernel: ACPI: FACP 0x00000000786B0000 000114 (v06 AMAZON AMZNFACP 00000001 AMZN 00000001) May 7 23:46:34.165350 kernel: ACPI: DSDT 0x0000000078640000 00159D (v02 AMAZON AMZNDSDT 00000001 INTL 20160527) May 7 23:46:34.165365 kernel: ACPI: APIC 0x00000000786C0000 000108 (v04 AMAZON AMZNAPIC 00000001 AMZN 00000001) May 7 23:46:34.165380 kernel: ACPI: SPCR 0x00000000786A0000 000050 (v02 AMAZON AMZNSPCR 00000001 AMZN 00000001) May 7 23:46:34.165395 kernel: ACPI: GTDT 0x0000000078690000 000060 (v02 AMAZON AMZNGTDT 00000001 AMZN 00000001) May 7 23:46:34.165413 kernel: ACPI: MCFG 0x0000000078680000 00003C (v02 AMAZON AMZNMCFG 00000001 AMZN 00000001) May 7 23:46:34.165434 kernel: ACPI: SLIT 0x0000000078670000 00002D (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) May 7 23:46:34.165451 kernel: ACPI: IORT 0x0000000078660000 000078 (v01 AMAZON AMZNIORT 00000001 AMZN 00000001) May 7 23:46:34.165467 kernel: ACPI: PPTT 0x0000000078650000 0000EC (v01 AMAZON AMZNPPTT 00000001 AMZN 00000001) May 7 23:46:34.165483 kernel: ACPI: SPCR: console: uart,mmio,0x90a0000,115200 May 7 23:46:34.165499 kernel: earlycon: uart0 at MMIO 0x00000000090a0000 (options '115200') May 7 23:46:34.165515 kernel: printk: bootconsole [uart0] enabled May 7 23:46:34.165531 kernel: NUMA: Failed to initialise from firmware May 7 23:46:34.165548 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000004b5ffffff] May 7 23:46:34.165564 kernel: NUMA: NODE_DATA [mem 0x4b583f800-0x4b5844fff] May 7 23:46:34.165580 kernel: Zone ranges: May 7 23:46:34.165612 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] May 7 23:46:34.165635 kernel: DMA32 empty May 7 23:46:34.165652 kernel: Normal [mem 0x0000000100000000-0x00000004b5ffffff] May 7 23:46:34.165668 kernel: Movable zone start for each node May 7 23:46:34.165684 kernel: Early memory node ranges May 7 23:46:34.165700 kernel: node 0: [mem 0x0000000040000000-0x000000007862ffff] May 7 23:46:34.165716 kernel: node 0: [mem 0x0000000078630000-0x000000007863ffff] May 7 23:46:34.165732 kernel: node 0: [mem 0x0000000078640000-0x00000000786effff] May 7 23:46:34.165748 kernel: node 0: [mem 0x00000000786f0000-0x000000007872ffff] May 7 23:46:34.165764 kernel: node 0: [mem 0x0000000078730000-0x000000007bbfffff] May 7 23:46:34.165780 kernel: node 0: [mem 0x000000007bc00000-0x000000007bfdffff] May 7 23:46:34.165796 kernel: node 0: [mem 0x000000007bfe0000-0x000000007fffffff] May 7 23:46:34.165812 kernel: node 0: [mem 0x0000000400000000-0x00000004b5ffffff] May 7 23:46:34.165832 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000004b5ffffff] May 7 23:46:34.165849 kernel: On node 0, zone Normal: 8192 pages in unavailable ranges May 7 23:46:34.165872 kernel: psci: probing for conduit method from ACPI. May 7 23:46:34.165889 kernel: psci: PSCIv1.0 detected in firmware. May 7 23:46:34.165906 kernel: psci: Using standard PSCI v0.2 function IDs May 7 23:46:34.165927 kernel: psci: Trusted OS migration not required May 7 23:46:34.165944 kernel: psci: SMC Calling Convention v1.1 May 7 23:46:34.165960 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 May 7 23:46:34.165978 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 May 7 23:46:34.165995 kernel: pcpu-alloc: [0] 0 [0] 1 May 7 23:46:34.166011 kernel: Detected PIPT I-cache on CPU0 May 7 23:46:34.166028 kernel: CPU features: detected: GIC system register CPU interface May 7 23:46:34.166045 kernel: CPU features: detected: Spectre-v2 May 7 23:46:34.166061 kernel: CPU features: detected: Spectre-v3a May 7 23:46:34.166094 kernel: CPU features: detected: Spectre-BHB May 7 23:46:34.173391 kernel: CPU features: detected: ARM erratum 1742098 May 7 23:46:34.173422 kernel: CPU features: detected: ARM errata 1165522, 1319367, or 1530923 May 7 23:46:34.173452 kernel: alternatives: applying boot alternatives May 7 23:46:34.173471 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=82f9441f083668f7b43f8fe99c3dc9ee441b8a3ef2f63ecd1e548de4dde5b207 May 7 23:46:34.173490 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 7 23:46:34.173508 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 7 23:46:34.173525 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 7 23:46:34.173542 kernel: Fallback order for Node 0: 0 May 7 23:46:34.173559 kernel: Built 1 zonelists, mobility grouping on. Total pages: 991872 May 7 23:46:34.173576 kernel: Policy zone: Normal May 7 23:46:34.173592 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 7 23:46:34.173609 kernel: software IO TLB: area num 2. May 7 23:46:34.173631 kernel: software IO TLB: mapped [mem 0x000000007c000000-0x0000000080000000] (64MB) May 7 23:46:34.173649 kernel: Memory: 3821176K/4030464K available (10368K kernel code, 2186K rwdata, 8100K rodata, 38336K init, 897K bss, 209288K reserved, 0K cma-reserved) May 7 23:46:34.173666 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 7 23:46:34.173683 kernel: rcu: Preemptible hierarchical RCU implementation. May 7 23:46:34.173701 kernel: rcu: RCU event tracing is enabled. May 7 23:46:34.173718 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 7 23:46:34.173736 kernel: Trampoline variant of Tasks RCU enabled. May 7 23:46:34.173753 kernel: Tracing variant of Tasks RCU enabled. May 7 23:46:34.173770 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 7 23:46:34.173787 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 7 23:46:34.173804 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 May 7 23:46:34.173826 kernel: GICv3: 96 SPIs implemented May 7 23:46:34.173843 kernel: GICv3: 0 Extended SPIs implemented May 7 23:46:34.173859 kernel: Root IRQ handler: gic_handle_irq May 7 23:46:34.173876 kernel: GICv3: GICv3 features: 16 PPIs May 7 23:46:34.173892 kernel: GICv3: CPU0: found redistributor 0 region 0:0x0000000010200000 May 7 23:46:34.173909 kernel: ITS [mem 0x10080000-0x1009ffff] May 7 23:46:34.173926 kernel: ITS@0x0000000010080000: allocated 8192 Devices @4000b0000 (indirect, esz 8, psz 64K, shr 1) May 7 23:46:34.173943 kernel: ITS@0x0000000010080000: allocated 8192 Interrupt Collections @4000c0000 (flat, esz 8, psz 64K, shr 1) May 7 23:46:34.173960 kernel: GICv3: using LPI property table @0x00000004000d0000 May 7 23:46:34.173977 kernel: ITS: Using hypervisor restricted LPI range [128] May 7 23:46:34.173994 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000004000e0000 May 7 23:46:34.174011 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 7 23:46:34.174033 kernel: arch_timer: cp15 timer(s) running at 83.33MHz (virt). May 7 23:46:34.174050 kernel: clocksource: arch_sys_counter: mask: 0x1ffffffffffffff max_cycles: 0x13381ebeec, max_idle_ns: 440795203145 ns May 7 23:46:34.174067 kernel: sched_clock: 57 bits at 83MHz, resolution 12ns, wraps every 4398046511100ns May 7 23:46:34.174105 kernel: Console: colour dummy device 80x25 May 7 23:46:34.174159 kernel: printk: console [tty1] enabled May 7 23:46:34.174177 kernel: ACPI: Core revision 20230628 May 7 23:46:34.174195 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 166.66 BogoMIPS (lpj=83333) May 7 23:46:34.174213 kernel: pid_max: default: 32768 minimum: 301 May 7 23:46:34.174230 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity May 7 23:46:34.174248 kernel: landlock: Up and running. May 7 23:46:34.174272 kernel: SELinux: Initializing. May 7 23:46:34.174290 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 7 23:46:34.174308 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 7 23:46:34.174328 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 7 23:46:34.174345 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 7 23:46:34.174363 kernel: rcu: Hierarchical SRCU implementation. May 7 23:46:34.174381 kernel: rcu: Max phase no-delay instances is 400. May 7 23:46:34.174398 kernel: Platform MSI: ITS@0x10080000 domain created May 7 23:46:34.174420 kernel: PCI/MSI: ITS@0x10080000 domain created May 7 23:46:34.174438 kernel: Remapping and enabling EFI services. May 7 23:46:34.174456 kernel: smp: Bringing up secondary CPUs ... May 7 23:46:34.174473 kernel: Detected PIPT I-cache on CPU1 May 7 23:46:34.174490 kernel: GICv3: CPU1: found redistributor 1 region 0:0x0000000010220000 May 7 23:46:34.174508 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000004000f0000 May 7 23:46:34.174525 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd083] May 7 23:46:34.174542 kernel: smp: Brought up 1 node, 2 CPUs May 7 23:46:34.174560 kernel: SMP: Total of 2 processors activated. May 7 23:46:34.174577 kernel: CPU features: detected: 32-bit EL0 Support May 7 23:46:34.174600 kernel: CPU features: detected: 32-bit EL1 Support May 7 23:46:34.174618 kernel: CPU features: detected: CRC32 instructions May 7 23:46:34.174647 kernel: CPU: All CPU(s) started at EL1 May 7 23:46:34.174669 kernel: alternatives: applying system-wide alternatives May 7 23:46:34.174702 kernel: devtmpfs: initialized May 7 23:46:34.174723 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 7 23:46:34.174742 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 7 23:46:34.174763 kernel: pinctrl core: initialized pinctrl subsystem May 7 23:46:34.174809 kernel: SMBIOS 3.0.0 present. May 7 23:46:34.174880 kernel: DMI: Amazon EC2 a1.large/, BIOS 1.0 11/1/2018 May 7 23:46:34.174920 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 7 23:46:34.174943 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations May 7 23:46:34.174967 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations May 7 23:46:34.174987 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations May 7 23:46:34.175006 kernel: audit: initializing netlink subsys (disabled) May 7 23:46:34.175024 kernel: audit: type=2000 audit(0.219:1): state=initialized audit_enabled=0 res=1 May 7 23:46:34.175049 kernel: thermal_sys: Registered thermal governor 'step_wise' May 7 23:46:34.175067 kernel: cpuidle: using governor menu May 7 23:46:34.175085 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. May 7 23:46:34.175103 kernel: ASID allocator initialised with 65536 entries May 7 23:46:34.176355 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 7 23:46:34.176375 kernel: Serial: AMBA PL011 UART driver May 7 23:46:34.176393 kernel: Modules: 17744 pages in range for non-PLT usage May 7 23:46:34.176411 kernel: Modules: 509264 pages in range for PLT usage May 7 23:46:34.176430 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages May 7 23:46:34.176457 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page May 7 23:46:34.176476 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages May 7 23:46:34.176494 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page May 7 23:46:34.176512 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 7 23:46:34.176530 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page May 7 23:46:34.176548 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages May 7 23:46:34.176566 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page May 7 23:46:34.176584 kernel: ACPI: Added _OSI(Module Device) May 7 23:46:34.176602 kernel: ACPI: Added _OSI(Processor Device) May 7 23:46:34.176625 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 7 23:46:34.176643 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 7 23:46:34.176661 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 7 23:46:34.176679 kernel: ACPI: Interpreter enabled May 7 23:46:34.176697 kernel: ACPI: Using GIC for interrupt routing May 7 23:46:34.176715 kernel: ACPI: MCFG table detected, 1 entries May 7 23:46:34.176733 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-0f]) May 7 23:46:34.177044 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] May 7 23:46:34.177290 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] May 7 23:46:34.177491 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] May 7 23:46:34.177694 kernel: acpi PNP0A08:00: ECAM area [mem 0x20000000-0x20ffffff] reserved by PNP0C02:00 May 7 23:46:34.177891 kernel: acpi PNP0A08:00: ECAM at [mem 0x20000000-0x20ffffff] for [bus 00-0f] May 7 23:46:34.177916 kernel: ACPI: Remapped I/O 0x000000001fff0000 to [io 0x0000-0xffff window] May 7 23:46:34.177935 kernel: acpiphp: Slot [1] registered May 7 23:46:34.177953 kernel: acpiphp: Slot [2] registered May 7 23:46:34.177971 kernel: acpiphp: Slot [3] registered May 7 23:46:34.177995 kernel: acpiphp: Slot [4] registered May 7 23:46:34.178026 kernel: acpiphp: Slot [5] registered May 7 23:46:34.178046 kernel: acpiphp: Slot [6] registered May 7 23:46:34.178065 kernel: acpiphp: Slot [7] registered May 7 23:46:34.178100 kernel: acpiphp: Slot [8] registered May 7 23:46:34.178138 kernel: acpiphp: Slot [9] registered May 7 23:46:34.178157 kernel: acpiphp: Slot [10] registered May 7 23:46:34.178176 kernel: acpiphp: Slot [11] registered May 7 23:46:34.178194 kernel: acpiphp: Slot [12] registered May 7 23:46:34.178212 kernel: acpiphp: Slot [13] registered May 7 23:46:34.178236 kernel: acpiphp: Slot [14] registered May 7 23:46:34.178254 kernel: acpiphp: Slot [15] registered May 7 23:46:34.178271 kernel: acpiphp: Slot [16] registered May 7 23:46:34.178289 kernel: acpiphp: Slot [17] registered May 7 23:46:34.178307 kernel: acpiphp: Slot [18] registered May 7 23:46:34.178325 kernel: acpiphp: Slot [19] registered May 7 23:46:34.178343 kernel: acpiphp: Slot [20] registered May 7 23:46:34.178361 kernel: acpiphp: Slot [21] registered May 7 23:46:34.178378 kernel: acpiphp: Slot [22] registered May 7 23:46:34.178400 kernel: acpiphp: Slot [23] registered May 7 23:46:34.178418 kernel: acpiphp: Slot [24] registered May 7 23:46:34.178436 kernel: acpiphp: Slot [25] registered May 7 23:46:34.178454 kernel: acpiphp: Slot [26] registered May 7 23:46:34.178472 kernel: acpiphp: Slot [27] registered May 7 23:46:34.178490 kernel: acpiphp: Slot [28] registered May 7 23:46:34.183235 kernel: acpiphp: Slot [29] registered May 7 23:46:34.183258 kernel: acpiphp: Slot [30] registered May 7 23:46:34.183277 kernel: acpiphp: Slot [31] registered May 7 23:46:34.183310 kernel: PCI host bridge to bus 0000:00 May 7 23:46:34.183565 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xffffffff window] May 7 23:46:34.183764 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] May 7 23:46:34.183959 kernel: pci_bus 0000:00: root bus resource [mem 0x400000000000-0x407fffffffff window] May 7 23:46:34.184627 kernel: pci_bus 0000:00: root bus resource [bus 00-0f] May 7 23:46:34.184882 kernel: pci 0000:00:00.0: [1d0f:0200] type 00 class 0x060000 May 7 23:46:34.186315 kernel: pci 0000:00:01.0: [1d0f:8250] type 00 class 0x070003 May 7 23:46:34.186576 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80118000-0x80118fff] May 7 23:46:34.186796 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 May 7 23:46:34.187003 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80114000-0x80117fff] May 7 23:46:34.187285 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D2 D3hot D3cold May 7 23:46:34.187537 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 May 7 23:46:34.187755 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80110000-0x80113fff] May 7 23:46:34.187962 kernel: pci 0000:00:05.0: reg 0x18: [mem 0x80000000-0x800fffff pref] May 7 23:46:34.188484 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x80100000-0x8010ffff] May 7 23:46:34.188731 kernel: pci 0000:00:05.0: PME# supported from D0 D1 D2 D3hot D3cold May 7 23:46:34.188938 kernel: pci 0000:00:05.0: BAR 2: assigned [mem 0x80000000-0x800fffff pref] May 7 23:46:34.189173 kernel: pci 0000:00:05.0: BAR 4: assigned [mem 0x80100000-0x8010ffff] May 7 23:46:34.189411 kernel: pci 0000:00:04.0: BAR 0: assigned [mem 0x80110000-0x80113fff] May 7 23:46:34.189643 kernel: pci 0000:00:05.0: BAR 0: assigned [mem 0x80114000-0x80117fff] May 7 23:46:34.189901 kernel: pci 0000:00:01.0: BAR 0: assigned [mem 0x80118000-0x80118fff] May 7 23:46:34.190168 kernel: pci_bus 0000:00: resource 4 [mem 0x80000000-0xffffffff window] May 7 23:46:34.190762 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] May 7 23:46:34.190967 kernel: pci_bus 0000:00: resource 6 [mem 0x400000000000-0x407fffffffff window] May 7 23:46:34.190993 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 May 7 23:46:34.191013 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 May 7 23:46:34.191034 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 May 7 23:46:34.191072 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 May 7 23:46:34.191092 kernel: iommu: Default domain type: Translated May 7 23:46:34.191152 kernel: iommu: DMA domain TLB invalidation policy: strict mode May 7 23:46:34.191173 kernel: efivars: Registered efivars operations May 7 23:46:34.191193 kernel: vgaarb: loaded May 7 23:46:34.191214 kernel: clocksource: Switched to clocksource arch_sys_counter May 7 23:46:34.191233 kernel: VFS: Disk quotas dquot_6.6.0 May 7 23:46:34.191251 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 7 23:46:34.191270 kernel: pnp: PnP ACPI init May 7 23:46:34.191518 kernel: system 00:00: [mem 0x20000000-0x2fffffff] could not be reserved May 7 23:46:34.191556 kernel: pnp: PnP ACPI: found 1 devices May 7 23:46:34.191577 kernel: NET: Registered PF_INET protocol family May 7 23:46:34.191597 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 7 23:46:34.191616 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 7 23:46:34.191635 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 7 23:46:34.191654 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 7 23:46:34.191682 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) May 7 23:46:34.191702 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 7 23:46:34.191721 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 7 23:46:34.191745 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 7 23:46:34.191764 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 7 23:46:34.191782 kernel: PCI: CLS 0 bytes, default 64 May 7 23:46:34.191800 kernel: kvm [1]: HYP mode not available May 7 23:46:34.191818 kernel: Initialise system trusted keyrings May 7 23:46:34.191836 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 7 23:46:34.191855 kernel: Key type asymmetric registered May 7 23:46:34.191872 kernel: Asymmetric key parser 'x509' registered May 7 23:46:34.191891 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 7 23:46:34.191915 kernel: io scheduler mq-deadline registered May 7 23:46:34.191934 kernel: io scheduler kyber registered May 7 23:46:34.191953 kernel: io scheduler bfq registered May 7 23:46:34.194536 kernel: pl061_gpio ARMH0061:00: PL061 GPIO chip registered May 7 23:46:34.194580 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 May 7 23:46:34.194599 kernel: ACPI: button: Power Button [PWRB] May 7 23:46:34.194618 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input1 May 7 23:46:34.194636 kernel: ACPI: button: Sleep Button [SLPB] May 7 23:46:34.194664 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 7 23:46:34.194684 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 May 7 23:46:34.194923 kernel: serial 0000:00:01.0: enabling device (0010 -> 0012) May 7 23:46:34.194951 kernel: printk: console [ttyS0] disabled May 7 23:46:34.194970 kernel: 0000:00:01.0: ttyS0 at MMIO 0x80118000 (irq = 14, base_baud = 115200) is a 16550A May 7 23:46:34.194989 kernel: printk: console [ttyS0] enabled May 7 23:46:34.195007 kernel: printk: bootconsole [uart0] disabled May 7 23:46:34.195025 kernel: thunder_xcv, ver 1.0 May 7 23:46:34.195043 kernel: thunder_bgx, ver 1.0 May 7 23:46:34.195061 kernel: nicpf, ver 1.0 May 7 23:46:34.195085 kernel: nicvf, ver 1.0 May 7 23:46:34.195400 kernel: rtc-efi rtc-efi.0: registered as rtc0 May 7 23:46:34.195611 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-05-07T23:46:33 UTC (1746661593) May 7 23:46:34.195639 kernel: hid: raw HID events driver (C) Jiri Kosina May 7 23:46:34.195659 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 3 counters available May 7 23:46:34.195679 kernel: watchdog: Delayed init of the lockup detector failed: -19 May 7 23:46:34.195698 kernel: watchdog: Hard watchdog permanently disabled May 7 23:46:34.195727 kernel: NET: Registered PF_INET6 protocol family May 7 23:46:34.195747 kernel: Segment Routing with IPv6 May 7 23:46:34.195766 kernel: In-situ OAM (IOAM) with IPv6 May 7 23:46:34.195784 kernel: NET: Registered PF_PACKET protocol family May 7 23:46:34.195802 kernel: Key type dns_resolver registered May 7 23:46:34.195820 kernel: registered taskstats version 1 May 7 23:46:34.195838 kernel: Loading compiled-in X.509 certificates May 7 23:46:34.195857 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.88-flatcar: f45666b1b2057b901dda15e57012558a26abdeb0' May 7 23:46:34.195875 kernel: Key type .fscrypt registered May 7 23:46:34.195893 kernel: Key type fscrypt-provisioning registered May 7 23:46:34.195916 kernel: ima: No TPM chip found, activating TPM-bypass! May 7 23:46:34.195935 kernel: ima: Allocated hash algorithm: sha1 May 7 23:46:34.195953 kernel: ima: No architecture policies found May 7 23:46:34.195971 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) May 7 23:46:34.195989 kernel: clk: Disabling unused clocks May 7 23:46:34.196007 kernel: Freeing unused kernel memory: 38336K May 7 23:46:34.196025 kernel: Run /init as init process May 7 23:46:34.196043 kernel: with arguments: May 7 23:46:34.196077 kernel: /init May 7 23:46:34.196105 kernel: with environment: May 7 23:46:34.196211 kernel: HOME=/ May 7 23:46:34.196230 kernel: TERM=linux May 7 23:46:34.196248 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 7 23:46:34.196268 systemd[1]: Successfully made /usr/ read-only. May 7 23:46:34.196292 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 7 23:46:34.196313 systemd[1]: Detected virtualization amazon. May 7 23:46:34.196340 systemd[1]: Detected architecture arm64. May 7 23:46:34.196360 systemd[1]: Running in initrd. May 7 23:46:34.196379 systemd[1]: No hostname configured, using default hostname. May 7 23:46:34.196399 systemd[1]: Hostname set to . May 7 23:46:34.196418 systemd[1]: Initializing machine ID from VM UUID. May 7 23:46:34.196438 systemd[1]: Queued start job for default target initrd.target. May 7 23:46:34.196457 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 7 23:46:34.196477 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 7 23:46:34.196498 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 7 23:46:34.196523 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 7 23:46:34.196543 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 7 23:46:34.196565 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 7 23:46:34.196587 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 7 23:46:34.196607 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 7 23:46:34.196627 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 7 23:46:34.196651 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 7 23:46:34.196671 systemd[1]: Reached target paths.target - Path Units. May 7 23:46:34.196691 systemd[1]: Reached target slices.target - Slice Units. May 7 23:46:34.196710 systemd[1]: Reached target swap.target - Swaps. May 7 23:46:34.196730 systemd[1]: Reached target timers.target - Timer Units. May 7 23:46:34.196749 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 7 23:46:34.196769 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 7 23:46:34.196789 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 7 23:46:34.196808 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 7 23:46:34.196833 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 7 23:46:34.196852 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 7 23:46:34.196872 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 7 23:46:34.196892 systemd[1]: Reached target sockets.target - Socket Units. May 7 23:46:34.196911 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 7 23:46:34.196931 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 7 23:46:34.196950 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 7 23:46:34.196970 systemd[1]: Starting systemd-fsck-usr.service... May 7 23:46:34.196996 systemd[1]: Starting systemd-journald.service - Journal Service... May 7 23:46:34.197016 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 7 23:46:34.197036 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 7 23:46:34.197104 systemd-journald[252]: Collecting audit messages is disabled. May 7 23:46:34.197168 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 7 23:46:34.197190 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 7 23:46:34.197210 systemd[1]: Finished systemd-fsck-usr.service. May 7 23:46:34.197230 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 7 23:46:34.197261 kernel: Bridge firewalling registered May 7 23:46:34.197289 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 7 23:46:34.197310 systemd-journald[252]: Journal started May 7 23:46:34.197347 systemd-journald[252]: Runtime Journal (/run/log/journal/ec26214e429a6f04acfd1063e3a1b4ba) is 8M, max 75.3M, 67.3M free. May 7 23:46:34.158854 systemd-modules-load[253]: Inserted module 'overlay' May 7 23:46:34.189190 systemd-modules-load[253]: Inserted module 'br_netfilter' May 7 23:46:34.215249 systemd[1]: Started systemd-journald.service - Journal Service. May 7 23:46:34.217811 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 7 23:46:34.222970 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 7 23:46:34.237436 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 7 23:46:34.245025 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 7 23:46:34.260374 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 7 23:46:34.265155 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 7 23:46:34.273346 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 7 23:46:34.302904 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 7 23:46:34.309158 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 7 23:46:34.320437 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 7 23:46:34.328184 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 7 23:46:34.333750 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 7 23:46:34.352745 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 7 23:46:34.361469 dracut-cmdline[286]: dracut-dracut-053 May 7 23:46:34.367577 dracut-cmdline[286]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=82f9441f083668f7b43f8fe99c3dc9ee441b8a3ef2f63ecd1e548de4dde5b207 May 7 23:46:34.439153 systemd-resolved[294]: Positive Trust Anchors: May 7 23:46:34.439181 systemd-resolved[294]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 7 23:46:34.439241 systemd-resolved[294]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 7 23:46:34.530654 kernel: SCSI subsystem initialized May 7 23:46:34.538235 kernel: Loading iSCSI transport class v2.0-870. May 7 23:46:34.551224 kernel: iscsi: registered transport (tcp) May 7 23:46:34.572831 kernel: iscsi: registered transport (qla4xxx) May 7 23:46:34.572903 kernel: QLogic iSCSI HBA Driver May 7 23:46:34.675144 kernel: random: crng init done May 7 23:46:34.673394 systemd-resolved[294]: Defaulting to hostname 'linux'. May 7 23:46:34.675250 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 7 23:46:34.681154 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 7 23:46:34.705023 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 7 23:46:34.714503 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 7 23:46:34.758912 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 7 23:46:34.758986 kernel: device-mapper: uevent: version 1.0.3 May 7 23:46:34.761144 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com May 7 23:46:34.825153 kernel: raid6: neonx8 gen() 6581 MB/s May 7 23:46:34.842146 kernel: raid6: neonx4 gen() 6527 MB/s May 7 23:46:34.859147 kernel: raid6: neonx2 gen() 5405 MB/s May 7 23:46:34.876142 kernel: raid6: neonx1 gen() 3930 MB/s May 7 23:46:34.893141 kernel: raid6: int64x8 gen() 3602 MB/s May 7 23:46:34.910141 kernel: raid6: int64x4 gen() 3701 MB/s May 7 23:46:34.927141 kernel: raid6: int64x2 gen() 3597 MB/s May 7 23:46:34.944948 kernel: raid6: int64x1 gen() 2734 MB/s May 7 23:46:34.944987 kernel: raid6: using algorithm neonx8 gen() 6581 MB/s May 7 23:46:34.962906 kernel: raid6: .... xor() 4674 MB/s, rmw enabled May 7 23:46:34.962953 kernel: raid6: using neon recovery algorithm May 7 23:46:34.970145 kernel: xor: measuring software checksum speed May 7 23:46:34.971141 kernel: 8regs : 11490 MB/sec May 7 23:46:34.972142 kernel: 32regs : 11753 MB/sec May 7 23:46:34.974238 kernel: arm64_neon : 9022 MB/sec May 7 23:46:34.974281 kernel: xor: using function: 32regs (11753 MB/sec) May 7 23:46:35.057159 kernel: Btrfs loaded, zoned=no, fsverity=no May 7 23:46:35.075934 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 7 23:46:35.085429 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 7 23:46:35.131570 systemd-udevd[473]: Using default interface naming scheme 'v255'. May 7 23:46:35.142020 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 7 23:46:35.156351 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 7 23:46:35.189684 dracut-pre-trigger[478]: rd.md=0: removing MD RAID activation May 7 23:46:35.243001 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 7 23:46:35.253391 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 7 23:46:35.374923 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 7 23:46:35.398253 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 7 23:46:35.437887 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 7 23:46:35.442367 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 7 23:46:35.444752 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 7 23:46:35.446993 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 7 23:46:35.470244 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 7 23:46:35.500923 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 7 23:46:35.593240 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 May 7 23:46:35.593302 kernel: ena 0000:00:05.0: enabling device (0010 -> 0012) May 7 23:46:35.626334 kernel: ena 0000:00:05.0: ENA device version: 0.10 May 7 23:46:35.626686 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 May 7 23:46:35.626923 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80114000, mac addr 06:69:02:c4:41:e9 May 7 23:46:35.627182 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 May 7 23:46:35.627221 kernel: nvme nvme0: pci function 0000:00:04.0 May 7 23:46:35.596570 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 7 23:46:35.596809 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 7 23:46:35.601900 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 7 23:46:35.606432 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 7 23:46:35.606729 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 7 23:46:35.609174 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 7 23:46:35.618730 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 7 23:46:35.623142 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 7 23:46:35.653164 kernel: nvme nvme0: 2/0/0 default/read/poll queues May 7 23:46:35.664155 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 7 23:46:35.664233 kernel: GPT:9289727 != 16777215 May 7 23:46:35.665372 kernel: GPT:Alternate GPT header not at the end of the disk. May 7 23:46:35.667050 kernel: GPT:9289727 != 16777215 May 7 23:46:35.667084 kernel: GPT: Use GNU Parted to correct GPT errors. May 7 23:46:35.668140 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 7 23:46:35.671645 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 7 23:46:35.675939 (udev-worker)[536]: Network interface NamePolicy= disabled on kernel command line. May 7 23:46:35.688406 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 7 23:46:35.724749 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 7 23:46:35.781605 kernel: BTRFS: device fsid a4d66dad-2d34-4ed0-87a7-f6519531b08f devid 1 transid 42 /dev/nvme0n1p3 scanned by (udev-worker) (520) May 7 23:46:35.792268 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by (udev-worker) (525) May 7 23:46:35.885302 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. May 7 23:46:35.924982 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. May 7 23:46:35.930025 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. May 7 23:46:35.956684 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. May 7 23:46:35.981599 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. May 7 23:46:36.005389 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 7 23:46:36.020339 disk-uuid[661]: Primary Header is updated. May 7 23:46:36.020339 disk-uuid[661]: Secondary Entries is updated. May 7 23:46:36.020339 disk-uuid[661]: Secondary Header is updated. May 7 23:46:36.030174 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 7 23:46:37.050142 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 May 7 23:46:37.051434 disk-uuid[662]: The operation has completed successfully. May 7 23:46:37.237209 systemd[1]: disk-uuid.service: Deactivated successfully. May 7 23:46:37.239195 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 7 23:46:37.345376 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 7 23:46:37.354037 sh[922]: Success May 7 23:46:37.378161 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" May 7 23:46:37.500189 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 7 23:46:37.515360 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 7 23:46:37.523040 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 7 23:46:37.570211 kernel: BTRFS info (device dm-0): first mount of filesystem a4d66dad-2d34-4ed0-87a7-f6519531b08f May 7 23:46:37.570284 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm May 7 23:46:37.572002 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead May 7 23:46:37.573295 kernel: BTRFS info (device dm-0): disabling log replay at mount time May 7 23:46:37.573332 kernel: BTRFS info (device dm-0): using free space tree May 7 23:46:37.602151 kernel: BTRFS info (device dm-0): enabling ssd optimizations May 7 23:46:37.616922 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 7 23:46:37.620632 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 7 23:46:37.633561 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 7 23:46:37.639487 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 7 23:46:37.678560 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 28594331-30e6-4c58-8ddc-9d8448a320bb May 7 23:46:37.678634 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm May 7 23:46:37.679856 kernel: BTRFS info (device nvme0n1p6): using free space tree May 7 23:46:37.687148 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations May 7 23:46:37.695180 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 28594331-30e6-4c58-8ddc-9d8448a320bb May 7 23:46:37.704091 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 7 23:46:37.716412 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 7 23:46:37.830291 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 7 23:46:37.846765 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 7 23:46:37.885801 ignition[1025]: Ignition 2.20.0 May 7 23:46:37.888767 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 7 23:46:37.885830 ignition[1025]: Stage: fetch-offline May 7 23:46:37.886278 ignition[1025]: no configs at "/usr/lib/ignition/base.d" May 7 23:46:37.886303 ignition[1025]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 7 23:46:37.886739 ignition[1025]: Ignition finished successfully May 7 23:46:37.931770 systemd-networkd[1117]: lo: Link UP May 7 23:46:37.931785 systemd-networkd[1117]: lo: Gained carrier May 7 23:46:37.935362 systemd-networkd[1117]: Enumeration completed May 7 23:46:37.936944 systemd-networkd[1117]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 7 23:46:37.936952 systemd-networkd[1117]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 7 23:46:37.937255 systemd[1]: Started systemd-networkd.service - Network Configuration. May 7 23:46:37.940639 systemd[1]: Reached target network.target - Network. May 7 23:46:37.949794 systemd-networkd[1117]: eth0: Link UP May 7 23:46:37.949801 systemd-networkd[1117]: eth0: Gained carrier May 7 23:46:37.949817 systemd-networkd[1117]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 7 23:46:37.963547 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 7 23:46:37.977246 systemd-networkd[1117]: eth0: DHCPv4 address 172.31.28.15/20, gateway 172.31.16.1 acquired from 172.31.16.1 May 7 23:46:37.994600 ignition[1121]: Ignition 2.20.0 May 7 23:46:37.994642 ignition[1121]: Stage: fetch May 7 23:46:37.995631 ignition[1121]: no configs at "/usr/lib/ignition/base.d" May 7 23:46:37.995658 ignition[1121]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 7 23:46:37.995829 ignition[1121]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 7 23:46:38.007266 ignition[1121]: PUT result: OK May 7 23:46:38.010124 ignition[1121]: parsed url from cmdline: "" May 7 23:46:38.010152 ignition[1121]: no config URL provided May 7 23:46:38.010168 ignition[1121]: reading system config file "/usr/lib/ignition/user.ign" May 7 23:46:38.010195 ignition[1121]: no config at "/usr/lib/ignition/user.ign" May 7 23:46:38.010232 ignition[1121]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 7 23:46:38.011914 ignition[1121]: PUT result: OK May 7 23:46:38.013925 ignition[1121]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 May 7 23:46:38.021139 ignition[1121]: GET result: OK May 7 23:46:38.022496 ignition[1121]: parsing config with SHA512: 92ed82bfcc8201f247ba46c6328d99e234f8692aa8c4a1119017d0f5a77c1a00bbb1c6e21f0eefcf6cff194d21863a21256940b0ef6d7f5114a0aba79e0446a7 May 7 23:46:38.032226 unknown[1121]: fetched base config from "system" May 7 23:46:38.032268 unknown[1121]: fetched base config from "system" May 7 23:46:38.032282 unknown[1121]: fetched user config from "aws" May 7 23:46:38.034496 ignition[1121]: fetch: fetch complete May 7 23:46:38.034513 ignition[1121]: fetch: fetch passed May 7 23:46:38.034604 ignition[1121]: Ignition finished successfully May 7 23:46:38.043905 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 7 23:46:38.055429 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 7 23:46:38.085597 ignition[1128]: Ignition 2.20.0 May 7 23:46:38.086444 ignition[1128]: Stage: kargs May 7 23:46:38.087092 ignition[1128]: no configs at "/usr/lib/ignition/base.d" May 7 23:46:38.087165 ignition[1128]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 7 23:46:38.087318 ignition[1128]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 7 23:46:38.089945 ignition[1128]: PUT result: OK May 7 23:46:38.100330 ignition[1128]: kargs: kargs passed May 7 23:46:38.100918 ignition[1128]: Ignition finished successfully May 7 23:46:38.106198 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 7 23:46:38.118375 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 7 23:46:38.142924 ignition[1134]: Ignition 2.20.0 May 7 23:46:38.142949 ignition[1134]: Stage: disks May 7 23:46:38.144284 ignition[1134]: no configs at "/usr/lib/ignition/base.d" May 7 23:46:38.144313 ignition[1134]: no config dir at "/usr/lib/ignition/base.platform.d/aws" May 7 23:46:38.144505 ignition[1134]: PUT http://169.254.169.254/latest/api/token: attempt #1 May 7 23:46:38.147415 ignition[1134]: PUT result: OK May 7 23:46:38.157254 ignition[1134]: disks: disks passed May 7 23:46:38.157352 ignition[1134]: Ignition finished successfully May 7 23:46:38.159601 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 7 23:46:38.163979 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 7 23:46:38.166983 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 7 23:46:38.175909 systemd[1]: Reached target local-fs.target - Local File Systems. May 7 23:46:38.177972 systemd[1]: Reached target sysinit.target - System Initialization. May 7 23:46:38.179908 systemd[1]: Reached target basic.target - Basic System. May 7 23:46:38.193471 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 7 23:46:38.235917 systemd-fsck[1143]: ROOT: clean, 14/553520 files, 52654/553472 blocks May 7 23:46:38.244677 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 7 23:46:38.256302 systemd[1]: Mounting sysroot.mount - /sysroot... May 7 23:46:38.352131 kernel: EXT4-fs (nvme0n1p9): mounted filesystem f291ddc8-664e-45dc-bbf9-8344dca1a297 r/w with ordered data mode. Quota mode: none. May 7 23:46:38.353160 systemd[1]: Mounted sysroot.mount - /sysroot. May 7 23:46:38.356718 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 7 23:46:38.380271 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 7 23:46:38.386020 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 7 23:46:38.390304 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. May 7 23:46:38.390402 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 7 23:46:38.390457 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 7 23:46:38.409155 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 scanned by mount (1162) May 7 23:46:38.412786 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 28594331-30e6-4c58-8ddc-9d8448a320bb May 7 23:46:38.412855 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm May 7 23:46:38.414328 kernel: BTRFS info (device nvme0n1p6): using free space tree May 7 23:46:38.417423 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 7 23:46:38.426483 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 7 23:46:38.433149 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations May 7 23:46:38.437245 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 7 23:46:38.532817 initrd-setup-root[1186]: cut: /sysroot/etc/passwd: No such file or directory May 7 23:46:38.542596 initrd-setup-root[1193]: cut: /sysroot/etc/group: No such file or directory May 7 23:46:38.552264 initrd-setup-root[1200]: cut: /sysroot/etc/shadow: No such file or directory May 7 23:46:38.561450 initrd-setup-root[1207]: cut: /sysroot/etc/gshadow: No such file or directory May 7 23:46:38.727506 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 7 23:46:38.741409 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 7 23:46:38.750428 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 7 23:46:38.762713 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 7 23:46:38.765317 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 28594331-30e6-4c58-8ddc-9d8448a320bb May 7 23:46:38.805243 ignition[1280]: INFO : Ignition 2.20.0 May 7 23:46:38.805243 ignition[1280]: INFO : Stage: mount May 7 23:46:38.805243 ignition[1280]: INFO : no configs at "/usr/lib/ignition/base.d" May 7 23:46:38.805243 ignition[1280]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 7 23:46:38.805205 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 7 23:46:38.818263 ignition[1280]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 7 23:46:38.818263 ignition[1280]: INFO : PUT result: OK May 7 23:46:38.823581 ignition[1280]: INFO : mount: mount passed May 7 23:46:38.826756 ignition[1280]: INFO : Ignition finished successfully May 7 23:46:38.826546 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 7 23:46:38.838308 systemd[1]: Starting ignition-files.service - Ignition (files)... May 7 23:46:38.866480 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 7 23:46:38.891640 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/nvme0n1p6 scanned by mount (1292) May 7 23:46:38.891707 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 28594331-30e6-4c58-8ddc-9d8448a320bb May 7 23:46:38.891733 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm May 7 23:46:38.894271 kernel: BTRFS info (device nvme0n1p6): using free space tree May 7 23:46:38.899155 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations May 7 23:46:38.903276 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 7 23:46:38.933808 ignition[1308]: INFO : Ignition 2.20.0 May 7 23:46:38.933808 ignition[1308]: INFO : Stage: files May 7 23:46:38.937969 ignition[1308]: INFO : no configs at "/usr/lib/ignition/base.d" May 7 23:46:38.937969 ignition[1308]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 7 23:46:38.937969 ignition[1308]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 7 23:46:38.937969 ignition[1308]: INFO : PUT result: OK May 7 23:46:38.948463 ignition[1308]: DEBUG : files: compiled without relabeling support, skipping May 7 23:46:38.952355 ignition[1308]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 7 23:46:38.952355 ignition[1308]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 7 23:46:38.960771 ignition[1308]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 7 23:46:38.963458 ignition[1308]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 7 23:46:38.966516 unknown[1308]: wrote ssh authorized keys file for user: core May 7 23:46:38.970294 ignition[1308]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 7 23:46:38.973203 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" May 7 23:46:38.973203 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 May 7 23:46:39.058647 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 7 23:46:39.212775 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" May 7 23:46:39.216396 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 7 23:46:39.216396 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 May 7 23:46:39.390306 systemd-networkd[1117]: eth0: Gained IPv6LL May 7 23:46:39.677914 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 7 23:46:39.835527 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 7 23:46:39.838839 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" May 7 23:46:39.838839 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" May 7 23:46:39.838839 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" May 7 23:46:39.838839 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" May 7 23:46:39.838839 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" May 7 23:46:39.855151 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-arm64.raw: attempt #1 May 7 23:46:40.181743 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK May 7 23:46:40.522959 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" May 7 23:46:40.522959 ignition[1308]: INFO : files: op(c): [started] processing unit "prepare-helm.service" May 7 23:46:40.529398 ignition[1308]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 7 23:46:40.529398 ignition[1308]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 7 23:46:40.529398 ignition[1308]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" May 7 23:46:40.529398 ignition[1308]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" May 7 23:46:40.529398 ignition[1308]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" May 7 23:46:40.529398 ignition[1308]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" May 7 23:46:40.529398 ignition[1308]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" May 7 23:46:40.529398 ignition[1308]: INFO : files: files passed May 7 23:46:40.529398 ignition[1308]: INFO : Ignition finished successfully May 7 23:46:40.554411 systemd[1]: Finished ignition-files.service - Ignition (files). May 7 23:46:40.563418 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 7 23:46:40.573683 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 7 23:46:40.594793 systemd[1]: ignition-quench.service: Deactivated successfully. May 7 23:46:40.594997 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 7 23:46:40.611886 initrd-setup-root-after-ignition[1337]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 7 23:46:40.611886 initrd-setup-root-after-ignition[1337]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 7 23:46:40.618590 initrd-setup-root-after-ignition[1341]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 7 23:46:40.625188 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 7 23:46:40.628448 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 7 23:46:40.648452 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 7 23:46:40.712043 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 7 23:46:40.712347 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 7 23:46:40.718936 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 7 23:46:40.721066 systemd[1]: Reached target initrd.target - Initrd Default Target. May 7 23:46:40.723158 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 7 23:46:40.732437 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 7 23:46:40.761181 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 7 23:46:40.773443 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 7 23:46:40.797977 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 7 23:46:40.802682 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 7 23:46:40.807161 systemd[1]: Stopped target timers.target - Timer Units. May 7 23:46:40.809407 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 7 23:46:40.809737 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 7 23:46:40.816896 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 7 23:46:40.819791 systemd[1]: Stopped target basic.target - Basic System. May 7 23:46:40.825517 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 7 23:46:40.828173 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 7 23:46:40.834152 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 7 23:46:40.836807 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 7 23:46:40.842094 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 7 23:46:40.845077 systemd[1]: Stopped target sysinit.target - System Initialization. May 7 23:46:40.849366 systemd[1]: Stopped target local-fs.target - Local File Systems. May 7 23:46:40.854830 systemd[1]: Stopped target swap.target - Swaps. May 7 23:46:40.856896 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 7 23:46:40.857197 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 7 23:46:40.863540 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 7 23:46:40.865821 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 7 23:46:40.873036 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 7 23:46:40.873299 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 7 23:46:40.877568 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 7 23:46:40.877786 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 7 23:46:40.882172 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 7 23:46:40.882507 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 7 23:46:40.888695 systemd[1]: ignition-files.service: Deactivated successfully. May 7 23:46:40.889470 systemd[1]: Stopped ignition-files.service - Ignition (files). May 7 23:46:40.907156 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 7 23:46:40.917523 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 7 23:46:40.927981 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 7 23:46:40.932617 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 7 23:46:40.941263 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 7 23:46:40.943499 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 7 23:46:40.955840 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 7 23:46:40.963066 ignition[1361]: INFO : Ignition 2.20.0 May 7 23:46:40.963066 ignition[1361]: INFO : Stage: umount May 7 23:46:40.963870 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 7 23:46:40.964523 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 7 23:46:40.974323 ignition[1361]: INFO : no configs at "/usr/lib/ignition/base.d" May 7 23:46:40.974323 ignition[1361]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" May 7 23:46:40.974323 ignition[1361]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 May 7 23:46:40.980765 ignition[1361]: INFO : PUT result: OK May 7 23:46:40.985308 ignition[1361]: INFO : umount: umount passed May 7 23:46:40.987976 ignition[1361]: INFO : Ignition finished successfully May 7 23:46:40.988659 systemd[1]: ignition-mount.service: Deactivated successfully. May 7 23:46:40.988896 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 7 23:46:40.996318 systemd[1]: ignition-disks.service: Deactivated successfully. May 7 23:46:40.996493 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 7 23:46:41.003085 systemd[1]: ignition-kargs.service: Deactivated successfully. May 7 23:46:41.004363 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 7 23:46:41.006985 systemd[1]: ignition-fetch.service: Deactivated successfully. May 7 23:46:41.007074 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 7 23:46:41.011204 systemd[1]: Stopped target network.target - Network. May 7 23:46:41.011397 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 7 23:46:41.011492 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 7 23:46:41.011749 systemd[1]: Stopped target paths.target - Path Units. May 7 23:46:41.012003 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 7 23:46:41.019620 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 7 23:46:41.024631 systemd[1]: Stopped target slices.target - Slice Units. May 7 23:46:41.026295 systemd[1]: Stopped target sockets.target - Socket Units. May 7 23:46:41.030535 systemd[1]: iscsid.socket: Deactivated successfully. May 7 23:46:41.030625 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 7 23:46:41.048797 systemd[1]: iscsiuio.socket: Deactivated successfully. May 7 23:46:41.048896 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 7 23:46:41.050865 systemd[1]: ignition-setup.service: Deactivated successfully. May 7 23:46:41.050957 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 7 23:46:41.053293 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 7 23:46:41.053386 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 7 23:46:41.056193 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 7 23:46:41.064838 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 7 23:46:41.090816 systemd[1]: systemd-networkd.service: Deactivated successfully. May 7 23:46:41.091022 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 7 23:46:41.103633 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 7 23:46:41.106459 systemd[1]: systemd-resolved.service: Deactivated successfully. May 7 23:46:41.108276 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 7 23:46:41.116306 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 7 23:46:41.116847 systemd[1]: sysroot-boot.service: Deactivated successfully. May 7 23:46:41.117266 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 7 23:46:41.127889 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 7 23:46:41.127981 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 7 23:46:41.130570 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 7 23:46:41.130669 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 7 23:46:41.140064 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 7 23:46:41.147044 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 7 23:46:41.147193 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 7 23:46:41.150296 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 7 23:46:41.150382 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 7 23:46:41.166161 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 7 23:46:41.166259 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 7 23:46:41.168349 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 7 23:46:41.168440 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 7 23:46:41.174604 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 7 23:46:41.188741 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 7 23:46:41.188885 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 7 23:46:41.203642 systemd[1]: systemd-udevd.service: Deactivated successfully. May 7 23:46:41.203922 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 7 23:46:41.210448 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 7 23:46:41.210566 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 7 23:46:41.220843 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 7 23:46:41.220926 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 7 23:46:41.222980 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 7 23:46:41.223071 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 7 23:46:41.225905 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 7 23:46:41.225991 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 7 23:46:41.237384 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 7 23:46:41.237481 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 7 23:46:41.257291 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 7 23:46:41.263278 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 7 23:46:41.263432 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 7 23:46:41.267899 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. May 7 23:46:41.270247 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 7 23:46:41.277022 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 7 23:46:41.277263 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 7 23:46:41.283493 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 7 23:46:41.283598 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 7 23:46:41.289513 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 7 23:46:41.289628 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 7 23:46:41.293752 systemd[1]: network-cleanup.service: Deactivated successfully. May 7 23:46:41.294837 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 7 23:46:41.308820 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 7 23:46:41.308998 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 7 23:46:41.314768 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 7 23:46:41.332752 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 7 23:46:41.349263 systemd[1]: Switching root. May 7 23:46:41.385488 systemd-journald[252]: Journal stopped May 7 23:46:43.390015 systemd-journald[252]: Received SIGTERM from PID 1 (systemd). May 7 23:46:43.390182 kernel: SELinux: policy capability network_peer_controls=1 May 7 23:46:43.390228 kernel: SELinux: policy capability open_perms=1 May 7 23:46:43.390260 kernel: SELinux: policy capability extended_socket_class=1 May 7 23:46:43.390289 kernel: SELinux: policy capability always_check_network=0 May 7 23:46:43.390317 kernel: SELinux: policy capability cgroup_seclabel=1 May 7 23:46:43.390347 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 7 23:46:43.390376 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 7 23:46:43.390405 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 7 23:46:43.390441 kernel: audit: type=1403 audit(1746661601.718:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 7 23:46:43.390480 systemd[1]: Successfully loaded SELinux policy in 49.368ms. May 7 23:46:43.390532 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 24.708ms. May 7 23:46:43.390564 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 7 23:46:43.390594 systemd[1]: Detected virtualization amazon. May 7 23:46:43.390625 systemd[1]: Detected architecture arm64. May 7 23:46:43.390657 systemd[1]: Detected first boot. May 7 23:46:43.390688 systemd[1]: Initializing machine ID from VM UUID. May 7 23:46:43.390718 zram_generator::config[1407]: No configuration found. May 7 23:46:43.390765 kernel: NET: Registered PF_VSOCK protocol family May 7 23:46:43.390797 systemd[1]: Populated /etc with preset unit settings. May 7 23:46:43.390829 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 7 23:46:43.390860 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 7 23:46:43.390894 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 7 23:46:43.390932 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 7 23:46:43.390961 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 7 23:46:43.390994 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 7 23:46:43.391027 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 7 23:46:43.391058 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 7 23:46:43.391092 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 7 23:46:43.393491 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 7 23:46:43.393542 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 7 23:46:43.393576 systemd[1]: Created slice user.slice - User and Session Slice. May 7 23:46:43.393613 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 7 23:46:43.393648 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 7 23:46:43.393678 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 7 23:46:43.393722 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 7 23:46:43.393753 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 7 23:46:43.393790 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 7 23:46:43.393825 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 7 23:46:43.393861 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 7 23:46:43.393896 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 7 23:46:43.393927 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 7 23:46:43.393974 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 7 23:46:43.394006 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 7 23:46:43.394065 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 7 23:46:43.394102 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 7 23:46:43.394160 systemd[1]: Reached target slices.target - Slice Units. May 7 23:46:43.394194 systemd[1]: Reached target swap.target - Swaps. May 7 23:46:43.394225 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 7 23:46:43.394256 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 7 23:46:43.394288 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 7 23:46:43.394325 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 7 23:46:43.394357 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 7 23:46:43.394387 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 7 23:46:43.394416 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 7 23:46:43.394444 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 7 23:46:43.394475 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 7 23:46:43.394506 systemd[1]: Mounting media.mount - External Media Directory... May 7 23:46:43.394613 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 7 23:46:43.395099 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 7 23:46:43.400261 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 7 23:46:43.400298 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 7 23:46:43.400330 systemd[1]: Reached target machines.target - Containers. May 7 23:46:43.400362 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 7 23:46:43.400393 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 7 23:46:43.400423 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 7 23:46:43.400452 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 7 23:46:43.400480 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 7 23:46:43.400511 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 7 23:46:43.400545 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 7 23:46:43.400579 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 7 23:46:43.400610 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 7 23:46:43.400638 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 7 23:46:43.400667 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 7 23:46:43.400699 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 7 23:46:43.400730 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 7 23:46:43.400758 systemd[1]: Stopped systemd-fsck-usr.service. May 7 23:46:43.400794 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 7 23:46:43.400824 systemd[1]: Starting systemd-journald.service - Journal Service... May 7 23:46:43.400852 kernel: loop: module loaded May 7 23:46:43.400880 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 7 23:46:43.400909 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 7 23:46:43.400937 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 7 23:46:43.400970 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 7 23:46:43.400997 kernel: fuse: init (API version 7.39) May 7 23:46:43.401025 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 7 23:46:43.401059 systemd[1]: verity-setup.service: Deactivated successfully. May 7 23:46:43.401087 systemd[1]: Stopped verity-setup.service. May 7 23:46:43.401135 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 7 23:46:43.401171 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 7 23:46:43.401205 systemd[1]: Mounted media.mount - External Media Directory. May 7 23:46:43.401233 kernel: ACPI: bus type drm_connector registered May 7 23:46:43.401261 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 7 23:46:43.401289 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 7 23:46:43.401317 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 7 23:46:43.401351 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 7 23:46:43.401384 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 7 23:46:43.401424 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 7 23:46:43.401453 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 7 23:46:43.401481 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 7 23:46:43.401509 systemd[1]: modprobe@drm.service: Deactivated successfully. May 7 23:46:43.401538 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 7 23:46:43.401566 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 7 23:46:43.401641 systemd-journald[1497]: Collecting audit messages is disabled. May 7 23:46:43.401697 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 7 23:46:43.401729 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 7 23:46:43.401758 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 7 23:46:43.401788 systemd[1]: modprobe@loop.service: Deactivated successfully. May 7 23:46:43.401817 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 7 23:46:43.401846 systemd-journald[1497]: Journal started May 7 23:46:43.401895 systemd-journald[1497]: Runtime Journal (/run/log/journal/ec26214e429a6f04acfd1063e3a1b4ba) is 8M, max 75.3M, 67.3M free. May 7 23:46:42.836675 systemd[1]: Queued start job for default target multi-user.target. May 7 23:46:42.847444 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. May 7 23:46:42.848304 systemd[1]: systemd-journald.service: Deactivated successfully. May 7 23:46:43.415241 systemd[1]: Started systemd-journald.service - Journal Service. May 7 23:46:43.413781 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 7 23:46:43.419338 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 7 23:46:43.429580 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 7 23:46:43.456175 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 7 23:46:43.464703 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 7 23:46:43.467828 systemd[1]: Reached target network-pre.target - Preparation for Network. May 7 23:46:43.478234 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 7 23:46:43.489002 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 7 23:46:43.492321 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 7 23:46:43.492381 systemd[1]: Reached target local-fs.target - Local File Systems. May 7 23:46:43.499328 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 7 23:46:43.509470 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 7 23:46:43.520544 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 7 23:46:43.522872 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 7 23:46:43.527997 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 7 23:46:43.533495 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 7 23:46:43.535810 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 7 23:46:43.539472 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 7 23:46:43.541811 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 7 23:46:43.553595 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 7 23:46:43.561419 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 7 23:46:43.566402 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 7 23:46:43.575697 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 7 23:46:43.578339 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 7 23:46:43.581221 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 7 23:46:43.598491 systemd-journald[1497]: Time spent on flushing to /var/log/journal/ec26214e429a6f04acfd1063e3a1b4ba is 84.181ms for 923 entries. May 7 23:46:43.598491 systemd-journald[1497]: System Journal (/var/log/journal/ec26214e429a6f04acfd1063e3a1b4ba) is 8M, max 195.6M, 187.6M free. May 7 23:46:43.708901 systemd-journald[1497]: Received client request to flush runtime journal. May 7 23:46:43.708979 kernel: loop0: detected capacity change from 0 to 53784 May 7 23:46:43.680338 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 7 23:46:43.683598 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 7 23:46:43.709494 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 7 23:46:43.719482 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 7 23:46:43.723920 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 7 23:46:43.761035 systemd-tmpfiles[1543]: ACLs are not supported, ignoring. May 7 23:46:43.761068 systemd-tmpfiles[1543]: ACLs are not supported, ignoring. May 7 23:46:43.781424 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 7 23:46:43.793162 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 7 23:46:43.799083 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 7 23:46:43.803203 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 7 23:46:43.816437 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 7 23:46:43.826215 kernel: loop1: detected capacity change from 0 to 123192 May 7 23:46:43.828483 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... May 7 23:46:43.852236 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 7 23:46:43.893173 udevadm[1562]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. May 7 23:46:43.897173 kernel: loop2: detected capacity change from 0 to 201592 May 7 23:46:43.914238 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 7 23:46:43.924934 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 7 23:46:43.983035 kernel: loop3: detected capacity change from 0 to 113512 May 7 23:46:43.999947 systemd-tmpfiles[1566]: ACLs are not supported, ignoring. May 7 23:46:43.999986 systemd-tmpfiles[1566]: ACLs are not supported, ignoring. May 7 23:46:44.020934 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 7 23:46:44.070174 kernel: loop4: detected capacity change from 0 to 53784 May 7 23:46:44.091154 kernel: loop5: detected capacity change from 0 to 123192 May 7 23:46:44.126176 kernel: loop6: detected capacity change from 0 to 201592 May 7 23:46:44.183153 kernel: loop7: detected capacity change from 0 to 113512 May 7 23:46:44.223062 (sd-merge)[1571]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. May 7 23:46:44.224154 (sd-merge)[1571]: Merged extensions into '/usr'. May 7 23:46:44.233485 systemd[1]: Reload requested from client PID 1542 ('systemd-sysext') (unit systemd-sysext.service)... May 7 23:46:44.233519 systemd[1]: Reloading... May 7 23:46:44.475140 zram_generator::config[1602]: No configuration found. May 7 23:46:44.548710 ldconfig[1537]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 7 23:46:44.776472 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 7 23:46:44.932055 systemd[1]: Reloading finished in 697 ms. May 7 23:46:44.951854 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 7 23:46:44.954784 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 7 23:46:44.957736 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 7 23:46:44.973338 systemd[1]: Starting ensure-sysext.service... May 7 23:46:44.977498 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 7 23:46:44.990471 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 7 23:46:45.016351 systemd[1]: Reload requested from client PID 1652 ('systemctl') (unit ensure-sysext.service)... May 7 23:46:45.016552 systemd[1]: Reloading... May 7 23:46:45.070542 systemd-udevd[1654]: Using default interface naming scheme 'v255'. May 7 23:46:45.074782 systemd-tmpfiles[1653]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 7 23:46:45.075354 systemd-tmpfiles[1653]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 7 23:46:45.077822 systemd-tmpfiles[1653]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 7 23:46:45.078693 systemd-tmpfiles[1653]: ACLs are not supported, ignoring. May 7 23:46:45.078950 systemd-tmpfiles[1653]: ACLs are not supported, ignoring. May 7 23:46:45.088059 systemd-tmpfiles[1653]: Detected autofs mount point /boot during canonicalization of boot. May 7 23:46:45.088303 systemd-tmpfiles[1653]: Skipping /boot May 7 23:46:45.125217 systemd-tmpfiles[1653]: Detected autofs mount point /boot during canonicalization of boot. May 7 23:46:45.125239 systemd-tmpfiles[1653]: Skipping /boot May 7 23:46:45.245965 zram_generator::config[1689]: No configuration found. May 7 23:46:45.405315 (udev-worker)[1695]: Network interface NamePolicy= disabled on kernel command line. May 7 23:46:45.591381 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 7 23:46:45.717137 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 42 scanned by (udev-worker) (1690) May 7 23:46:45.805502 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 7 23:46:45.805896 systemd[1]: Reloading finished in 788 ms. May 7 23:46:45.831450 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 7 23:46:45.868840 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 7 23:46:45.924338 systemd[1]: Finished ensure-sysext.service. May 7 23:46:46.000876 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. May 7 23:46:46.004013 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. May 7 23:46:46.019414 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 7 23:46:46.028437 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 7 23:46:46.031636 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 7 23:46:46.039425 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... May 7 23:46:46.046184 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 7 23:46:46.051472 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 7 23:46:46.057491 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 7 23:46:46.063554 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 7 23:46:46.066639 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 7 23:46:46.073526 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 7 23:46:46.078355 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 7 23:46:46.085474 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 7 23:46:46.096232 lvm[1854]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 7 23:46:46.112261 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 7 23:46:46.120667 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 7 23:46:46.123374 systemd[1]: Reached target time-set.target - System Time Set. May 7 23:46:46.135364 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 7 23:46:46.142719 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 7 23:46:46.152291 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 7 23:46:46.174644 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 7 23:46:46.175188 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 7 23:46:46.177686 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 7 23:46:46.178955 systemd[1]: modprobe@drm.service: Deactivated successfully. May 7 23:46:46.179754 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 7 23:46:46.210320 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 7 23:46:46.212920 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 7 23:46:46.220244 systemd[1]: modprobe@loop.service: Deactivated successfully. May 7 23:46:46.220700 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 7 23:46:46.224516 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 7 23:46:46.232486 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 7 23:46:46.236244 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. May 7 23:46:46.245253 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 7 23:46:46.253486 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... May 7 23:46:46.256980 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 7 23:46:46.285742 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 7 23:46:46.298577 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 7 23:46:46.307391 lvm[1886]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 7 23:46:46.347535 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 7 23:46:46.350610 augenrules[1897]: No rules May 7 23:46:46.354880 systemd[1]: audit-rules.service: Deactivated successfully. May 7 23:46:46.355927 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 7 23:46:46.387835 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. May 7 23:46:46.400850 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 7 23:46:46.406963 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 7 23:46:46.411456 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 7 23:46:46.435849 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 7 23:46:46.545033 systemd-networkd[1866]: lo: Link UP May 7 23:46:46.545060 systemd-networkd[1866]: lo: Gained carrier May 7 23:46:46.547283 systemd-resolved[1867]: Positive Trust Anchors: May 7 23:46:46.547313 systemd-resolved[1867]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 7 23:46:46.547378 systemd-resolved[1867]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 7 23:46:46.548157 systemd-networkd[1866]: Enumeration completed May 7 23:46:46.548323 systemd[1]: Started systemd-networkd.service - Network Configuration. May 7 23:46:46.554155 systemd-networkd[1866]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 7 23:46:46.554175 systemd-networkd[1866]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 7 23:46:46.556777 systemd-networkd[1866]: eth0: Link UP May 7 23:46:46.557146 systemd-networkd[1866]: eth0: Gained carrier May 7 23:46:46.557199 systemd-networkd[1866]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 7 23:46:46.561416 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 7 23:46:46.564635 systemd-resolved[1867]: Defaulting to hostname 'linux'. May 7 23:46:46.567444 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 7 23:46:46.572850 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 7 23:46:46.575162 systemd[1]: Reached target network.target - Network. May 7 23:46:46.577349 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 7 23:46:46.579658 systemd[1]: Reached target sysinit.target - System Initialization. May 7 23:46:46.581825 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 7 23:46:46.584257 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 7 23:46:46.586896 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 7 23:46:46.588311 systemd-networkd[1866]: eth0: DHCPv4 address 172.31.28.15/20, gateway 172.31.16.1 acquired from 172.31.16.1 May 7 23:46:46.589142 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 7 23:46:46.591464 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 7 23:46:46.594242 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 7 23:46:46.594292 systemd[1]: Reached target paths.target - Path Units. May 7 23:46:46.596090 systemd[1]: Reached target timers.target - Timer Units. May 7 23:46:46.600046 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 7 23:46:46.607493 systemd[1]: Starting docker.socket - Docker Socket for the API... May 7 23:46:46.614941 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 7 23:46:46.620584 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 7 23:46:46.623081 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 7 23:46:46.634379 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 7 23:46:46.637044 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 7 23:46:46.641170 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 7 23:46:46.644512 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 7 23:46:46.647637 systemd[1]: Reached target sockets.target - Socket Units. May 7 23:46:46.649907 systemd[1]: Reached target basic.target - Basic System. May 7 23:46:46.652479 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 7 23:46:46.652537 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 7 23:46:46.661234 systemd[1]: Starting containerd.service - containerd container runtime... May 7 23:46:46.668910 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 7 23:46:46.675455 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 7 23:46:46.679518 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 7 23:46:46.685852 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 7 23:46:46.688530 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 7 23:46:46.700505 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 7 23:46:46.721422 systemd[1]: Started ntpd.service - Network Time Service. May 7 23:46:46.728327 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 7 23:46:46.741031 systemd[1]: Starting setup-oem.service - Setup OEM... May 7 23:46:46.756393 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 7 23:46:46.781625 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 7 23:46:46.786756 jq[1925]: false May 7 23:46:46.795046 systemd[1]: Starting systemd-logind.service - User Login Management... May 7 23:46:46.803840 extend-filesystems[1926]: Found loop4 May 7 23:46:46.803840 extend-filesystems[1926]: Found loop5 May 7 23:46:46.803840 extend-filesystems[1926]: Found loop6 May 7 23:46:46.803840 extend-filesystems[1926]: Found loop7 May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1 May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1p1 May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1p2 May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1p3 May 7 23:46:46.803840 extend-filesystems[1926]: Found usr May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1p4 May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1p6 May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1p7 May 7 23:46:46.803840 extend-filesystems[1926]: Found nvme0n1p9 May 7 23:46:46.803840 extend-filesystems[1926]: Checking size of /dev/nvme0n1p9 May 7 23:46:46.802820 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 7 23:46:46.810789 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 7 23:46:46.815433 systemd[1]: Starting update-engine.service - Update Engine... May 7 23:46:46.828356 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 7 23:46:46.855451 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 7 23:46:46.855931 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: ntpd 4.2.8p17@1.4004-o Wed May 7 21:39:07 UTC 2025 (1): Starting May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: ---------------------------------------------------- May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: ntp-4 is maintained by Network Time Foundation, May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Inc. (NTF), a non-profit 501(c)(3) public-benefit May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: corporation. Support and training for ntp-4 are May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: available at https://www.nwtime.org/support May 7 23:46:46.885315 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: ---------------------------------------------------- May 7 23:46:46.884332 ntpd[1928]: ntpd 4.2.8p17@1.4004-o Wed May 7 21:39:07 UTC 2025 (1): Starting May 7 23:46:46.884379 ntpd[1928]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp May 7 23:46:46.884399 ntpd[1928]: ---------------------------------------------------- May 7 23:46:46.884417 ntpd[1928]: ntp-4 is maintained by Network Time Foundation, May 7 23:46:46.884435 ntpd[1928]: Inc. (NTF), a non-profit 501(c)(3) public-benefit May 7 23:46:46.884452 ntpd[1928]: corporation. Support and training for ntp-4 are May 7 23:46:46.884470 ntpd[1928]: available at https://www.nwtime.org/support May 7 23:46:46.884487 ntpd[1928]: ---------------------------------------------------- May 7 23:46:46.887558 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 7 23:46:46.888072 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 7 23:46:46.900459 ntpd[1928]: proto: precision = 0.096 usec (-23) May 7 23:46:46.907245 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: proto: precision = 0.096 usec (-23) May 7 23:46:46.913635 extend-filesystems[1926]: Resized partition /dev/nvme0n1p9 May 7 23:46:46.917528 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: basedate set to 2025-04-25 May 7 23:46:46.917528 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: gps base set to 2025-04-27 (week 2364) May 7 23:46:46.916362 ntpd[1928]: basedate set to 2025-04-25 May 7 23:46:46.916396 ntpd[1928]: gps base set to 2025-04-27 (week 2364) May 7 23:46:46.927156 extend-filesystems[1961]: resize2fs 1.47.1 (20-May-2024) May 7 23:46:46.932846 ntpd[1928]: Listen and drop on 0 v6wildcard [::]:123 May 7 23:46:46.933024 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Listen and drop on 0 v6wildcard [::]:123 May 7 23:46:46.940896 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks May 7 23:46:46.935773 ntpd[1928]: Listen and drop on 1 v4wildcard 0.0.0.0:123 May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Listen and drop on 1 v4wildcard 0.0.0.0:123 May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Listen normally on 2 lo 127.0.0.1:123 May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Listen normally on 3 eth0 172.31.28.15:123 May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Listen normally on 4 lo [::1]:123 May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: bind(21) AF_INET6 fe80::469:2ff:fec4:41e9%2#123 flags 0x11 failed: Cannot assign requested address May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: unable to create socket on eth0 (5) for fe80::469:2ff:fec4:41e9%2#123 May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: failed to init interface for address fe80::469:2ff:fec4:41e9%2 May 7 23:46:46.941180 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: Listening on routing socket on fd #21 for interface updates May 7 23:46:46.937332 ntpd[1928]: Listen normally on 2 lo 127.0.0.1:123 May 7 23:46:46.937399 ntpd[1928]: Listen normally on 3 eth0 172.31.28.15:123 May 7 23:46:46.937463 ntpd[1928]: Listen normally on 4 lo [::1]:123 May 7 23:46:46.937541 ntpd[1928]: bind(21) AF_INET6 fe80::469:2ff:fec4:41e9%2#123 flags 0x11 failed: Cannot assign requested address May 7 23:46:46.937580 ntpd[1928]: unable to create socket on eth0 (5) for fe80::469:2ff:fec4:41e9%2#123 May 7 23:46:46.937607 ntpd[1928]: failed to init interface for address fe80::469:2ff:fec4:41e9%2 May 7 23:46:46.937661 ntpd[1928]: Listening on routing socket on fd #21 for interface updates May 7 23:46:46.954161 jq[1941]: true May 7 23:46:46.978409 ntpd[1928]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 7 23:46:46.979076 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 7 23:46:46.979076 ntpd[1928]: 7 May 23:46:46 ntpd[1928]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 7 23:46:46.978462 ntpd[1928]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized May 7 23:46:46.987231 systemd[1]: motdgen.service: Deactivated successfully. May 7 23:46:46.987737 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 7 23:46:46.992921 dbus-daemon[1924]: [system] SELinux support is enabled May 7 23:46:46.994264 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 7 23:46:47.001451 (ntainerd)[1963]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 7 23:46:47.001650 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 7 23:46:47.001699 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 7 23:46:47.004193 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 7 23:46:47.004231 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 7 23:46:47.042313 dbus-daemon[1924]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.1' (uid=244 pid=1866 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") May 7 23:46:47.037894 systemd[1]: Finished setup-oem.service - Setup OEM. May 7 23:46:47.049174 tar[1958]: linux-arm64/LICENSE May 7 23:46:47.049174 tar[1958]: linux-arm64/helm May 7 23:46:47.049400 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... May 7 23:46:47.068944 update_engine[1939]: I20250507 23:46:47.062937 1939 main.cc:92] Flatcar Update Engine starting May 7 23:46:47.074022 systemd[1]: Started update-engine.service - Update Engine. May 7 23:46:47.077478 update_engine[1939]: I20250507 23:46:47.077404 1939 update_check_scheduler.cc:74] Next update check in 3m25s May 7 23:46:47.079440 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 7 23:46:47.083366 jq[1970]: true May 7 23:46:47.098268 coreos-metadata[1923]: May 07 23:46:47.096 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 May 7 23:46:47.103575 coreos-metadata[1923]: May 07 23:46:47.102 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 May 7 23:46:47.107044 coreos-metadata[1923]: May 07 23:46:47.105 INFO Fetch successful May 7 23:46:47.107044 coreos-metadata[1923]: May 07 23:46:47.105 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 May 7 23:46:47.107044 coreos-metadata[1923]: May 07 23:46:47.105 INFO Fetch successful May 7 23:46:47.107044 coreos-metadata[1923]: May 07 23:46:47.105 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 May 7 23:46:47.114442 coreos-metadata[1923]: May 07 23:46:47.112 INFO Fetch successful May 7 23:46:47.114442 coreos-metadata[1923]: May 07 23:46:47.112 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 May 7 23:46:47.114442 coreos-metadata[1923]: May 07 23:46:47.114 INFO Fetch successful May 7 23:46:47.114442 coreos-metadata[1923]: May 07 23:46:47.114 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 May 7 23:46:47.116144 coreos-metadata[1923]: May 07 23:46:47.115 INFO Fetch failed with 404: resource not found May 7 23:46:47.116144 coreos-metadata[1923]: May 07 23:46:47.115 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 May 7 23:46:47.119440 coreos-metadata[1923]: May 07 23:46:47.119 INFO Fetch successful May 7 23:46:47.119440 coreos-metadata[1923]: May 07 23:46:47.119 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 May 7 23:46:47.119440 coreos-metadata[1923]: May 07 23:46:47.119 INFO Fetch successful May 7 23:46:47.119440 coreos-metadata[1923]: May 07 23:46:47.119 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 May 7 23:46:47.120532 coreos-metadata[1923]: May 07 23:46:47.119 INFO Fetch successful May 7 23:46:47.120532 coreos-metadata[1923]: May 07 23:46:47.119 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 May 7 23:46:47.126042 coreos-metadata[1923]: May 07 23:46:47.124 INFO Fetch successful May 7 23:46:47.126042 coreos-metadata[1923]: May 07 23:46:47.124 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 May 7 23:46:47.130737 coreos-metadata[1923]: May 07 23:46:47.127 INFO Fetch successful May 7 23:46:47.132721 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 May 7 23:46:47.152682 extend-filesystems[1961]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required May 7 23:46:47.152682 extend-filesystems[1961]: old_desc_blocks = 1, new_desc_blocks = 1 May 7 23:46:47.152682 extend-filesystems[1961]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. May 7 23:46:47.176007 extend-filesystems[1926]: Resized filesystem in /dev/nvme0n1p9 May 7 23:46:47.161195 systemd[1]: extend-filesystems.service: Deactivated successfully. May 7 23:46:47.162723 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 7 23:46:47.241589 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 7 23:46:47.244706 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 7 23:46:47.275245 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 42 scanned by (udev-worker) (1698) May 7 23:46:47.357161 bash[2023]: Updated "/home/core/.ssh/authorized_keys" May 7 23:46:47.362993 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 7 23:46:47.386802 systemd[1]: Starting sshkeys.service... May 7 23:46:47.407704 systemd-logind[1936]: Watching system buttons on /dev/input/event0 (Power Button) May 7 23:46:47.407757 systemd-logind[1936]: Watching system buttons on /dev/input/event1 (Sleep Button) May 7 23:46:47.408167 systemd-logind[1936]: New seat seat0. May 7 23:46:47.415332 systemd[1]: Started systemd-logind.service - User Login Management. May 7 23:46:47.447373 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. May 7 23:46:47.568564 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... May 7 23:46:47.575646 systemd[1]: Started systemd-hostnamed.service - Hostname Service. May 7 23:46:47.592190 dbus-daemon[1924]: [system] Successfully activated service 'org.freedesktop.hostname1' May 7 23:46:47.598227 dbus-daemon[1924]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=1976 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") May 7 23:46:47.610695 systemd[1]: Starting polkit.service - Authorization Manager... May 7 23:46:47.709258 polkitd[2052]: Started polkitd version 121 May 7 23:46:47.736538 polkitd[2052]: Loading rules from directory /etc/polkit-1/rules.d May 7 23:46:47.736665 polkitd[2052]: Loading rules from directory /usr/share/polkit-1/rules.d May 7 23:46:47.738274 polkitd[2052]: Finished loading, compiling and executing 2 rules May 7 23:46:47.745329 dbus-daemon[1924]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' May 7 23:46:47.745619 systemd[1]: Started polkit.service - Authorization Manager. May 7 23:46:47.749962 polkitd[2052]: Acquired the name org.freedesktop.PolicyKit1 on the system bus May 7 23:46:47.774300 systemd-networkd[1866]: eth0: Gained IPv6LL May 7 23:46:47.784978 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 7 23:46:47.790377 systemd[1]: Reached target network-online.target - Network is Online. May 7 23:46:47.864493 containerd[1963]: time="2025-05-07T23:46:47.864344495Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 May 7 23:46:47.871283 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. May 7 23:46:47.882808 coreos-metadata[2036]: May 07 23:46:47.872 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 May 7 23:46:47.882808 coreos-metadata[2036]: May 07 23:46:47.878 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 May 7 23:46:47.880075 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:46:47.893865 coreos-metadata[2036]: May 07 23:46:47.885 INFO Fetch successful May 7 23:46:47.893865 coreos-metadata[2036]: May 07 23:46:47.885 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 May 7 23:46:47.893865 coreos-metadata[2036]: May 07 23:46:47.892 INFO Fetch successful May 7 23:46:47.886459 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 7 23:46:47.899578 unknown[2036]: wrote ssh authorized keys file for user: core May 7 23:46:47.911512 systemd-hostnamed[1976]: Hostname set to (transient) May 7 23:46:47.912156 systemd-resolved[1867]: System hostname changed to 'ip-172-31-28-15'. May 7 23:46:48.092140 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 7 23:46:48.102730 amazon-ssm-agent[2081]: Initializing new seelog logger May 7 23:46:48.102730 amazon-ssm-agent[2081]: New Seelog Logger Creation Complete May 7 23:46:48.102730 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.102730 amazon-ssm-agent[2081]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.102730 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 processing appconfig overrides May 7 23:46:48.112363 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.112363 amazon-ssm-agent[2081]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.112363 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 processing appconfig overrides May 7 23:46:48.112363 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.112363 amazon-ssm-agent[2081]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.112363 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 processing appconfig overrides May 7 23:46:48.114250 update-ssh-keys[2117]: Updated "/home/core/.ssh/authorized_keys" May 7 23:46:48.114717 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO Proxy environment variables: May 7 23:46:48.118476 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). May 7 23:46:48.128469 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.128469 amazon-ssm-agent[2081]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. May 7 23:46:48.128469 amazon-ssm-agent[2081]: 2025/05/07 23:46:48 processing appconfig overrides May 7 23:46:48.130220 systemd[1]: Finished sshkeys.service. May 7 23:46:48.138158 containerd[1963]: time="2025-05-07T23:46:48.131898212Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 May 7 23:46:48.149900 containerd[1963]: time="2025-05-07T23:46:48.149787896Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.88-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 May 7 23:46:48.150024 containerd[1963]: time="2025-05-07T23:46:48.149884400Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 May 7 23:46:48.150024 containerd[1963]: time="2025-05-07T23:46:48.149946860Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 May 7 23:46:48.153778 containerd[1963]: time="2025-05-07T23:46:48.153702008Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 May 7 23:46:48.153918 containerd[1963]: time="2025-05-07T23:46:48.153798008Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 May 7 23:46:48.156140 containerd[1963]: time="2025-05-07T23:46:48.155248496Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 May 7 23:46:48.156140 containerd[1963]: time="2025-05-07T23:46:48.155295680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 May 7 23:46:48.157010 locksmithd[1979]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 7 23:46:48.161502 containerd[1963]: time="2025-05-07T23:46:48.159811088Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 7 23:46:48.161502 containerd[1963]: time="2025-05-07T23:46:48.159874604Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 May 7 23:46:48.161502 containerd[1963]: time="2025-05-07T23:46:48.159911888Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 May 7 23:46:48.161502 containerd[1963]: time="2025-05-07T23:46:48.159937472Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 May 7 23:46:48.166134 containerd[1963]: time="2025-05-07T23:46:48.164342588Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 May 7 23:46:48.168672 containerd[1963]: time="2025-05-07T23:46:48.168602804Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 May 7 23:46:48.169369 containerd[1963]: time="2025-05-07T23:46:48.169285820Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 7 23:46:48.169369 containerd[1963]: time="2025-05-07T23:46:48.169359764Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 May 7 23:46:48.172561 containerd[1963]: time="2025-05-07T23:46:48.172491680Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 May 7 23:46:48.172944 containerd[1963]: time="2025-05-07T23:46:48.172900616Z" level=info msg="metadata content store policy set" policy=shared May 7 23:46:48.186237 containerd[1963]: time="2025-05-07T23:46:48.186157820Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 May 7 23:46:48.186343 containerd[1963]: time="2025-05-07T23:46:48.186272336Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 May 7 23:46:48.186392 containerd[1963]: time="2025-05-07T23:46:48.186338720Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 May 7 23:46:48.186438 containerd[1963]: time="2025-05-07T23:46:48.186408200Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 May 7 23:46:48.186483 containerd[1963]: time="2025-05-07T23:46:48.186444440Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 May 7 23:46:48.186749 containerd[1963]: time="2025-05-07T23:46:48.186701804Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 May 7 23:46:48.189138 containerd[1963]: time="2025-05-07T23:46:48.187267772Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 May 7 23:46:48.195401 containerd[1963]: time="2025-05-07T23:46:48.189472772Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 May 7 23:46:48.195565 containerd[1963]: time="2025-05-07T23:46:48.195523316Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 May 7 23:46:48.195621 containerd[1963]: time="2025-05-07T23:46:48.195574712Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 May 7 23:46:48.195669 containerd[1963]: time="2025-05-07T23:46:48.195637532Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 May 7 23:46:48.195715 containerd[1963]: time="2025-05-07T23:46:48.195693176Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 May 7 23:46:48.195786 containerd[1963]: time="2025-05-07T23:46:48.195727688Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 May 7 23:46:48.195834 containerd[1963]: time="2025-05-07T23:46:48.195783992Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 May 7 23:46:48.195834 containerd[1963]: time="2025-05-07T23:46:48.195819620Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.195967760Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196010852Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196066592Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196199648Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196237844Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196382432Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196414688Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196475084Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196530968Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196563236Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196593524Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.196651268Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.198139125Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198273 containerd[1963]: time="2025-05-07T23:46:48.198186717Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 May 7 23:46:48.198921 containerd[1963]: time="2025-05-07T23:46:48.198245277Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 May 7 23:46:48.201523 containerd[1963]: time="2025-05-07T23:46:48.198276297Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 May 7 23:46:48.201523 containerd[1963]: time="2025-05-07T23:46:48.199194057Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 May 7 23:46:48.201523 containerd[1963]: time="2025-05-07T23:46:48.199276977Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 May 7 23:46:48.201523 containerd[1963]: time="2025-05-07T23:46:48.200164185Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 May 7 23:46:48.203990 containerd[1963]: time="2025-05-07T23:46:48.200203269Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 May 7 23:46:48.204506 containerd[1963]: time="2025-05-07T23:46:48.204446205Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 May 7 23:46:48.204586 containerd[1963]: time="2025-05-07T23:46:48.204541617Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 May 7 23:46:48.204586 containerd[1963]: time="2025-05-07T23:46:48.204574677Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 May 7 23:46:48.204783 containerd[1963]: time="2025-05-07T23:46:48.204724845Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 May 7 23:46:48.204868 containerd[1963]: time="2025-05-07T23:46:48.204765237Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 May 7 23:46:48.204868 containerd[1963]: time="2025-05-07T23:46:48.204824925Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 May 7 23:46:48.208353 containerd[1963]: time="2025-05-07T23:46:48.204850941Z" level=info msg="NRI interface is disabled by configuration." May 7 23:46:48.208490 containerd[1963]: time="2025-05-07T23:46:48.208369725Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 May 7 23:46:48.210663 containerd[1963]: time="2025-05-07T23:46:48.209170665Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" May 7 23:46:48.210663 containerd[1963]: time="2025-05-07T23:46:48.209933937Z" level=info msg="Connect containerd service" May 7 23:46:48.210663 containerd[1963]: time="2025-05-07T23:46:48.210173841Z" level=info msg="using legacy CRI server" May 7 23:46:48.210663 containerd[1963]: time="2025-05-07T23:46:48.210197937Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 7 23:46:48.211090 containerd[1963]: time="2025-05-07T23:46:48.210664269Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.218456505Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.221242857Z" level=info msg="Start subscribing containerd event" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.221326341Z" level=info msg="Start recovering state" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.221447397Z" level=info msg="Start event monitor" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.221470497Z" level=info msg="Start snapshots syncer" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.221497761Z" level=info msg="Start cni network conf syncer for default" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.221519409Z" level=info msg="Start streaming server" May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.225224289Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.225341589Z" level=info msg=serving... address=/run/containerd/containerd.sock May 7 23:46:48.225921 containerd[1963]: time="2025-05-07T23:46:48.225557733Z" level=info msg="containerd successfully booted in 0.369706s" May 7 23:46:48.235632 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO https_proxy: May 7 23:46:48.228852 systemd[1]: Started containerd.service - containerd container runtime. May 7 23:46:48.332161 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO http_proxy: May 7 23:46:48.431368 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO no_proxy: May 7 23:46:48.531017 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO Checking if agent identity type OnPrem can be assumed May 7 23:46:48.629803 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO Checking if agent identity type EC2 can be assumed May 7 23:46:48.701562 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 7 23:46:48.728774 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO Agent will take identity from EC2 May 7 23:46:48.835132 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [amazon-ssm-agent] using named pipe channel for IPC May 7 23:46:48.931985 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [amazon-ssm-agent] using named pipe channel for IPC May 7 23:46:49.032452 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [amazon-ssm-agent] using named pipe channel for IPC May 7 23:46:49.075227 tar[1958]: linux-arm64/README.md May 7 23:46:49.113745 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 7 23:46:49.131890 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [amazon-ssm-agent] OS: linux, Arch: arm64 May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [amazon-ssm-agent] Starting Core Agent May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [amazon-ssm-agent] registrar detected. Attempting registration May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [Registrar] Starting registrar module May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:48 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:49 INFO [EC2Identity] EC2 registration was successful. May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:49 INFO [CredentialRefresher] credentialRefresher has started May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:49 INFO [CredentialRefresher] Starting credentials refresher loop May 7 23:46:49.222964 amazon-ssm-agent[2081]: 2025-05-07 23:46:49 INFO EC2RoleProvider Successfully connected with instance profile role credentials May 7 23:46:49.232966 amazon-ssm-agent[2081]: 2025-05-07 23:46:49 INFO [CredentialRefresher] Next credential rotation will be in 30.616657088266667 minutes May 7 23:46:49.325963 sshd_keygen[1966]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 7 23:46:49.369285 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 7 23:46:49.379653 systemd[1]: Starting issuegen.service - Generate /run/issue... May 7 23:46:49.386208 systemd[1]: Started sshd@0-172.31.28.15:22-147.75.109.163:44530.service - OpenSSH per-connection server daemon (147.75.109.163:44530). May 7 23:46:49.420838 systemd[1]: issuegen.service: Deactivated successfully. May 7 23:46:49.421552 systemd[1]: Finished issuegen.service - Generate /run/issue. May 7 23:46:49.435312 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 7 23:46:49.475530 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 7 23:46:49.490781 systemd[1]: Started getty@tty1.service - Getty on tty1. May 7 23:46:49.501416 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 7 23:46:49.504679 systemd[1]: Reached target getty.target - Login Prompts. May 7 23:46:49.637163 sshd[2163]: Accepted publickey for core from 147.75.109.163 port 44530 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:46:49.638734 sshd-session[2163]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:46:49.660872 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 7 23:46:49.671728 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 7 23:46:49.681692 systemd-logind[1936]: New session 1 of user core. May 7 23:46:49.710637 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 7 23:46:49.723745 systemd[1]: Starting user@500.service - User Manager for UID 500... May 7 23:46:49.747515 (systemd)[2174]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 7 23:46:49.753352 systemd-logind[1936]: New session c1 of user core. May 7 23:46:49.885067 ntpd[1928]: Listen normally on 6 eth0 [fe80::469:2ff:fec4:41e9%2]:123 May 7 23:46:49.886367 ntpd[1928]: 7 May 23:46:49 ntpd[1928]: Listen normally on 6 eth0 [fe80::469:2ff:fec4:41e9%2]:123 May 7 23:46:49.925496 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:46:49.931818 systemd[1]: Reached target multi-user.target - Multi-User System. May 7 23:46:49.940912 (kubelet)[2185]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 7 23:46:50.067580 systemd[2174]: Queued start job for default target default.target. May 7 23:46:50.076199 systemd[2174]: Created slice app.slice - User Application Slice. May 7 23:46:50.076267 systemd[2174]: Reached target paths.target - Paths. May 7 23:46:50.076387 systemd[2174]: Reached target timers.target - Timers. May 7 23:46:50.080348 systemd[2174]: Starting dbus.socket - D-Bus User Message Bus Socket... May 7 23:46:50.118558 systemd[2174]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 7 23:46:50.120011 systemd[2174]: Reached target sockets.target - Sockets. May 7 23:46:50.120216 systemd[2174]: Reached target basic.target - Basic System. May 7 23:46:50.120322 systemd[2174]: Reached target default.target - Main User Target. May 7 23:46:50.120381 systemd[2174]: Startup finished in 351ms. May 7 23:46:50.120401 systemd[1]: Started user@500.service - User Manager for UID 500. May 7 23:46:50.133420 systemd[1]: Started session-1.scope - Session 1 of User core. May 7 23:46:50.135946 systemd[1]: Startup finished in 1.095s (kernel) + 7.936s (initrd) + 8.465s (userspace) = 17.496s. May 7 23:46:50.274988 amazon-ssm-agent[2081]: 2025-05-07 23:46:50 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process May 7 23:46:50.318312 systemd[1]: Started sshd@1-172.31.28.15:22-147.75.109.163:44534.service - OpenSSH per-connection server daemon (147.75.109.163:44534). May 7 23:46:50.376223 amazon-ssm-agent[2081]: 2025-05-07 23:46:50 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2200) started May 7 23:46:50.476781 amazon-ssm-agent[2081]: 2025-05-07 23:46:50 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds May 7 23:46:50.548864 sshd[2202]: Accepted publickey for core from 147.75.109.163 port 44534 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:46:50.552513 sshd-session[2202]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:46:50.564000 systemd-logind[1936]: New session 2 of user core. May 7 23:46:50.571407 systemd[1]: Started session-2.scope - Session 2 of User core. May 7 23:46:50.699316 sshd[2212]: Connection closed by 147.75.109.163 port 44534 May 7 23:46:50.700893 sshd-session[2202]: pam_unix(sshd:session): session closed for user core May 7 23:46:50.709325 systemd[1]: sshd@1-172.31.28.15:22-147.75.109.163:44534.service: Deactivated successfully. May 7 23:46:50.713857 systemd[1]: session-2.scope: Deactivated successfully. May 7 23:46:50.715511 systemd-logind[1936]: Session 2 logged out. Waiting for processes to exit. May 7 23:46:50.718221 systemd-logind[1936]: Removed session 2. May 7 23:46:50.746358 systemd[1]: Started sshd@2-172.31.28.15:22-147.75.109.163:44550.service - OpenSSH per-connection server daemon (147.75.109.163:44550). May 7 23:46:50.912840 kubelet[2185]: E0507 23:46:50.912541 2185 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 7 23:46:50.917195 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 7 23:46:50.917534 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 7 23:46:50.918198 systemd[1]: kubelet.service: Consumed 1.306s CPU time, 248.5M memory peak. May 7 23:46:50.932855 sshd[2219]: Accepted publickey for core from 147.75.109.163 port 44550 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:46:50.935362 sshd-session[2219]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:46:50.944657 systemd-logind[1936]: New session 3 of user core. May 7 23:46:50.951375 systemd[1]: Started session-3.scope - Session 3 of User core. May 7 23:46:51.071777 sshd[2222]: Connection closed by 147.75.109.163 port 44550 May 7 23:46:51.069221 sshd-session[2219]: pam_unix(sshd:session): session closed for user core May 7 23:46:51.153590 systemd[1]: sshd@2-172.31.28.15:22-147.75.109.163:44550.service: Deactivated successfully. May 7 23:46:51.157683 systemd[1]: session-3.scope: Deactivated successfully. May 7 23:46:51.162457 systemd-logind[1936]: Session 3 logged out. Waiting for processes to exit. May 7 23:46:51.170657 systemd[1]: Started sshd@3-172.31.28.15:22-147.75.109.163:44554.service - OpenSSH per-connection server daemon (147.75.109.163:44554). May 7 23:46:51.191351 systemd-logind[1936]: Removed session 3. May 7 23:46:51.359212 sshd[2227]: Accepted publickey for core from 147.75.109.163 port 44554 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:46:51.361646 sshd-session[2227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:46:51.370857 systemd-logind[1936]: New session 4 of user core. May 7 23:46:51.380384 systemd[1]: Started session-4.scope - Session 4 of User core. May 7 23:46:51.506262 sshd[2230]: Connection closed by 147.75.109.163 port 44554 May 7 23:46:51.507041 sshd-session[2227]: pam_unix(sshd:session): session closed for user core May 7 23:46:51.513518 systemd[1]: sshd@3-172.31.28.15:22-147.75.109.163:44554.service: Deactivated successfully. May 7 23:46:51.517418 systemd[1]: session-4.scope: Deactivated successfully. May 7 23:46:51.519162 systemd-logind[1936]: Session 4 logged out. Waiting for processes to exit. May 7 23:46:51.521090 systemd-logind[1936]: Removed session 4. May 7 23:46:51.555580 systemd[1]: Started sshd@4-172.31.28.15:22-147.75.109.163:44560.service - OpenSSH per-connection server daemon (147.75.109.163:44560). May 7 23:46:51.734448 sshd[2236]: Accepted publickey for core from 147.75.109.163 port 44560 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:46:51.736904 sshd-session[2236]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:46:51.748254 systemd-logind[1936]: New session 5 of user core. May 7 23:46:51.754700 systemd[1]: Started session-5.scope - Session 5 of User core. May 7 23:46:51.870039 sudo[2240]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 7 23:46:51.870756 sudo[2240]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 7 23:46:51.889828 sudo[2240]: pam_unix(sudo:session): session closed for user root May 7 23:46:51.913165 sshd[2239]: Connection closed by 147.75.109.163 port 44560 May 7 23:46:51.914337 sshd-session[2236]: pam_unix(sshd:session): session closed for user core May 7 23:46:51.920169 systemd-logind[1936]: Session 5 logged out. Waiting for processes to exit. May 7 23:46:51.920533 systemd[1]: sshd@4-172.31.28.15:22-147.75.109.163:44560.service: Deactivated successfully. May 7 23:46:51.924230 systemd[1]: session-5.scope: Deactivated successfully. May 7 23:46:51.928760 systemd-logind[1936]: Removed session 5. May 7 23:46:51.956628 systemd[1]: Started sshd@5-172.31.28.15:22-147.75.109.163:44574.service - OpenSSH per-connection server daemon (147.75.109.163:44574). May 7 23:46:52.151507 sshd[2246]: Accepted publickey for core from 147.75.109.163 port 44574 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:46:52.154037 sshd-session[2246]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:46:52.165385 systemd-logind[1936]: New session 6 of user core. May 7 23:46:52.167401 systemd[1]: Started session-6.scope - Session 6 of User core. May 7 23:46:52.274054 sudo[2250]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 7 23:46:52.274750 sudo[2250]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 7 23:46:52.280986 sudo[2250]: pam_unix(sudo:session): session closed for user root May 7 23:46:52.290973 sudo[2249]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 7 23:46:52.291634 sudo[2249]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 7 23:46:52.312737 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 7 23:46:52.360947 augenrules[2272]: No rules May 7 23:46:52.363258 systemd[1]: audit-rules.service: Deactivated successfully. May 7 23:46:52.365210 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 7 23:46:52.367455 sudo[2249]: pam_unix(sudo:session): session closed for user root May 7 23:46:52.392834 sshd[2248]: Connection closed by 147.75.109.163 port 44574 May 7 23:46:52.392098 sshd-session[2246]: pam_unix(sshd:session): session closed for user core May 7 23:46:52.398448 systemd[1]: sshd@5-172.31.28.15:22-147.75.109.163:44574.service: Deactivated successfully. May 7 23:46:52.401509 systemd[1]: session-6.scope: Deactivated successfully. May 7 23:46:52.403708 systemd-logind[1936]: Session 6 logged out. Waiting for processes to exit. May 7 23:46:52.405935 systemd-logind[1936]: Removed session 6. May 7 23:46:52.438852 systemd[1]: Started sshd@6-172.31.28.15:22-147.75.109.163:44578.service - OpenSSH per-connection server daemon (147.75.109.163:44578). May 7 23:46:52.634569 sshd[2281]: Accepted publickey for core from 147.75.109.163 port 44578 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:46:52.636972 sshd-session[2281]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:46:52.645989 systemd-logind[1936]: New session 7 of user core. May 7 23:46:52.664388 systemd[1]: Started session-7.scope - Session 7 of User core. May 7 23:46:52.768474 sudo[2284]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 7 23:46:52.769100 sudo[2284]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 7 23:46:53.257621 systemd[1]: Starting docker.service - Docker Application Container Engine... May 7 23:46:53.260470 (dockerd)[2300]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 7 23:46:53.594263 dockerd[2300]: time="2025-05-07T23:46:53.594015903Z" level=info msg="Starting up" May 7 23:46:53.822895 dockerd[2300]: time="2025-05-07T23:46:53.822821116Z" level=info msg="Loading containers: start." May 7 23:46:54.069149 kernel: Initializing XFRM netlink socket May 7 23:46:54.105765 (udev-worker)[2323]: Network interface NamePolicy= disabled on kernel command line. May 7 23:46:54.198491 systemd-networkd[1866]: docker0: Link UP May 7 23:46:54.240559 dockerd[2300]: time="2025-05-07T23:46:54.240503387Z" level=info msg="Loading containers: done." May 7 23:46:54.263209 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck37028471-merged.mount: Deactivated successfully. May 7 23:46:54.270016 dockerd[2300]: time="2025-05-07T23:46:54.269903320Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 7 23:46:54.270280 dockerd[2300]: time="2025-05-07T23:46:54.270041557Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 May 7 23:46:54.270358 dockerd[2300]: time="2025-05-07T23:46:54.270302615Z" level=info msg="Daemon has completed initialization" May 7 23:46:54.326045 dockerd[2300]: time="2025-05-07T23:46:54.325863982Z" level=info msg="API listen on /run/docker.sock" May 7 23:46:54.326609 systemd[1]: Started docker.service - Docker Application Container Engine. May 7 23:46:55.409874 containerd[1963]: time="2025-05-07T23:46:55.409818220Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\"" May 7 23:46:55.997980 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4051824730.mount: Deactivated successfully. May 7 23:46:57.345163 containerd[1963]: time="2025-05-07T23:46:57.344671337Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:57.346368 containerd[1963]: time="2025-05-07T23:46:57.346142660Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.4: active requests=0, bytes read=26233118" May 7 23:46:57.347495 containerd[1963]: time="2025-05-07T23:46:57.347409341Z" level=info msg="ImageCreate event name:\"sha256:ab579d62aa850c7d0eca948aad11fcf813743e3b6c9742241c32cb4f1638968b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:57.352984 containerd[1963]: time="2025-05-07T23:46:57.352904379Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:57.355827 containerd[1963]: time="2025-05-07T23:46:57.355195196Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.4\" with image id \"sha256:ab579d62aa850c7d0eca948aad11fcf813743e3b6c9742241c32cb4f1638968b\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\", size \"26229918\" in 1.945314664s" May 7 23:46:57.355827 containerd[1963]: time="2025-05-07T23:46:57.355262045Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\" returns image reference \"sha256:ab579d62aa850c7d0eca948aad11fcf813743e3b6c9742241c32cb4f1638968b\"" May 7 23:46:57.356690 containerd[1963]: time="2025-05-07T23:46:57.356649147Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\"" May 7 23:46:58.721548 containerd[1963]: time="2025-05-07T23:46:58.720970260Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:58.722659 containerd[1963]: time="2025-05-07T23:46:58.722594047Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.4: active requests=0, bytes read=22529571" May 7 23:46:58.723060 containerd[1963]: time="2025-05-07T23:46:58.723006826Z" level=info msg="ImageCreate event name:\"sha256:79534fade29d07745acc698bbf598b0604a9ea1fd7917822c816a74fc0b55965\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:58.730975 containerd[1963]: time="2025-05-07T23:46:58.730907098Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:58.733383 containerd[1963]: time="2025-05-07T23:46:58.733331445Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.4\" with image id \"sha256:79534fade29d07745acc698bbf598b0604a9ea1fd7917822c816a74fc0b55965\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\", size \"23971132\" in 1.376490791s" May 7 23:46:58.733662 containerd[1963]: time="2025-05-07T23:46:58.733520588Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\" returns image reference \"sha256:79534fade29d07745acc698bbf598b0604a9ea1fd7917822c816a74fc0b55965\"" May 7 23:46:58.735162 containerd[1963]: time="2025-05-07T23:46:58.735019033Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\"" May 7 23:46:59.968178 containerd[1963]: time="2025-05-07T23:46:59.967674377Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:59.969836 containerd[1963]: time="2025-05-07T23:46:59.969753552Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.4: active requests=0, bytes read=17482173" May 7 23:46:59.971387 containerd[1963]: time="2025-05-07T23:46:59.971302674Z" level=info msg="ImageCreate event name:\"sha256:730fbc2590716b8202fcdd928a813b847575ebf03911a059979257cd6cbb8245\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:59.976832 containerd[1963]: time="2025-05-07T23:46:59.976736302Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:46:59.979295 containerd[1963]: time="2025-05-07T23:46:59.979136697Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.4\" with image id \"sha256:730fbc2590716b8202fcdd928a813b847575ebf03911a059979257cd6cbb8245\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\", size \"18923752\" in 1.244036948s" May 7 23:46:59.979295 containerd[1963]: time="2025-05-07T23:46:59.979194110Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\" returns image reference \"sha256:730fbc2590716b8202fcdd928a813b847575ebf03911a059979257cd6cbb8245\"" May 7 23:46:59.981161 containerd[1963]: time="2025-05-07T23:46:59.980940287Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\"" May 7 23:47:01.168887 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 7 23:47:01.178557 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:47:01.275246 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3612543661.mount: Deactivated successfully. May 7 23:47:01.600519 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:01.608633 (kubelet)[2565]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 7 23:47:01.707970 kubelet[2565]: E0507 23:47:01.707225 2565 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 7 23:47:01.715215 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 7 23:47:01.715565 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 7 23:47:01.716682 systemd[1]: kubelet.service: Consumed 321ms CPU time, 102.8M memory peak. May 7 23:47:02.036247 containerd[1963]: time="2025-05-07T23:47:02.036041622Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:02.038012 containerd[1963]: time="2025-05-07T23:47:02.037889550Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.4: active requests=0, bytes read=27370351" May 7 23:47:02.039255 containerd[1963]: time="2025-05-07T23:47:02.039169773Z" level=info msg="ImageCreate event name:\"sha256:62c496efa595c8eb7d098e43430b2b94ad66812214759a7ea9daaaa1ed901fc7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:02.043158 containerd[1963]: time="2025-05-07T23:47:02.043070187Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:02.044896 containerd[1963]: time="2025-05-07T23:47:02.044687347Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.4\" with image id \"sha256:62c496efa595c8eb7d098e43430b2b94ad66812214759a7ea9daaaa1ed901fc7\", repo tag \"registry.k8s.io/kube-proxy:v1.32.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\", size \"27369370\" in 2.063687546s" May 7 23:47:02.044896 containerd[1963]: time="2025-05-07T23:47:02.044747965Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\" returns image reference \"sha256:62c496efa595c8eb7d098e43430b2b94ad66812214759a7ea9daaaa1ed901fc7\"" May 7 23:47:02.046197 containerd[1963]: time="2025-05-07T23:47:02.046134214Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" May 7 23:47:02.580610 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount546873481.mount: Deactivated successfully. May 7 23:47:03.720007 containerd[1963]: time="2025-05-07T23:47:03.719939346Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:03.722222 containerd[1963]: time="2025-05-07T23:47:03.722149182Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951622" May 7 23:47:03.723170 containerd[1963]: time="2025-05-07T23:47:03.723070907Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:03.730290 containerd[1963]: time="2025-05-07T23:47:03.730238213Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:03.732559 containerd[1963]: time="2025-05-07T23:47:03.732266519Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.68606566s" May 7 23:47:03.732559 containerd[1963]: time="2025-05-07T23:47:03.732349216Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" May 7 23:47:03.733517 containerd[1963]: time="2025-05-07T23:47:03.733218690Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 7 23:47:04.230389 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4290643749.mount: Deactivated successfully. May 7 23:47:04.239416 containerd[1963]: time="2025-05-07T23:47:04.237894326Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:04.240576 containerd[1963]: time="2025-05-07T23:47:04.240519750Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" May 7 23:47:04.242932 containerd[1963]: time="2025-05-07T23:47:04.242889794Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:04.248096 containerd[1963]: time="2025-05-07T23:47:04.248044931Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:04.249759 containerd[1963]: time="2025-05-07T23:47:04.249701843Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 516.428681ms" May 7 23:47:04.249915 containerd[1963]: time="2025-05-07T23:47:04.249755029Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" May 7 23:47:04.250450 containerd[1963]: time="2025-05-07T23:47:04.250366159Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" May 7 23:47:04.749440 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2925226245.mount: Deactivated successfully. May 7 23:47:06.869776 containerd[1963]: time="2025-05-07T23:47:06.869710238Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:06.898253 containerd[1963]: time="2025-05-07T23:47:06.898174075Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67812469" May 7 23:47:06.942375 containerd[1963]: time="2025-05-07T23:47:06.941139336Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:07.003579 containerd[1963]: time="2025-05-07T23:47:07.003326838Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:07.007238 containerd[1963]: time="2025-05-07T23:47:07.006462001Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 2.756039882s" May 7 23:47:07.007679 containerd[1963]: time="2025-05-07T23:47:07.007600422Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" May 7 23:47:11.967966 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 7 23:47:11.977418 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:47:12.330484 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:12.341003 (kubelet)[2711]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 7 23:47:12.418147 kubelet[2711]: E0507 23:47:12.416893 2711 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 7 23:47:12.421461 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 7 23:47:12.421982 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 7 23:47:12.422918 systemd[1]: kubelet.service: Consumed 284ms CPU time, 104.1M memory peak. May 7 23:47:13.467722 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:13.468669 systemd[1]: kubelet.service: Consumed 284ms CPU time, 104.1M memory peak. May 7 23:47:13.486598 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:47:13.549147 systemd[1]: Reload requested from client PID 2725 ('systemctl') (unit session-7.scope)... May 7 23:47:13.549174 systemd[1]: Reloading... May 7 23:47:13.815156 zram_generator::config[2774]: No configuration found. May 7 23:47:14.046849 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 7 23:47:14.278013 systemd[1]: Reloading finished in 728 ms. May 7 23:47:14.374721 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:14.382671 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:47:14.386882 systemd[1]: kubelet.service: Deactivated successfully. May 7 23:47:14.387583 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:14.387664 systemd[1]: kubelet.service: Consumed 225ms CPU time, 90.1M memory peak. May 7 23:47:14.394850 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:47:14.715379 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:14.730654 (kubelet)[2836]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 7 23:47:14.802966 kubelet[2836]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 7 23:47:14.802966 kubelet[2836]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 7 23:47:14.802966 kubelet[2836]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 7 23:47:14.803548 kubelet[2836]: I0507 23:47:14.803077 2836 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 7 23:47:15.909975 kubelet[2836]: I0507 23:47:15.909925 2836 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 7 23:47:15.910641 kubelet[2836]: I0507 23:47:15.910520 2836 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 7 23:47:15.911059 kubelet[2836]: I0507 23:47:15.911008 2836 server.go:954] "Client rotation is on, will bootstrap in background" May 7 23:47:15.952869 kubelet[2836]: E0507 23:47:15.952768 2836 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.28.15:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:15.956533 kubelet[2836]: I0507 23:47:15.956275 2836 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 7 23:47:15.969997 kubelet[2836]: E0507 23:47:15.969926 2836 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" May 7 23:47:15.969997 kubelet[2836]: I0507 23:47:15.969992 2836 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." May 7 23:47:15.976786 kubelet[2836]: I0507 23:47:15.976716 2836 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 7 23:47:15.977334 kubelet[2836]: I0507 23:47:15.977243 2836 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 7 23:47:15.977682 kubelet[2836]: I0507 23:47:15.977327 2836 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-28-15","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 7 23:47:15.977863 kubelet[2836]: I0507 23:47:15.977706 2836 topology_manager.go:138] "Creating topology manager with none policy" May 7 23:47:15.977863 kubelet[2836]: I0507 23:47:15.977727 2836 container_manager_linux.go:304] "Creating device plugin manager" May 7 23:47:15.977993 kubelet[2836]: I0507 23:47:15.977950 2836 state_mem.go:36] "Initialized new in-memory state store" May 7 23:47:15.983602 kubelet[2836]: I0507 23:47:15.983416 2836 kubelet.go:446] "Attempting to sync node with API server" May 7 23:47:15.983602 kubelet[2836]: I0507 23:47:15.983461 2836 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 7 23:47:15.983602 kubelet[2836]: I0507 23:47:15.983496 2836 kubelet.go:352] "Adding apiserver pod source" May 7 23:47:15.983602 kubelet[2836]: I0507 23:47:15.983518 2836 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 7 23:47:15.991932 kubelet[2836]: W0507 23:47:15.991784 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.28.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-28-15&limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:15.994179 kubelet[2836]: E0507 23:47:15.992276 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.28.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-28-15&limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:15.994179 kubelet[2836]: I0507 23:47:15.992457 2836 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" May 7 23:47:15.994179 kubelet[2836]: I0507 23:47:15.993276 2836 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 7 23:47:15.994179 kubelet[2836]: W0507 23:47:15.993385 2836 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 7 23:47:15.996339 kubelet[2836]: I0507 23:47:15.996292 2836 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 7 23:47:15.996483 kubelet[2836]: I0507 23:47:15.996362 2836 server.go:1287] "Started kubelet" May 7 23:47:15.998673 kubelet[2836]: W0507 23:47:15.998609 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.28.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:16.000247 kubelet[2836]: E0507 23:47:15.998882 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.28.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:16.000247 kubelet[2836]: I0507 23:47:15.999053 2836 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 7 23:47:16.004176 kubelet[2836]: I0507 23:47:16.002885 2836 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 7 23:47:16.004176 kubelet[2836]: I0507 23:47:16.003393 2836 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 7 23:47:16.004176 kubelet[2836]: I0507 23:47:16.003840 2836 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 7 23:47:16.006051 kubelet[2836]: E0507 23:47:16.005795 2836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.28.15:6443/api/v1/namespaces/default/events\": dial tcp 172.31.28.15:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-28-15.183d6380f45d12ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-28-15,UID:ip-172-31-28-15,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-28-15,},FirstTimestamp:2025-05-07 23:47:15.996324589 +0000 UTC m=+1.259614603,LastTimestamp:2025-05-07 23:47:15.996324589 +0000 UTC m=+1.259614603,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-28-15,}" May 7 23:47:16.009243 kubelet[2836]: I0507 23:47:16.009069 2836 server.go:490] "Adding debug handlers to kubelet server" May 7 23:47:16.010997 kubelet[2836]: I0507 23:47:16.010956 2836 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 7 23:47:16.015190 kubelet[2836]: I0507 23:47:16.015147 2836 volume_manager.go:297] "Starting Kubelet Volume Manager" May 7 23:47:16.015671 kubelet[2836]: E0507 23:47:16.015622 2836 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-28-15\" not found" May 7 23:47:16.017145 kubelet[2836]: E0507 23:47:16.016333 2836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.28.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-28-15?timeout=10s\": dial tcp 172.31.28.15:6443: connect: connection refused" interval="200ms" May 7 23:47:16.017145 kubelet[2836]: I0507 23:47:16.016434 2836 reconciler.go:26] "Reconciler: start to sync state" May 7 23:47:16.017145 kubelet[2836]: I0507 23:47:16.016480 2836 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 7 23:47:16.017145 kubelet[2836]: W0507 23:47:16.016982 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.28.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:16.017145 kubelet[2836]: E0507 23:47:16.017060 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.28.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:16.022389 kubelet[2836]: E0507 23:47:16.022157 2836 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 7 23:47:16.024148 kubelet[2836]: I0507 23:47:16.022715 2836 factory.go:221] Registration of the containerd container factory successfully May 7 23:47:16.024148 kubelet[2836]: I0507 23:47:16.022744 2836 factory.go:221] Registration of the systemd container factory successfully May 7 23:47:16.024148 kubelet[2836]: I0507 23:47:16.022864 2836 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 7 23:47:16.039065 kubelet[2836]: I0507 23:47:16.038974 2836 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 7 23:47:16.041249 kubelet[2836]: I0507 23:47:16.041195 2836 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 7 23:47:16.041249 kubelet[2836]: I0507 23:47:16.041242 2836 status_manager.go:227] "Starting to sync pod status with apiserver" May 7 23:47:16.041443 kubelet[2836]: I0507 23:47:16.041284 2836 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 7 23:47:16.041443 kubelet[2836]: I0507 23:47:16.041299 2836 kubelet.go:2388] "Starting kubelet main sync loop" May 7 23:47:16.041443 kubelet[2836]: E0507 23:47:16.041370 2836 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 7 23:47:16.064823 kubelet[2836]: W0507 23:47:16.064290 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.28.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:16.064823 kubelet[2836]: E0507 23:47:16.064637 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.28.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:16.072178 kubelet[2836]: I0507 23:47:16.072139 2836 cpu_manager.go:221] "Starting CPU manager" policy="none" May 7 23:47:16.072337 kubelet[2836]: I0507 23:47:16.072317 2836 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 7 23:47:16.072443 kubelet[2836]: I0507 23:47:16.072424 2836 state_mem.go:36] "Initialized new in-memory state store" May 7 23:47:16.075441 kubelet[2836]: I0507 23:47:16.075404 2836 policy_none.go:49] "None policy: Start" May 7 23:47:16.075597 kubelet[2836]: I0507 23:47:16.075578 2836 memory_manager.go:186] "Starting memorymanager" policy="None" May 7 23:47:16.075717 kubelet[2836]: I0507 23:47:16.075699 2836 state_mem.go:35] "Initializing new in-memory state store" May 7 23:47:16.086795 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 7 23:47:16.104865 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 7 23:47:16.116320 kubelet[2836]: E0507 23:47:16.116224 2836 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-28-15\" not found" May 7 23:47:16.123006 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 7 23:47:16.125535 kubelet[2836]: I0507 23:47:16.125488 2836 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 7 23:47:16.125958 kubelet[2836]: I0507 23:47:16.125775 2836 eviction_manager.go:189] "Eviction manager: starting control loop" May 7 23:47:16.125958 kubelet[2836]: I0507 23:47:16.125807 2836 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 7 23:47:16.126810 kubelet[2836]: I0507 23:47:16.126555 2836 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 7 23:47:16.129850 kubelet[2836]: E0507 23:47:16.129800 2836 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 7 23:47:16.130012 kubelet[2836]: E0507 23:47:16.129869 2836 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-28-15\" not found" May 7 23:47:16.161447 systemd[1]: Created slice kubepods-burstable-podeb44c7e6e41f00eddc76740dcfa6b38c.slice - libcontainer container kubepods-burstable-podeb44c7e6e41f00eddc76740dcfa6b38c.slice. May 7 23:47:16.175225 kubelet[2836]: E0507 23:47:16.175030 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:16.180190 systemd[1]: Created slice kubepods-burstable-pod9873d30601be48fd51168bc4b7f5df81.slice - libcontainer container kubepods-burstable-pod9873d30601be48fd51168bc4b7f5df81.slice. May 7 23:47:16.184503 kubelet[2836]: E0507 23:47:16.184165 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:16.189540 systemd[1]: Created slice kubepods-burstable-pod2a55cfd6270de58396bda8347fda87e4.slice - libcontainer container kubepods-burstable-pod2a55cfd6270de58396bda8347fda87e4.slice. May 7 23:47:16.193671 kubelet[2836]: E0507 23:47:16.193632 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:16.217092 kubelet[2836]: E0507 23:47:16.217019 2836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.28.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-28-15?timeout=10s\": dial tcp 172.31.28.15:6443: connect: connection refused" interval="400ms" May 7 23:47:16.218216 kubelet[2836]: I0507 23:47:16.218091 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:16.218216 kubelet[2836]: I0507 23:47:16.218167 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:16.218383 kubelet[2836]: I0507 23:47:16.218234 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9873d30601be48fd51168bc4b7f5df81-kubeconfig\") pod \"kube-scheduler-ip-172-31-28-15\" (UID: \"9873d30601be48fd51168bc4b7f5df81\") " pod="kube-system/kube-scheduler-ip-172-31-28-15" May 7 23:47:16.218383 kubelet[2836]: I0507 23:47:16.218275 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2a55cfd6270de58396bda8347fda87e4-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-28-15\" (UID: \"2a55cfd6270de58396bda8347fda87e4\") " pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:16.218383 kubelet[2836]: I0507 23:47:16.218315 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-ca-certs\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:16.218383 kubelet[2836]: I0507 23:47:16.218354 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-kubeconfig\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:16.218570 kubelet[2836]: I0507 23:47:16.218392 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2a55cfd6270de58396bda8347fda87e4-ca-certs\") pod \"kube-apiserver-ip-172-31-28-15\" (UID: \"2a55cfd6270de58396bda8347fda87e4\") " pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:16.218570 kubelet[2836]: I0507 23:47:16.218428 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2a55cfd6270de58396bda8347fda87e4-k8s-certs\") pod \"kube-apiserver-ip-172-31-28-15\" (UID: \"2a55cfd6270de58396bda8347fda87e4\") " pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:16.218570 kubelet[2836]: I0507 23:47:16.218468 2836 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-k8s-certs\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:16.228727 kubelet[2836]: I0507 23:47:16.228682 2836 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-28-15" May 7 23:47:16.229433 kubelet[2836]: E0507 23:47:16.229386 2836 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.28.15:6443/api/v1/nodes\": dial tcp 172.31.28.15:6443: connect: connection refused" node="ip-172-31-28-15" May 7 23:47:16.432807 kubelet[2836]: I0507 23:47:16.432631 2836 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-28-15" May 7 23:47:16.433245 kubelet[2836]: E0507 23:47:16.433178 2836 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.28.15:6443/api/v1/nodes\": dial tcp 172.31.28.15:6443: connect: connection refused" node="ip-172-31-28-15" May 7 23:47:16.477252 containerd[1963]: time="2025-05-07T23:47:16.477168218Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-28-15,Uid:eb44c7e6e41f00eddc76740dcfa6b38c,Namespace:kube-system,Attempt:0,}" May 7 23:47:16.486262 containerd[1963]: time="2025-05-07T23:47:16.485871703Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-28-15,Uid:9873d30601be48fd51168bc4b7f5df81,Namespace:kube-system,Attempt:0,}" May 7 23:47:16.495592 containerd[1963]: time="2025-05-07T23:47:16.495536353Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-28-15,Uid:2a55cfd6270de58396bda8347fda87e4,Namespace:kube-system,Attempt:0,}" May 7 23:47:16.618430 kubelet[2836]: E0507 23:47:16.618357 2836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.28.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-28-15?timeout=10s\": dial tcp 172.31.28.15:6443: connect: connection refused" interval="800ms" May 7 23:47:16.836233 kubelet[2836]: I0507 23:47:16.836012 2836 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-28-15" May 7 23:47:16.836957 kubelet[2836]: E0507 23:47:16.836896 2836 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.28.15:6443/api/v1/nodes\": dial tcp 172.31.28.15:6443: connect: connection refused" node="ip-172-31-28-15" May 7 23:47:16.928692 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount315506645.mount: Deactivated successfully. May 7 23:47:16.936205 containerd[1963]: time="2025-05-07T23:47:16.935797169Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 7 23:47:16.937943 containerd[1963]: time="2025-05-07T23:47:16.937886309Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 7 23:47:16.939818 containerd[1963]: time="2025-05-07T23:47:16.939753050Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" May 7 23:47:16.940177 containerd[1963]: time="2025-05-07T23:47:16.940089086Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" May 7 23:47:16.942990 containerd[1963]: time="2025-05-07T23:47:16.942894480Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 7 23:47:16.944316 containerd[1963]: time="2025-05-07T23:47:16.944159300Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" May 7 23:47:16.948167 containerd[1963]: time="2025-05-07T23:47:16.948089968Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 7 23:47:16.953241 containerd[1963]: time="2025-05-07T23:47:16.952857419Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 466.882271ms" May 7 23:47:16.955503 containerd[1963]: time="2025-05-07T23:47:16.954880298Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 7 23:47:16.960135 containerd[1963]: time="2025-05-07T23:47:16.959709759Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 482.424327ms" May 7 23:47:16.962807 containerd[1963]: time="2025-05-07T23:47:16.962730601Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 467.082159ms" May 7 23:47:17.098254 kubelet[2836]: E0507 23:47:17.097536 2836 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.28.15:6443/api/v1/namespaces/default/events\": dial tcp 172.31.28.15:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-28-15.183d6380f45d12ed default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-28-15,UID:ip-172-31-28-15,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-28-15,},FirstTimestamp:2025-05-07 23:47:15.996324589 +0000 UTC m=+1.259614603,LastTimestamp:2025-05-07 23:47:15.996324589 +0000 UTC m=+1.259614603,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-28-15,}" May 7 23:47:17.158052 containerd[1963]: time="2025-05-07T23:47:17.157556731Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:17.158401 containerd[1963]: time="2025-05-07T23:47:17.158266837Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:17.158801 containerd[1963]: time="2025-05-07T23:47:17.158574227Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:17.165315 containerd[1963]: time="2025-05-07T23:47:17.164957301Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:17.168145 containerd[1963]: time="2025-05-07T23:47:17.165702573Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:17.168145 containerd[1963]: time="2025-05-07T23:47:17.165278990Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:17.168145 containerd[1963]: time="2025-05-07T23:47:17.165555092Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:17.168597 containerd[1963]: time="2025-05-07T23:47:17.166491584Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:17.168597 containerd[1963]: time="2025-05-07T23:47:17.166569347Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:17.168597 containerd[1963]: time="2025-05-07T23:47:17.166594019Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:17.168597 containerd[1963]: time="2025-05-07T23:47:17.166706167Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:17.169008 containerd[1963]: time="2025-05-07T23:47:17.168889854Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:17.209331 kubelet[2836]: W0507 23:47:17.209057 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.28.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:17.210250 kubelet[2836]: E0507 23:47:17.210180 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.28.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:17.210876 systemd[1]: Started cri-containerd-685cd4ce7a815c0569471176a5d2450a60fc822c19abd1fff5ba5fe80a949e81.scope - libcontainer container 685cd4ce7a815c0569471176a5d2450a60fc822c19abd1fff5ba5fe80a949e81. May 7 23:47:17.220605 systemd[1]: Started cri-containerd-22397ac0021320ed2392e2473c4ba3fde4d7af1df0acb997885e0925bb0e4013.scope - libcontainer container 22397ac0021320ed2392e2473c4ba3fde4d7af1df0acb997885e0925bb0e4013. May 7 23:47:17.232933 systemd[1]: Started cri-containerd-a3dbf5d57359bd2ee76ef79e0fd93d501bb8a920f9aa0b309752bc6c55f24f24.scope - libcontainer container a3dbf5d57359bd2ee76ef79e0fd93d501bb8a920f9aa0b309752bc6c55f24f24. May 7 23:47:17.268308 kubelet[2836]: W0507 23:47:17.268210 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.28.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:17.268563 kubelet[2836]: E0507 23:47:17.268511 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.28.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:17.352867 containerd[1963]: time="2025-05-07T23:47:17.352695136Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-28-15,Uid:eb44c7e6e41f00eddc76740dcfa6b38c,Namespace:kube-system,Attempt:0,} returns sandbox id \"a3dbf5d57359bd2ee76ef79e0fd93d501bb8a920f9aa0b309752bc6c55f24f24\"" May 7 23:47:17.354687 containerd[1963]: time="2025-05-07T23:47:17.354603310Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-28-15,Uid:2a55cfd6270de58396bda8347fda87e4,Namespace:kube-system,Attempt:0,} returns sandbox id \"685cd4ce7a815c0569471176a5d2450a60fc822c19abd1fff5ba5fe80a949e81\"" May 7 23:47:17.361969 containerd[1963]: time="2025-05-07T23:47:17.361899764Z" level=info msg="CreateContainer within sandbox \"a3dbf5d57359bd2ee76ef79e0fd93d501bb8a920f9aa0b309752bc6c55f24f24\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 7 23:47:17.362555 containerd[1963]: time="2025-05-07T23:47:17.362229629Z" level=info msg="CreateContainer within sandbox \"685cd4ce7a815c0569471176a5d2450a60fc822c19abd1fff5ba5fe80a949e81\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 7 23:47:17.371711 containerd[1963]: time="2025-05-07T23:47:17.371643726Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-28-15,Uid:9873d30601be48fd51168bc4b7f5df81,Namespace:kube-system,Attempt:0,} returns sandbox id \"22397ac0021320ed2392e2473c4ba3fde4d7af1df0acb997885e0925bb0e4013\"" May 7 23:47:17.373285 kubelet[2836]: W0507 23:47:17.373080 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.28.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-28-15&limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:17.373442 kubelet[2836]: E0507 23:47:17.373309 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.28.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-28-15&limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:17.380696 containerd[1963]: time="2025-05-07T23:47:17.380632090Z" level=info msg="CreateContainer within sandbox \"22397ac0021320ed2392e2473c4ba3fde4d7af1df0acb997885e0925bb0e4013\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 7 23:47:17.387099 containerd[1963]: time="2025-05-07T23:47:17.387024024Z" level=info msg="CreateContainer within sandbox \"685cd4ce7a815c0569471176a5d2450a60fc822c19abd1fff5ba5fe80a949e81\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"6f1698649396812800e3b7ebdb141b43436cf5dde49f7563477abb5b862d296b\"" May 7 23:47:17.388827 containerd[1963]: time="2025-05-07T23:47:17.388026549Z" level=info msg="StartContainer for \"6f1698649396812800e3b7ebdb141b43436cf5dde49f7563477abb5b862d296b\"" May 7 23:47:17.392808 containerd[1963]: time="2025-05-07T23:47:17.392742830Z" level=info msg="CreateContainer within sandbox \"a3dbf5d57359bd2ee76ef79e0fd93d501bb8a920f9aa0b309752bc6c55f24f24\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"804e961ed0d391bbcae82bd292c5880bf994855b130e03f1e149001cf0dd6791\"" May 7 23:47:17.394788 containerd[1963]: time="2025-05-07T23:47:17.394738912Z" level=info msg="StartContainer for \"804e961ed0d391bbcae82bd292c5880bf994855b130e03f1e149001cf0dd6791\"" May 7 23:47:17.406676 containerd[1963]: time="2025-05-07T23:47:17.406623411Z" level=info msg="CreateContainer within sandbox \"22397ac0021320ed2392e2473c4ba3fde4d7af1df0acb997885e0925bb0e4013\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"54ad642a90e7ff3354f23c4158e408e65a4c9ffc8954185207a41d419e882528\"" May 7 23:47:17.408428 containerd[1963]: time="2025-05-07T23:47:17.408333666Z" level=info msg="StartContainer for \"54ad642a90e7ff3354f23c4158e408e65a4c9ffc8954185207a41d419e882528\"" May 7 23:47:17.419846 kubelet[2836]: E0507 23:47:17.419727 2836 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.28.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-28-15?timeout=10s\": dial tcp 172.31.28.15:6443: connect: connection refused" interval="1.6s" May 7 23:47:17.458613 systemd[1]: Started cri-containerd-6f1698649396812800e3b7ebdb141b43436cf5dde49f7563477abb5b862d296b.scope - libcontainer container 6f1698649396812800e3b7ebdb141b43436cf5dde49f7563477abb5b862d296b. May 7 23:47:17.479826 systemd[1]: Started cri-containerd-804e961ed0d391bbcae82bd292c5880bf994855b130e03f1e149001cf0dd6791.scope - libcontainer container 804e961ed0d391bbcae82bd292c5880bf994855b130e03f1e149001cf0dd6791. May 7 23:47:17.515180 systemd[1]: Started cri-containerd-54ad642a90e7ff3354f23c4158e408e65a4c9ffc8954185207a41d419e882528.scope - libcontainer container 54ad642a90e7ff3354f23c4158e408e65a4c9ffc8954185207a41d419e882528. May 7 23:47:17.540161 kubelet[2836]: W0507 23:47:17.539618 2836 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.28.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.28.15:6443: connect: connection refused May 7 23:47:17.540338 kubelet[2836]: E0507 23:47:17.540098 2836 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.28.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.28.15:6443: connect: connection refused" logger="UnhandledError" May 7 23:47:17.575712 containerd[1963]: time="2025-05-07T23:47:17.575629966Z" level=info msg="StartContainer for \"6f1698649396812800e3b7ebdb141b43436cf5dde49f7563477abb5b862d296b\" returns successfully" May 7 23:47:17.632013 containerd[1963]: time="2025-05-07T23:47:17.631058234Z" level=info msg="StartContainer for \"804e961ed0d391bbcae82bd292c5880bf994855b130e03f1e149001cf0dd6791\" returns successfully" May 7 23:47:17.642723 kubelet[2836]: I0507 23:47:17.642240 2836 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-28-15" May 7 23:47:17.643344 kubelet[2836]: E0507 23:47:17.643025 2836 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.28.15:6443/api/v1/nodes\": dial tcp 172.31.28.15:6443: connect: connection refused" node="ip-172-31-28-15" May 7 23:47:17.655341 containerd[1963]: time="2025-05-07T23:47:17.655269678Z" level=info msg="StartContainer for \"54ad642a90e7ff3354f23c4158e408e65a4c9ffc8954185207a41d419e882528\" returns successfully" May 7 23:47:17.957676 systemd[1]: systemd-hostnamed.service: Deactivated successfully. May 7 23:47:18.081063 kubelet[2836]: E0507 23:47:18.080660 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:18.085680 kubelet[2836]: E0507 23:47:18.085623 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:18.089744 kubelet[2836]: E0507 23:47:18.089699 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:19.094184 kubelet[2836]: E0507 23:47:19.094096 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:19.096519 kubelet[2836]: E0507 23:47:19.096270 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:19.245970 kubelet[2836]: I0507 23:47:19.245885 2836 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-28-15" May 7 23:47:20.098383 kubelet[2836]: E0507 23:47:20.098104 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:20.240928 kubelet[2836]: E0507 23:47:20.240524 2836 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:20.960490 kubelet[2836]: E0507 23:47:20.960427 2836 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-28-15\" not found" node="ip-172-31-28-15" May 7 23:47:20.992675 kubelet[2836]: I0507 23:47:20.992246 2836 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-28-15" May 7 23:47:20.992675 kubelet[2836]: E0507 23:47:20.992326 2836 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"ip-172-31-28-15\": node \"ip-172-31-28-15\" not found" May 7 23:47:21.000662 kubelet[2836]: I0507 23:47:21.000618 2836 apiserver.go:52] "Watching apiserver" May 7 23:47:21.016440 kubelet[2836]: I0507 23:47:21.016314 2836 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:21.089673 kubelet[2836]: E0507 23:47:21.089328 2836 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-28-15\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:21.089673 kubelet[2836]: I0507 23:47:21.089375 2836 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-28-15" May 7 23:47:21.097213 kubelet[2836]: E0507 23:47:21.097156 2836 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-28-15\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-28-15" May 7 23:47:21.097213 kubelet[2836]: I0507 23:47:21.097204 2836 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:21.102088 kubelet[2836]: E0507 23:47:21.102030 2836 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-28-15\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:21.119147 kubelet[2836]: I0507 23:47:21.117261 2836 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 7 23:47:23.199944 systemd[1]: Reload requested from client PID 3115 ('systemctl') (unit session-7.scope)... May 7 23:47:23.199971 systemd[1]: Reloading... May 7 23:47:23.447218 zram_generator::config[3166]: No configuration found. May 7 23:47:23.669882 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 7 23:47:23.929332 systemd[1]: Reloading finished in 728 ms. May 7 23:47:23.968685 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:47:23.987216 systemd[1]: kubelet.service: Deactivated successfully. May 7 23:47:23.989182 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:23.989279 systemd[1]: kubelet.service: Consumed 1.945s CPU time, 122.7M memory peak. May 7 23:47:23.996645 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 7 23:47:24.383423 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 7 23:47:24.397725 (kubelet)[3220]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 7 23:47:24.498349 kubelet[3220]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 7 23:47:24.498838 kubelet[3220]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 7 23:47:24.499578 kubelet[3220]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 7 23:47:24.499578 kubelet[3220]: I0507 23:47:24.499067 3220 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 7 23:47:24.528294 kubelet[3220]: I0507 23:47:24.527689 3220 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 7 23:47:24.528294 kubelet[3220]: I0507 23:47:24.527751 3220 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 7 23:47:24.530042 kubelet[3220]: I0507 23:47:24.529975 3220 server.go:954] "Client rotation is on, will bootstrap in background" May 7 23:47:24.533352 kubelet[3220]: I0507 23:47:24.533302 3220 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 7 23:47:24.539090 kubelet[3220]: I0507 23:47:24.539004 3220 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 7 23:47:24.547923 kubelet[3220]: E0507 23:47:24.547293 3220 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" May 7 23:47:24.547923 kubelet[3220]: I0507 23:47:24.547347 3220 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." May 7 23:47:24.553196 kubelet[3220]: I0507 23:47:24.552381 3220 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 7 23:47:24.553196 kubelet[3220]: I0507 23:47:24.552741 3220 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 7 23:47:24.553438 kubelet[3220]: I0507 23:47:24.552793 3220 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-28-15","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 7 23:47:24.553666 kubelet[3220]: I0507 23:47:24.553641 3220 topology_manager.go:138] "Creating topology manager with none policy" May 7 23:47:24.553773 kubelet[3220]: I0507 23:47:24.553754 3220 container_manager_linux.go:304] "Creating device plugin manager" May 7 23:47:24.553955 kubelet[3220]: I0507 23:47:24.553935 3220 state_mem.go:36] "Initialized new in-memory state store" May 7 23:47:24.554439 kubelet[3220]: I0507 23:47:24.554410 3220 kubelet.go:446] "Attempting to sync node with API server" May 7 23:47:24.554744 kubelet[3220]: I0507 23:47:24.554718 3220 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 7 23:47:24.554951 kubelet[3220]: I0507 23:47:24.554932 3220 kubelet.go:352] "Adding apiserver pod source" May 7 23:47:24.555259 kubelet[3220]: I0507 23:47:24.555202 3220 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 7 23:47:24.565434 kubelet[3220]: I0507 23:47:24.565364 3220 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" May 7 23:47:24.566337 kubelet[3220]: I0507 23:47:24.566291 3220 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 7 23:47:24.567152 kubelet[3220]: I0507 23:47:24.567049 3220 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 7 23:47:24.575884 sudo[3235]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 7 23:47:24.576663 sudo[3235]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) May 7 23:47:24.580127 kubelet[3220]: I0507 23:47:24.579391 3220 server.go:1287] "Started kubelet" May 7 23:47:24.592401 kubelet[3220]: I0507 23:47:24.591918 3220 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 7 23:47:24.603291 kubelet[3220]: I0507 23:47:24.603235 3220 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 7 23:47:24.613317 kubelet[3220]: I0507 23:47:24.608364 3220 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 7 23:47:24.616870 kubelet[3220]: I0507 23:47:24.608860 3220 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 7 23:47:24.617088 kubelet[3220]: I0507 23:47:24.617049 3220 volume_manager.go:297] "Starting Kubelet Volume Manager" May 7 23:47:24.647414 kubelet[3220]: I0507 23:47:24.617274 3220 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 7 23:47:24.647414 kubelet[3220]: E0507 23:47:24.617453 3220 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-28-15\" not found" May 7 23:47:24.647414 kubelet[3220]: I0507 23:47:24.634574 3220 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 7 23:47:24.649560 kubelet[3220]: I0507 23:47:24.649226 3220 reconciler.go:26] "Reconciler: start to sync state" May 7 23:47:24.672346 kubelet[3220]: I0507 23:47:24.669727 3220 server.go:490] "Adding debug handlers to kubelet server" May 7 23:47:24.694343 kubelet[3220]: I0507 23:47:24.694295 3220 factory.go:221] Registration of the systemd container factory successfully May 7 23:47:24.694504 kubelet[3220]: I0507 23:47:24.694470 3220 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 7 23:47:24.697751 kubelet[3220]: I0507 23:47:24.697689 3220 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 7 23:47:24.703894 kubelet[3220]: I0507 23:47:24.703828 3220 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 7 23:47:24.703894 kubelet[3220]: I0507 23:47:24.703878 3220 status_manager.go:227] "Starting to sync pod status with apiserver" May 7 23:47:24.704099 kubelet[3220]: I0507 23:47:24.703912 3220 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 7 23:47:24.704099 kubelet[3220]: I0507 23:47:24.703928 3220 kubelet.go:2388] "Starting kubelet main sync loop" May 7 23:47:24.704099 kubelet[3220]: E0507 23:47:24.703999 3220 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 7 23:47:24.714514 kubelet[3220]: I0507 23:47:24.713793 3220 factory.go:221] Registration of the containerd container factory successfully May 7 23:47:24.736790 kubelet[3220]: E0507 23:47:24.736743 3220 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 7 23:47:24.807041 kubelet[3220]: E0507 23:47:24.804292 3220 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 7 23:47:24.885091 kubelet[3220]: I0507 23:47:24.885058 3220 cpu_manager.go:221] "Starting CPU manager" policy="none" May 7 23:47:24.885624 kubelet[3220]: I0507 23:47:24.885555 3220 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 7 23:47:24.885805 kubelet[3220]: I0507 23:47:24.885742 3220 state_mem.go:36] "Initialized new in-memory state store" May 7 23:47:24.886391 kubelet[3220]: I0507 23:47:24.886263 3220 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 7 23:47:24.886391 kubelet[3220]: I0507 23:47:24.886311 3220 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 7 23:47:24.886391 kubelet[3220]: I0507 23:47:24.886348 3220 policy_none.go:49] "None policy: Start" May 7 23:47:24.886677 kubelet[3220]: I0507 23:47:24.886367 3220 memory_manager.go:186] "Starting memorymanager" policy="None" May 7 23:47:24.886677 kubelet[3220]: I0507 23:47:24.886618 3220 state_mem.go:35] "Initializing new in-memory state store" May 7 23:47:24.887256 kubelet[3220]: I0507 23:47:24.887090 3220 state_mem.go:75] "Updated machine memory state" May 7 23:47:24.899184 kubelet[3220]: I0507 23:47:24.897849 3220 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 7 23:47:24.899184 kubelet[3220]: I0507 23:47:24.898149 3220 eviction_manager.go:189] "Eviction manager: starting control loop" May 7 23:47:24.899184 kubelet[3220]: I0507 23:47:24.898170 3220 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 7 23:47:24.899184 kubelet[3220]: I0507 23:47:24.898485 3220 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 7 23:47:24.906234 kubelet[3220]: E0507 23:47:24.906088 3220 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 7 23:47:25.006390 kubelet[3220]: I0507 23:47:25.006326 3220 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:25.007862 kubelet[3220]: I0507 23:47:25.007808 3220 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-28-15" May 7 23:47:25.008386 kubelet[3220]: I0507 23:47:25.008038 3220 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:25.034829 kubelet[3220]: I0507 23:47:25.034094 3220 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-28-15" May 7 23:47:25.052593 kubelet[3220]: I0507 23:47:25.051549 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2a55cfd6270de58396bda8347fda87e4-ca-certs\") pod \"kube-apiserver-ip-172-31-28-15\" (UID: \"2a55cfd6270de58396bda8347fda87e4\") " pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:25.052593 kubelet[3220]: I0507 23:47:25.051608 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2a55cfd6270de58396bda8347fda87e4-k8s-certs\") pod \"kube-apiserver-ip-172-31-28-15\" (UID: \"2a55cfd6270de58396bda8347fda87e4\") " pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:25.052593 kubelet[3220]: I0507 23:47:25.051652 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2a55cfd6270de58396bda8347fda87e4-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-28-15\" (UID: \"2a55cfd6270de58396bda8347fda87e4\") " pod="kube-system/kube-apiserver-ip-172-31-28-15" May 7 23:47:25.052593 kubelet[3220]: I0507 23:47:25.051690 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-k8s-certs\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:25.052593 kubelet[3220]: I0507 23:47:25.051727 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-kubeconfig\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:25.052965 kubelet[3220]: I0507 23:47:25.051765 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:25.052965 kubelet[3220]: I0507 23:47:25.051806 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9873d30601be48fd51168bc4b7f5df81-kubeconfig\") pod \"kube-scheduler-ip-172-31-28-15\" (UID: \"9873d30601be48fd51168bc4b7f5df81\") " pod="kube-system/kube-scheduler-ip-172-31-28-15" May 7 23:47:25.052965 kubelet[3220]: I0507 23:47:25.051841 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-ca-certs\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:25.052965 kubelet[3220]: I0507 23:47:25.051878 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/eb44c7e6e41f00eddc76740dcfa6b38c-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-28-15\" (UID: \"eb44c7e6e41f00eddc76740dcfa6b38c\") " pod="kube-system/kube-controller-manager-ip-172-31-28-15" May 7 23:47:25.054400 kubelet[3220]: I0507 23:47:25.053664 3220 kubelet_node_status.go:125] "Node was previously registered" node="ip-172-31-28-15" May 7 23:47:25.054400 kubelet[3220]: I0507 23:47:25.053921 3220 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-28-15" May 7 23:47:25.532374 sudo[3235]: pam_unix(sudo:session): session closed for user root May 7 23:47:25.576838 kubelet[3220]: I0507 23:47:25.576773 3220 apiserver.go:52] "Watching apiserver" May 7 23:47:25.647445 kubelet[3220]: I0507 23:47:25.647323 3220 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 7 23:47:25.846834 kubelet[3220]: I0507 23:47:25.846670 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-28-15" podStartSLOduration=0.846644596 podStartE2EDuration="846.644596ms" podCreationTimestamp="2025-05-07 23:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-07 23:47:25.831442094 +0000 UTC m=+1.423804493" watchObservedRunningTime="2025-05-07 23:47:25.846644596 +0000 UTC m=+1.439006971" May 7 23:47:25.868687 kubelet[3220]: I0507 23:47:25.867138 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-28-15" podStartSLOduration=0.86709181 podStartE2EDuration="867.09181ms" podCreationTimestamp="2025-05-07 23:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-07 23:47:25.851236206 +0000 UTC m=+1.443598581" watchObservedRunningTime="2025-05-07 23:47:25.86709181 +0000 UTC m=+1.459454197" May 7 23:47:25.886919 kubelet[3220]: I0507 23:47:25.886836 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-28-15" podStartSLOduration=0.88679655 podStartE2EDuration="886.79655ms" podCreationTimestamp="2025-05-07 23:47:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-07 23:47:25.867583768 +0000 UTC m=+1.459946143" watchObservedRunningTime="2025-05-07 23:47:25.88679655 +0000 UTC m=+1.479158913" May 7 23:47:27.897901 sudo[2284]: pam_unix(sudo:session): session closed for user root May 7 23:47:27.921689 sshd[2283]: Connection closed by 147.75.109.163 port 44578 May 7 23:47:27.922670 sshd-session[2281]: pam_unix(sshd:session): session closed for user core May 7 23:47:27.930667 systemd[1]: sshd@6-172.31.28.15:22-147.75.109.163:44578.service: Deactivated successfully. May 7 23:47:27.936035 systemd[1]: session-7.scope: Deactivated successfully. May 7 23:47:27.937476 systemd[1]: session-7.scope: Consumed 9.737s CPU time, 263M memory peak. May 7 23:47:27.941726 systemd-logind[1936]: Session 7 logged out. Waiting for processes to exit. May 7 23:47:27.944053 systemd-logind[1936]: Removed session 7. May 7 23:47:28.967471 kubelet[3220]: I0507 23:47:28.967388 3220 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 7 23:47:28.968896 containerd[1963]: time="2025-05-07T23:47:28.968769908Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 7 23:47:28.969577 kubelet[3220]: I0507 23:47:28.969367 3220 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 7 23:47:29.757719 systemd[1]: Created slice kubepods-besteffort-pod5f722b96_51a7_4b7e_bd42_cd07aceaaf8f.slice - libcontainer container kubepods-besteffort-pod5f722b96_51a7_4b7e_bd42_cd07aceaaf8f.slice. May 7 23:47:29.788778 kubelet[3220]: I0507 23:47:29.788143 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/5f722b96-51a7-4b7e-bd42-cd07aceaaf8f-xtables-lock\") pod \"kube-proxy-rxj22\" (UID: \"5f722b96-51a7-4b7e-bd42-cd07aceaaf8f\") " pod="kube-system/kube-proxy-rxj22" May 7 23:47:29.788778 kubelet[3220]: I0507 23:47:29.788216 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5f722b96-51a7-4b7e-bd42-cd07aceaaf8f-lib-modules\") pod \"kube-proxy-rxj22\" (UID: \"5f722b96-51a7-4b7e-bd42-cd07aceaaf8f\") " pod="kube-system/kube-proxy-rxj22" May 7 23:47:29.788778 kubelet[3220]: I0507 23:47:29.788255 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-746gn\" (UniqueName: \"kubernetes.io/projected/5f722b96-51a7-4b7e-bd42-cd07aceaaf8f-kube-api-access-746gn\") pod \"kube-proxy-rxj22\" (UID: \"5f722b96-51a7-4b7e-bd42-cd07aceaaf8f\") " pod="kube-system/kube-proxy-rxj22" May 7 23:47:29.788778 kubelet[3220]: I0507 23:47:29.788319 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/5f722b96-51a7-4b7e-bd42-cd07aceaaf8f-kube-proxy\") pod \"kube-proxy-rxj22\" (UID: \"5f722b96-51a7-4b7e-bd42-cd07aceaaf8f\") " pod="kube-system/kube-proxy-rxj22" May 7 23:47:29.792800 systemd[1]: Created slice kubepods-burstable-pode62d43b6_116a_4d7f_bd3f_0bbbd8bcaa8c.slice - libcontainer container kubepods-burstable-pode62d43b6_116a_4d7f_bd3f_0bbbd8bcaa8c.slice. May 7 23:47:29.891163 kubelet[3220]: I0507 23:47:29.889189 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8krcx\" (UniqueName: \"kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-kube-api-access-8krcx\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891163 kubelet[3220]: I0507 23:47:29.889267 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hostproc\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891163 kubelet[3220]: I0507 23:47:29.889309 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-etc-cni-netd\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891163 kubelet[3220]: I0507 23:47:29.889344 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-cgroup\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891163 kubelet[3220]: I0507 23:47:29.889380 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-config-path\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891163 kubelet[3220]: I0507 23:47:29.889415 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-net\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891550 kubelet[3220]: I0507 23:47:29.889452 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-bpf-maps\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891550 kubelet[3220]: I0507 23:47:29.889489 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cni-path\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891550 kubelet[3220]: I0507 23:47:29.889522 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hubble-tls\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891550 kubelet[3220]: I0507 23:47:29.889574 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-run\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891550 kubelet[3220]: I0507 23:47:29.889608 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-lib-modules\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891550 kubelet[3220]: I0507 23:47:29.889645 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-xtables-lock\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891851 kubelet[3220]: I0507 23:47:29.889679 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-clustermesh-secrets\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.891851 kubelet[3220]: I0507 23:47:29.889714 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-kernel\") pod \"cilium-j6rx4\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " pod="kube-system/cilium-j6rx4" May 7 23:47:29.983456 systemd[1]: Created slice kubepods-besteffort-pod00cd9891_bcda_4389_9118_a75d01bcd984.slice - libcontainer container kubepods-besteffort-pod00cd9891_bcda_4389_9118_a75d01bcd984.slice. May 7 23:47:29.990513 kubelet[3220]: I0507 23:47:29.990449 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/00cd9891-bcda-4389-9118-a75d01bcd984-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-jrvts\" (UID: \"00cd9891-bcda-4389-9118-a75d01bcd984\") " pod="kube-system/cilium-operator-6c4d7847fc-jrvts" May 7 23:47:29.991044 kubelet[3220]: I0507 23:47:29.990650 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnf2r\" (UniqueName: \"kubernetes.io/projected/00cd9891-bcda-4389-9118-a75d01bcd984-kube-api-access-bnf2r\") pod \"cilium-operator-6c4d7847fc-jrvts\" (UID: \"00cd9891-bcda-4389-9118-a75d01bcd984\") " pod="kube-system/cilium-operator-6c4d7847fc-jrvts" May 7 23:47:30.083216 containerd[1963]: time="2025-05-07T23:47:30.083158564Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-rxj22,Uid:5f722b96-51a7-4b7e-bd42-cd07aceaaf8f,Namespace:kube-system,Attempt:0,}" May 7 23:47:30.104979 containerd[1963]: time="2025-05-07T23:47:30.104361772Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-j6rx4,Uid:e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c,Namespace:kube-system,Attempt:0,}" May 7 23:47:30.170183 containerd[1963]: time="2025-05-07T23:47:30.169755045Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:30.170183 containerd[1963]: time="2025-05-07T23:47:30.169882344Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:30.170183 containerd[1963]: time="2025-05-07T23:47:30.169920848Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:30.171719 containerd[1963]: time="2025-05-07T23:47:30.170845610Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:30.175979 containerd[1963]: time="2025-05-07T23:47:30.175761767Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:30.176278 containerd[1963]: time="2025-05-07T23:47:30.175905779Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:30.176431 containerd[1963]: time="2025-05-07T23:47:30.176241959Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:30.177850 containerd[1963]: time="2025-05-07T23:47:30.177720486Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:30.211450 systemd[1]: Started cri-containerd-13149a66682b67e8f310bb00481200ac171e67284f6748b9573cc260ebb28f47.scope - libcontainer container 13149a66682b67e8f310bb00481200ac171e67284f6748b9573cc260ebb28f47. May 7 23:47:30.220895 systemd[1]: Started cri-containerd-8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a.scope - libcontainer container 8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a. May 7 23:47:30.291140 containerd[1963]: time="2025-05-07T23:47:30.291043704Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-rxj22,Uid:5f722b96-51a7-4b7e-bd42-cd07aceaaf8f,Namespace:kube-system,Attempt:0,} returns sandbox id \"13149a66682b67e8f310bb00481200ac171e67284f6748b9573cc260ebb28f47\"" May 7 23:47:30.292488 containerd[1963]: time="2025-05-07T23:47:30.290894541Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-j6rx4,Uid:e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c,Namespace:kube-system,Attempt:0,} returns sandbox id \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\"" May 7 23:47:30.292488 containerd[1963]: time="2025-05-07T23:47:30.291606929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-jrvts,Uid:00cd9891-bcda-4389-9118-a75d01bcd984,Namespace:kube-system,Attempt:0,}" May 7 23:47:30.300481 containerd[1963]: time="2025-05-07T23:47:30.300310523Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 7 23:47:30.301630 containerd[1963]: time="2025-05-07T23:47:30.301443001Z" level=info msg="CreateContainer within sandbox \"13149a66682b67e8f310bb00481200ac171e67284f6748b9573cc260ebb28f47\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 7 23:47:30.331651 containerd[1963]: time="2025-05-07T23:47:30.331456248Z" level=info msg="CreateContainer within sandbox \"13149a66682b67e8f310bb00481200ac171e67284f6748b9573cc260ebb28f47\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"95567531cfcb732616134375280a05053a588f3133ec32e8d6d37a46145e2a84\"" May 7 23:47:30.334542 containerd[1963]: time="2025-05-07T23:47:30.333968011Z" level=info msg="StartContainer for \"95567531cfcb732616134375280a05053a588f3133ec32e8d6d37a46145e2a84\"" May 7 23:47:30.345489 containerd[1963]: time="2025-05-07T23:47:30.345353505Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:30.345776 containerd[1963]: time="2025-05-07T23:47:30.345608069Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:30.345885 containerd[1963]: time="2025-05-07T23:47:30.345799456Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:30.346589 containerd[1963]: time="2025-05-07T23:47:30.346324179Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:30.383457 systemd[1]: Started cri-containerd-de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b.scope - libcontainer container de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b. May 7 23:47:30.402505 systemd[1]: Started cri-containerd-95567531cfcb732616134375280a05053a588f3133ec32e8d6d37a46145e2a84.scope - libcontainer container 95567531cfcb732616134375280a05053a588f3133ec32e8d6d37a46145e2a84. May 7 23:47:30.480627 containerd[1963]: time="2025-05-07T23:47:30.480576124Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-jrvts,Uid:00cd9891-bcda-4389-9118-a75d01bcd984,Namespace:kube-system,Attempt:0,} returns sandbox id \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\"" May 7 23:47:30.494546 containerd[1963]: time="2025-05-07T23:47:30.494490442Z" level=info msg="StartContainer for \"95567531cfcb732616134375280a05053a588f3133ec32e8d6d37a46145e2a84\" returns successfully" May 7 23:47:32.432512 kubelet[3220]: I0507 23:47:32.432427 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-rxj22" podStartSLOduration=3.432404716 podStartE2EDuration="3.432404716s" podCreationTimestamp="2025-05-07 23:47:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-07 23:47:30.828537044 +0000 UTC m=+6.420899443" watchObservedRunningTime="2025-05-07 23:47:32.432404716 +0000 UTC m=+8.024767091" May 7 23:47:32.642214 update_engine[1939]: I20250507 23:47:32.641603 1939 update_attempter.cc:509] Updating boot flags... May 7 23:47:32.776467 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 42 scanned by (udev-worker) (3598) May 7 23:47:33.210168 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 42 scanned by (udev-worker) (3600) May 7 23:47:37.312158 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3292239268.mount: Deactivated successfully. May 7 23:47:39.999488 containerd[1963]: time="2025-05-07T23:47:39.999241074Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:40.002291 containerd[1963]: time="2025-05-07T23:47:40.002214362Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" May 7 23:47:40.004462 containerd[1963]: time="2025-05-07T23:47:40.004368230Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:40.008747 containerd[1963]: time="2025-05-07T23:47:40.007910618Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 9.707138699s" May 7 23:47:40.008747 containerd[1963]: time="2025-05-07T23:47:40.007969802Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" May 7 23:47:40.011813 containerd[1963]: time="2025-05-07T23:47:40.011635838Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 7 23:47:40.014518 containerd[1963]: time="2025-05-07T23:47:40.014406002Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 7 23:47:40.047500 containerd[1963]: time="2025-05-07T23:47:40.047425190Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\"" May 7 23:47:40.049005 containerd[1963]: time="2025-05-07T23:47:40.048762218Z" level=info msg="StartContainer for \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\"" May 7 23:47:40.105439 systemd[1]: Started cri-containerd-b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436.scope - libcontainer container b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436. May 7 23:47:40.155538 containerd[1963]: time="2025-05-07T23:47:40.155460362Z" level=info msg="StartContainer for \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\" returns successfully" May 7 23:47:40.174080 systemd[1]: cri-containerd-b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436.scope: Deactivated successfully. May 7 23:47:40.916230 containerd[1963]: time="2025-05-07T23:47:40.915680382Z" level=info msg="shim disconnected" id=b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436 namespace=k8s.io May 7 23:47:40.916230 containerd[1963]: time="2025-05-07T23:47:40.915852762Z" level=warning msg="cleaning up after shim disconnected" id=b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436 namespace=k8s.io May 7 23:47:40.916230 containerd[1963]: time="2025-05-07T23:47:40.915872550Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:47:41.038856 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436-rootfs.mount: Deactivated successfully. May 7 23:47:41.878247 containerd[1963]: time="2025-05-07T23:47:41.878172127Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 7 23:47:41.919960 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2177160896.mount: Deactivated successfully. May 7 23:47:41.926159 containerd[1963]: time="2025-05-07T23:47:41.926016103Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\"" May 7 23:47:41.927041 containerd[1963]: time="2025-05-07T23:47:41.926939095Z" level=info msg="StartContainer for \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\"" May 7 23:47:41.986428 systemd[1]: Started cri-containerd-af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e.scope - libcontainer container af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e. May 7 23:47:42.044652 containerd[1963]: time="2025-05-07T23:47:42.044296456Z" level=info msg="StartContainer for \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\" returns successfully" May 7 23:47:42.067267 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 7 23:47:42.069577 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 7 23:47:42.070389 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... May 7 23:47:42.077751 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 7 23:47:42.083429 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 7 23:47:42.084792 systemd[1]: cri-containerd-af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e.scope: Deactivated successfully. May 7 23:47:42.143708 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 7 23:47:42.162070 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e-rootfs.mount: Deactivated successfully. May 7 23:47:42.182964 containerd[1963]: time="2025-05-07T23:47:42.182878228Z" level=info msg="shim disconnected" id=af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e namespace=k8s.io May 7 23:47:42.183237 containerd[1963]: time="2025-05-07T23:47:42.182984956Z" level=warning msg="cleaning up after shim disconnected" id=af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e namespace=k8s.io May 7 23:47:42.183237 containerd[1963]: time="2025-05-07T23:47:42.183006148Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:47:42.465962 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2166513660.mount: Deactivated successfully. May 7 23:47:42.889899 containerd[1963]: time="2025-05-07T23:47:42.889533824Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 7 23:47:42.945612 containerd[1963]: time="2025-05-07T23:47:42.945444860Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\"" May 7 23:47:42.946897 containerd[1963]: time="2025-05-07T23:47:42.946390952Z" level=info msg="StartContainer for \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\"" May 7 23:47:43.120833 systemd[1]: Started cri-containerd-bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6.scope - libcontainer container bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6. May 7 23:47:43.220451 containerd[1963]: time="2025-05-07T23:47:43.219985806Z" level=info msg="StartContainer for \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\" returns successfully" May 7 23:47:43.227472 systemd[1]: cri-containerd-bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6.scope: Deactivated successfully. May 7 23:47:43.292973 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6-rootfs.mount: Deactivated successfully. May 7 23:47:43.364917 containerd[1963]: time="2025-05-07T23:47:43.364794510Z" level=info msg="shim disconnected" id=bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6 namespace=k8s.io May 7 23:47:43.364917 containerd[1963]: time="2025-05-07T23:47:43.364870242Z" level=warning msg="cleaning up after shim disconnected" id=bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6 namespace=k8s.io May 7 23:47:43.364917 containerd[1963]: time="2025-05-07T23:47:43.364890690Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:47:43.501379 containerd[1963]: time="2025-05-07T23:47:43.501218179Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:43.503349 containerd[1963]: time="2025-05-07T23:47:43.503261191Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" May 7 23:47:43.505628 containerd[1963]: time="2025-05-07T23:47:43.505552447Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 7 23:47:43.509500 containerd[1963]: time="2025-05-07T23:47:43.508498795Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 3.496256453s" May 7 23:47:43.509500 containerd[1963]: time="2025-05-07T23:47:43.508556275Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" May 7 23:47:43.517276 containerd[1963]: time="2025-05-07T23:47:43.516903655Z" level=info msg="CreateContainer within sandbox \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 7 23:47:43.549664 containerd[1963]: time="2025-05-07T23:47:43.549612871Z" level=info msg="CreateContainer within sandbox \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\"" May 7 23:47:43.550877 containerd[1963]: time="2025-05-07T23:47:43.550527643Z" level=info msg="StartContainer for \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\"" May 7 23:47:43.596425 systemd[1]: Started cri-containerd-4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7.scope - libcontainer container 4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7. May 7 23:47:43.645931 containerd[1963]: time="2025-05-07T23:47:43.645792728Z" level=info msg="StartContainer for \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\" returns successfully" May 7 23:47:43.898147 containerd[1963]: time="2025-05-07T23:47:43.898042149Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 7 23:47:43.936189 containerd[1963]: time="2025-05-07T23:47:43.935482233Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\"" May 7 23:47:43.937341 containerd[1963]: time="2025-05-07T23:47:43.937238313Z" level=info msg="StartContainer for \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\"" May 7 23:47:44.004573 kubelet[3220]: I0507 23:47:44.004216 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-jrvts" podStartSLOduration=1.9759166320000001 podStartE2EDuration="15.004190453s" podCreationTimestamp="2025-05-07 23:47:29 +0000 UTC" firstStartedPulling="2025-05-07 23:47:30.482796958 +0000 UTC m=+6.075159345" lastFinishedPulling="2025-05-07 23:47:43.511070803 +0000 UTC m=+19.103433166" observedRunningTime="2025-05-07 23:47:44.003931277 +0000 UTC m=+19.596293664" watchObservedRunningTime="2025-05-07 23:47:44.004190453 +0000 UTC m=+19.596552912" May 7 23:47:44.027420 systemd[1]: Started cri-containerd-449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf.scope - libcontainer container 449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf. May 7 23:47:44.159868 containerd[1963]: time="2025-05-07T23:47:44.159694290Z" level=info msg="StartContainer for \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\" returns successfully" May 7 23:47:44.168826 systemd[1]: cri-containerd-449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf.scope: Deactivated successfully. May 7 23:47:44.235388 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf-rootfs.mount: Deactivated successfully. May 7 23:47:44.254789 containerd[1963]: time="2025-05-07T23:47:44.254714671Z" level=info msg="shim disconnected" id=449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf namespace=k8s.io May 7 23:47:44.255490 containerd[1963]: time="2025-05-07T23:47:44.255190303Z" level=warning msg="cleaning up after shim disconnected" id=449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf namespace=k8s.io May 7 23:47:44.255490 containerd[1963]: time="2025-05-07T23:47:44.255221743Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:47:44.290909 containerd[1963]: time="2025-05-07T23:47:44.290565835Z" level=warning msg="cleanup warnings time=\"2025-05-07T23:47:44Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io May 7 23:47:44.915695 containerd[1963]: time="2025-05-07T23:47:44.915632530Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 7 23:47:44.959245 containerd[1963]: time="2025-05-07T23:47:44.959014330Z" level=info msg="CreateContainer within sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\"" May 7 23:47:44.961544 containerd[1963]: time="2025-05-07T23:47:44.959878282Z" level=info msg="StartContainer for \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\"" May 7 23:47:44.967446 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2329809803.mount: Deactivated successfully. May 7 23:47:45.056475 systemd[1]: Started cri-containerd-4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367.scope - libcontainer container 4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367. May 7 23:47:45.207789 containerd[1963]: time="2025-05-07T23:47:45.207645127Z" level=info msg="StartContainer for \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\" returns successfully" May 7 23:47:45.574614 kubelet[3220]: I0507 23:47:45.574542 3220 kubelet_node_status.go:502] "Fast updating node status as it just became ready" May 7 23:47:45.639725 systemd[1]: Created slice kubepods-burstable-pod1e9f0f9d_d75d_4dd1_8624_6e8aa94a8cc6.slice - libcontainer container kubepods-burstable-pod1e9f0f9d_d75d_4dd1_8624_6e8aa94a8cc6.slice. May 7 23:47:45.671187 systemd[1]: Created slice kubepods-burstable-pod1846ca31_b989_4e8c_8209_b60c3b441a96.slice - libcontainer container kubepods-burstable-pod1846ca31_b989_4e8c_8209_b60c3b441a96.slice. May 7 23:47:45.724618 kubelet[3220]: I0507 23:47:45.724543 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1846ca31-b989-4e8c-8209-b60c3b441a96-config-volume\") pod \"coredns-668d6bf9bc-rkdcb\" (UID: \"1846ca31-b989-4e8c-8209-b60c3b441a96\") " pod="kube-system/coredns-668d6bf9bc-rkdcb" May 7 23:47:45.724793 kubelet[3220]: I0507 23:47:45.724627 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68r9p\" (UniqueName: \"kubernetes.io/projected/1e9f0f9d-d75d-4dd1-8624-6e8aa94a8cc6-kube-api-access-68r9p\") pod \"coredns-668d6bf9bc-sq9w9\" (UID: \"1e9f0f9d-d75d-4dd1-8624-6e8aa94a8cc6\") " pod="kube-system/coredns-668d6bf9bc-sq9w9" May 7 23:47:45.724793 kubelet[3220]: I0507 23:47:45.724674 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltttt\" (UniqueName: \"kubernetes.io/projected/1846ca31-b989-4e8c-8209-b60c3b441a96-kube-api-access-ltttt\") pod \"coredns-668d6bf9bc-rkdcb\" (UID: \"1846ca31-b989-4e8c-8209-b60c3b441a96\") " pod="kube-system/coredns-668d6bf9bc-rkdcb" May 7 23:47:45.724793 kubelet[3220]: I0507 23:47:45.724722 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e9f0f9d-d75d-4dd1-8624-6e8aa94a8cc6-config-volume\") pod \"coredns-668d6bf9bc-sq9w9\" (UID: \"1e9f0f9d-d75d-4dd1-8624-6e8aa94a8cc6\") " pod="kube-system/coredns-668d6bf9bc-sq9w9" May 7 23:47:45.988576 containerd[1963]: time="2025-05-07T23:47:45.988414967Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rkdcb,Uid:1846ca31-b989-4e8c-8209-b60c3b441a96,Namespace:kube-system,Attempt:0,}" May 7 23:47:46.265556 containerd[1963]: time="2025-05-07T23:47:46.265392177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-sq9w9,Uid:1e9f0f9d-d75d-4dd1-8624-6e8aa94a8cc6,Namespace:kube-system,Attempt:0,}" May 7 23:47:48.479877 (udev-worker)[4239]: Network interface NamePolicy= disabled on kernel command line. May 7 23:47:48.480781 (udev-worker)[4205]: Network interface NamePolicy= disabled on kernel command line. May 7 23:47:48.490070 systemd-networkd[1866]: cilium_host: Link UP May 7 23:47:48.490495 systemd-networkd[1866]: cilium_net: Link UP May 7 23:47:48.491036 systemd-networkd[1866]: cilium_net: Gained carrier May 7 23:47:48.491450 systemd-networkd[1866]: cilium_host: Gained carrier May 7 23:47:48.491690 systemd-networkd[1866]: cilium_net: Gained IPv6LL May 7 23:47:48.491964 systemd-networkd[1866]: cilium_host: Gained IPv6LL May 7 23:47:48.648999 (udev-worker)[4204]: Network interface NamePolicy= disabled on kernel command line. May 7 23:47:48.660387 systemd-networkd[1866]: cilium_vxlan: Link UP May 7 23:47:48.660401 systemd-networkd[1866]: cilium_vxlan: Gained carrier May 7 23:47:49.163228 kernel: NET: Registered PF_ALG protocol family May 7 23:47:50.496631 systemd-networkd[1866]: cilium_vxlan: Gained IPv6LL May 7 23:47:50.528720 (udev-worker)[4244]: Network interface NamePolicy= disabled on kernel command line. May 7 23:47:50.530175 systemd-networkd[1866]: lxc_health: Link UP May 7 23:47:50.534716 systemd-networkd[1866]: lxc_health: Gained carrier May 7 23:47:50.851972 (udev-worker)[4576]: Network interface NamePolicy= disabled on kernel command line. May 7 23:47:50.856819 systemd-networkd[1866]: lxc232061829539: Link UP May 7 23:47:50.866375 kernel: eth0: renamed from tmp5b134 May 7 23:47:50.869217 systemd-networkd[1866]: lxc232061829539: Gained carrier May 7 23:47:51.109368 systemd-networkd[1866]: lxc9ba94a06241f: Link UP May 7 23:47:51.114371 kernel: eth0: renamed from tmp64b6b May 7 23:47:51.119973 systemd-networkd[1866]: lxc9ba94a06241f: Gained carrier May 7 23:47:52.139229 kubelet[3220]: I0507 23:47:52.139052 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-j6rx4" podStartSLOduration=13.425748605 podStartE2EDuration="23.13902725s" podCreationTimestamp="2025-05-07 23:47:29 +0000 UTC" firstStartedPulling="2025-05-07 23:47:30.296673845 +0000 UTC m=+5.889036220" lastFinishedPulling="2025-05-07 23:47:40.009952478 +0000 UTC m=+15.602314865" observedRunningTime="2025-05-07 23:47:46.078254696 +0000 UTC m=+21.670617155" watchObservedRunningTime="2025-05-07 23:47:52.13902725 +0000 UTC m=+27.731389673" May 7 23:47:52.223411 systemd-networkd[1866]: lxc232061829539: Gained IPv6LL May 7 23:47:52.286377 systemd-networkd[1866]: lxc9ba94a06241f: Gained IPv6LL May 7 23:47:52.542470 systemd-networkd[1866]: lxc_health: Gained IPv6LL May 7 23:47:54.885149 ntpd[1928]: Listen normally on 7 cilium_host 192.168.0.128:123 May 7 23:47:54.885282 ntpd[1928]: Listen normally on 8 cilium_net [fe80::7c24:21ff:fe36:a4aa%4]:123 May 7 23:47:54.885760 ntpd[1928]: 7 May 23:47:54 ntpd[1928]: Listen normally on 7 cilium_host 192.168.0.128:123 May 7 23:47:54.885760 ntpd[1928]: 7 May 23:47:54 ntpd[1928]: Listen normally on 8 cilium_net [fe80::7c24:21ff:fe36:a4aa%4]:123 May 7 23:47:54.885760 ntpd[1928]: 7 May 23:47:54 ntpd[1928]: Listen normally on 9 cilium_host [fe80::1810:b5ff:fe8b:b2e4%5]:123 May 7 23:47:54.885760 ntpd[1928]: 7 May 23:47:54 ntpd[1928]: Listen normally on 10 cilium_vxlan [fe80::c834:9eff:fe80:d4c5%6]:123 May 7 23:47:54.885760 ntpd[1928]: 7 May 23:47:54 ntpd[1928]: Listen normally on 11 lxc_health [fe80::c3:56ff:fec5:d642%8]:123 May 7 23:47:54.885760 ntpd[1928]: 7 May 23:47:54 ntpd[1928]: Listen normally on 12 lxc232061829539 [fe80::14f0:5dff:fe01:a292%10]:123 May 7 23:47:54.885760 ntpd[1928]: 7 May 23:47:54 ntpd[1928]: Listen normally on 13 lxc9ba94a06241f [fe80::18e5:7bff:fe6d:7cd8%12]:123 May 7 23:47:54.885363 ntpd[1928]: Listen normally on 9 cilium_host [fe80::1810:b5ff:fe8b:b2e4%5]:123 May 7 23:47:54.885431 ntpd[1928]: Listen normally on 10 cilium_vxlan [fe80::c834:9eff:fe80:d4c5%6]:123 May 7 23:47:54.885497 ntpd[1928]: Listen normally on 11 lxc_health [fe80::c3:56ff:fec5:d642%8]:123 May 7 23:47:54.885565 ntpd[1928]: Listen normally on 12 lxc232061829539 [fe80::14f0:5dff:fe01:a292%10]:123 May 7 23:47:54.885631 ntpd[1928]: Listen normally on 13 lxc9ba94a06241f [fe80::18e5:7bff:fe6d:7cd8%12]:123 May 7 23:47:56.235943 kubelet[3220]: I0507 23:47:56.234715 3220 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 7 23:47:59.255416 containerd[1963]: time="2025-05-07T23:47:59.253983525Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:59.255416 containerd[1963]: time="2025-05-07T23:47:59.254088897Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:59.255416 containerd[1963]: time="2025-05-07T23:47:59.254163681Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:59.256134 containerd[1963]: time="2025-05-07T23:47:59.255339909Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:59.328402 containerd[1963]: time="2025-05-07T23:47:59.326589238Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:47:59.328402 containerd[1963]: time="2025-05-07T23:47:59.326838142Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:47:59.328402 containerd[1963]: time="2025-05-07T23:47:59.326973598Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:59.328402 containerd[1963]: time="2025-05-07T23:47:59.327446218Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:47:59.332399 systemd[1]: Started cri-containerd-64b6b8e1f9a763844745f0be3ff7448027a8bebcd26b23f6a56f58b62e7e8248.scope - libcontainer container 64b6b8e1f9a763844745f0be3ff7448027a8bebcd26b23f6a56f58b62e7e8248. May 7 23:47:59.391704 systemd[1]: Started cri-containerd-5b134e89cf4451b6fca30f2a730e09875d501f6a2f5ab16385a27c6235fa02a8.scope - libcontainer container 5b134e89cf4451b6fca30f2a730e09875d501f6a2f5ab16385a27c6235fa02a8. May 7 23:47:59.470556 containerd[1963]: time="2025-05-07T23:47:59.469257898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rkdcb,Uid:1846ca31-b989-4e8c-8209-b60c3b441a96,Namespace:kube-system,Attempt:0,} returns sandbox id \"64b6b8e1f9a763844745f0be3ff7448027a8bebcd26b23f6a56f58b62e7e8248\"" May 7 23:47:59.479865 containerd[1963]: time="2025-05-07T23:47:59.479331754Z" level=info msg="CreateContainer within sandbox \"64b6b8e1f9a763844745f0be3ff7448027a8bebcd26b23f6a56f58b62e7e8248\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 7 23:47:59.515236 containerd[1963]: time="2025-05-07T23:47:59.514056826Z" level=info msg="CreateContainer within sandbox \"64b6b8e1f9a763844745f0be3ff7448027a8bebcd26b23f6a56f58b62e7e8248\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"701fafbe8edab1767c59d276fef6ef881780cc175dafc58751515717482bbd71\"" May 7 23:47:59.517426 containerd[1963]: time="2025-05-07T23:47:59.517356766Z" level=info msg="StartContainer for \"701fafbe8edab1767c59d276fef6ef881780cc175dafc58751515717482bbd71\"" May 7 23:47:59.576928 containerd[1963]: time="2025-05-07T23:47:59.576862115Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-sq9w9,Uid:1e9f0f9d-d75d-4dd1-8624-6e8aa94a8cc6,Namespace:kube-system,Attempt:0,} returns sandbox id \"5b134e89cf4451b6fca30f2a730e09875d501f6a2f5ab16385a27c6235fa02a8\"" May 7 23:47:59.607418 containerd[1963]: time="2025-05-07T23:47:59.606806495Z" level=info msg="CreateContainer within sandbox \"5b134e89cf4451b6fca30f2a730e09875d501f6a2f5ab16385a27c6235fa02a8\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 7 23:47:59.608426 systemd[1]: Started cri-containerd-701fafbe8edab1767c59d276fef6ef881780cc175dafc58751515717482bbd71.scope - libcontainer container 701fafbe8edab1767c59d276fef6ef881780cc175dafc58751515717482bbd71. May 7 23:47:59.645704 containerd[1963]: time="2025-05-07T23:47:59.645625391Z" level=info msg="CreateContainer within sandbox \"5b134e89cf4451b6fca30f2a730e09875d501f6a2f5ab16385a27c6235fa02a8\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"16feca8454260568a1c682a8ef249a3c6e5a862ac22fa56751694d83fda89960\"" May 7 23:47:59.649055 containerd[1963]: time="2025-05-07T23:47:59.648974807Z" level=info msg="StartContainer for \"16feca8454260568a1c682a8ef249a3c6e5a862ac22fa56751694d83fda89960\"" May 7 23:47:59.709660 containerd[1963]: time="2025-05-07T23:47:59.709577759Z" level=info msg="StartContainer for \"701fafbe8edab1767c59d276fef6ef881780cc175dafc58751515717482bbd71\" returns successfully" May 7 23:47:59.751475 systemd[1]: Started cri-containerd-16feca8454260568a1c682a8ef249a3c6e5a862ac22fa56751694d83fda89960.scope - libcontainer container 16feca8454260568a1c682a8ef249a3c6e5a862ac22fa56751694d83fda89960. May 7 23:47:59.830405 containerd[1963]: time="2025-05-07T23:47:59.830334396Z" level=info msg="StartContainer for \"16feca8454260568a1c682a8ef249a3c6e5a862ac22fa56751694d83fda89960\" returns successfully" May 7 23:48:00.001185 kubelet[3220]: I0507 23:48:00.000316 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-sq9w9" podStartSLOduration=31.000290505 podStartE2EDuration="31.000290505s" podCreationTimestamp="2025-05-07 23:47:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-07 23:47:59.997255345 +0000 UTC m=+35.589617804" watchObservedRunningTime="2025-05-07 23:48:00.000290505 +0000 UTC m=+35.592652880" May 7 23:48:11.063645 systemd[1]: Started sshd@7-172.31.28.15:22-147.75.109.163:56820.service - OpenSSH per-connection server daemon (147.75.109.163:56820). May 7 23:48:11.264068 sshd[4784]: Accepted publickey for core from 147.75.109.163 port 56820 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:11.266819 sshd-session[4784]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:11.275253 systemd-logind[1936]: New session 8 of user core. May 7 23:48:11.286379 systemd[1]: Started session-8.scope - Session 8 of User core. May 7 23:48:11.548966 sshd[4786]: Connection closed by 147.75.109.163 port 56820 May 7 23:48:11.549891 sshd-session[4784]: pam_unix(sshd:session): session closed for user core May 7 23:48:11.556296 systemd[1]: sshd@7-172.31.28.15:22-147.75.109.163:56820.service: Deactivated successfully. May 7 23:48:11.562804 systemd[1]: session-8.scope: Deactivated successfully. May 7 23:48:11.565534 systemd-logind[1936]: Session 8 logged out. Waiting for processes to exit. May 7 23:48:11.568561 systemd-logind[1936]: Removed session 8. May 7 23:48:16.590715 systemd[1]: Started sshd@8-172.31.28.15:22-147.75.109.163:56828.service - OpenSSH per-connection server daemon (147.75.109.163:56828). May 7 23:48:16.772928 sshd[4799]: Accepted publickey for core from 147.75.109.163 port 56828 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:16.775397 sshd-session[4799]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:16.786290 systemd-logind[1936]: New session 9 of user core. May 7 23:48:16.795370 systemd[1]: Started session-9.scope - Session 9 of User core. May 7 23:48:17.031317 sshd[4801]: Connection closed by 147.75.109.163 port 56828 May 7 23:48:17.032482 sshd-session[4799]: pam_unix(sshd:session): session closed for user core May 7 23:48:17.039547 systemd-logind[1936]: Session 9 logged out. Waiting for processes to exit. May 7 23:48:17.041368 systemd[1]: sshd@8-172.31.28.15:22-147.75.109.163:56828.service: Deactivated successfully. May 7 23:48:17.045887 systemd[1]: session-9.scope: Deactivated successfully. May 7 23:48:17.048043 systemd-logind[1936]: Removed session 9. May 7 23:48:22.074654 systemd[1]: Started sshd@9-172.31.28.15:22-147.75.109.163:54964.service - OpenSSH per-connection server daemon (147.75.109.163:54964). May 7 23:48:22.256940 sshd[4815]: Accepted publickey for core from 147.75.109.163 port 54964 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:22.259412 sshd-session[4815]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:22.268213 systemd-logind[1936]: New session 10 of user core. May 7 23:48:22.278399 systemd[1]: Started session-10.scope - Session 10 of User core. May 7 23:48:22.519808 sshd[4817]: Connection closed by 147.75.109.163 port 54964 May 7 23:48:22.520943 sshd-session[4815]: pam_unix(sshd:session): session closed for user core May 7 23:48:22.527326 systemd[1]: sshd@9-172.31.28.15:22-147.75.109.163:54964.service: Deactivated successfully. May 7 23:48:22.531551 systemd[1]: session-10.scope: Deactivated successfully. May 7 23:48:22.533865 systemd-logind[1936]: Session 10 logged out. Waiting for processes to exit. May 7 23:48:22.535783 systemd-logind[1936]: Removed session 10. May 7 23:48:27.562040 systemd[1]: Started sshd@10-172.31.28.15:22-147.75.109.163:38328.service - OpenSSH per-connection server daemon (147.75.109.163:38328). May 7 23:48:27.750308 sshd[4832]: Accepted publickey for core from 147.75.109.163 port 38328 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:27.752730 sshd-session[4832]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:27.760310 systemd-logind[1936]: New session 11 of user core. May 7 23:48:27.769387 systemd[1]: Started session-11.scope - Session 11 of User core. May 7 23:48:28.012958 sshd[4834]: Connection closed by 147.75.109.163 port 38328 May 7 23:48:28.014157 sshd-session[4832]: pam_unix(sshd:session): session closed for user core May 7 23:48:28.024638 systemd-logind[1936]: Session 11 logged out. Waiting for processes to exit. May 7 23:48:28.027428 systemd[1]: sshd@10-172.31.28.15:22-147.75.109.163:38328.service: Deactivated successfully. May 7 23:48:28.034734 systemd[1]: session-11.scope: Deactivated successfully. May 7 23:48:28.039907 systemd-logind[1936]: Removed session 11. May 7 23:48:33.053709 systemd[1]: Started sshd@11-172.31.28.15:22-147.75.109.163:38334.service - OpenSSH per-connection server daemon (147.75.109.163:38334). May 7 23:48:33.249979 sshd[4851]: Accepted publickey for core from 147.75.109.163 port 38334 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:33.252529 sshd-session[4851]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:33.260277 systemd-logind[1936]: New session 12 of user core. May 7 23:48:33.266382 systemd[1]: Started session-12.scope - Session 12 of User core. May 7 23:48:33.509779 sshd[4853]: Connection closed by 147.75.109.163 port 38334 May 7 23:48:33.510665 sshd-session[4851]: pam_unix(sshd:session): session closed for user core May 7 23:48:33.517419 systemd[1]: sshd@11-172.31.28.15:22-147.75.109.163:38334.service: Deactivated successfully. May 7 23:48:33.522220 systemd[1]: session-12.scope: Deactivated successfully. May 7 23:48:33.525426 systemd-logind[1936]: Session 12 logged out. Waiting for processes to exit. May 7 23:48:33.527346 systemd-logind[1936]: Removed session 12. May 7 23:48:33.549624 systemd[1]: Started sshd@12-172.31.28.15:22-147.75.109.163:38336.service - OpenSSH per-connection server daemon (147.75.109.163:38336). May 7 23:48:33.737333 sshd[4866]: Accepted publickey for core from 147.75.109.163 port 38336 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:33.740057 sshd-session[4866]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:33.749274 systemd-logind[1936]: New session 13 of user core. May 7 23:48:33.763391 systemd[1]: Started session-13.scope - Session 13 of User core. May 7 23:48:34.069903 sshd[4868]: Connection closed by 147.75.109.163 port 38336 May 7 23:48:34.070245 sshd-session[4866]: pam_unix(sshd:session): session closed for user core May 7 23:48:34.082936 systemd[1]: sshd@12-172.31.28.15:22-147.75.109.163:38336.service: Deactivated successfully. May 7 23:48:34.090022 systemd[1]: session-13.scope: Deactivated successfully. May 7 23:48:34.097240 systemd-logind[1936]: Session 13 logged out. Waiting for processes to exit. May 7 23:48:34.119794 systemd[1]: Started sshd@13-172.31.28.15:22-147.75.109.163:38344.service - OpenSSH per-connection server daemon (147.75.109.163:38344). May 7 23:48:34.123581 systemd-logind[1936]: Removed session 13. May 7 23:48:34.317943 sshd[4877]: Accepted publickey for core from 147.75.109.163 port 38344 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:34.320529 sshd-session[4877]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:34.329275 systemd-logind[1936]: New session 14 of user core. May 7 23:48:34.342376 systemd[1]: Started session-14.scope - Session 14 of User core. May 7 23:48:34.595189 sshd[4880]: Connection closed by 147.75.109.163 port 38344 May 7 23:48:34.596041 sshd-session[4877]: pam_unix(sshd:session): session closed for user core May 7 23:48:34.601995 systemd[1]: sshd@13-172.31.28.15:22-147.75.109.163:38344.service: Deactivated successfully. May 7 23:48:34.606955 systemd[1]: session-14.scope: Deactivated successfully. May 7 23:48:34.608734 systemd-logind[1936]: Session 14 logged out. Waiting for processes to exit. May 7 23:48:34.610762 systemd-logind[1936]: Removed session 14. May 7 23:48:39.636625 systemd[1]: Started sshd@14-172.31.28.15:22-147.75.109.163:56588.service - OpenSSH per-connection server daemon (147.75.109.163:56588). May 7 23:48:39.834252 sshd[4893]: Accepted publickey for core from 147.75.109.163 port 56588 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:39.837347 sshd-session[4893]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:39.847809 systemd-logind[1936]: New session 15 of user core. May 7 23:48:39.856350 systemd[1]: Started session-15.scope - Session 15 of User core. May 7 23:48:40.111709 sshd[4895]: Connection closed by 147.75.109.163 port 56588 May 7 23:48:40.110745 sshd-session[4893]: pam_unix(sshd:session): session closed for user core May 7 23:48:40.116529 systemd-logind[1936]: Session 15 logged out. Waiting for processes to exit. May 7 23:48:40.117375 systemd[1]: sshd@14-172.31.28.15:22-147.75.109.163:56588.service: Deactivated successfully. May 7 23:48:40.120857 systemd[1]: session-15.scope: Deactivated successfully. May 7 23:48:40.125491 systemd-logind[1936]: Removed session 15. May 7 23:48:43.264662 systemd[1]: Started sshd@15-172.31.28.15:22-114.251.240.252:58251.service - OpenSSH per-connection server daemon (114.251.240.252:58251). May 7 23:48:45.153645 systemd[1]: Started sshd@16-172.31.28.15:22-147.75.109.163:56590.service - OpenSSH per-connection server daemon (147.75.109.163:56590). May 7 23:48:45.334241 sshd[4909]: Accepted publickey for core from 147.75.109.163 port 56590 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:45.336816 sshd-session[4909]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:45.344849 systemd-logind[1936]: New session 16 of user core. May 7 23:48:45.355372 systemd[1]: Started session-16.scope - Session 16 of User core. May 7 23:48:45.595412 sshd[4911]: Connection closed by 147.75.109.163 port 56590 May 7 23:48:45.596304 sshd-session[4909]: pam_unix(sshd:session): session closed for user core May 7 23:48:45.602653 systemd[1]: sshd@16-172.31.28.15:22-147.75.109.163:56590.service: Deactivated successfully. May 7 23:48:45.608787 systemd[1]: session-16.scope: Deactivated successfully. May 7 23:48:45.612572 systemd-logind[1936]: Session 16 logged out. Waiting for processes to exit. May 7 23:48:45.614640 systemd-logind[1936]: Removed session 16. May 7 23:48:49.611178 sshd[4907]: banner exchange: Connection from 114.251.240.252 port 58251: invalid format May 7 23:48:49.612151 systemd[1]: sshd@15-172.31.28.15:22-114.251.240.252:58251.service: Deactivated successfully. May 7 23:48:50.638641 systemd[1]: Started sshd@17-172.31.28.15:22-147.75.109.163:55014.service - OpenSSH per-connection server daemon (147.75.109.163:55014). May 7 23:48:50.833003 sshd[4925]: Accepted publickey for core from 147.75.109.163 port 55014 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:50.835094 sshd-session[4925]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:50.848213 systemd-logind[1936]: New session 17 of user core. May 7 23:48:50.855433 systemd[1]: Started session-17.scope - Session 17 of User core. May 7 23:48:51.101630 sshd[4927]: Connection closed by 147.75.109.163 port 55014 May 7 23:48:51.102717 sshd-session[4925]: pam_unix(sshd:session): session closed for user core May 7 23:48:51.109340 systemd-logind[1936]: Session 17 logged out. Waiting for processes to exit. May 7 23:48:51.110457 systemd[1]: sshd@17-172.31.28.15:22-147.75.109.163:55014.service: Deactivated successfully. May 7 23:48:51.115356 systemd[1]: session-17.scope: Deactivated successfully. May 7 23:48:51.117937 systemd-logind[1936]: Removed session 17. May 7 23:48:51.144375 systemd[1]: Started sshd@18-172.31.28.15:22-147.75.109.163:55022.service - OpenSSH per-connection server daemon (147.75.109.163:55022). May 7 23:48:51.337922 sshd[4939]: Accepted publickey for core from 147.75.109.163 port 55022 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:51.340442 sshd-session[4939]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:51.351030 systemd-logind[1936]: New session 18 of user core. May 7 23:48:51.358394 systemd[1]: Started session-18.scope - Session 18 of User core. May 7 23:48:51.650270 sshd[4941]: Connection closed by 147.75.109.163 port 55022 May 7 23:48:51.651031 sshd-session[4939]: pam_unix(sshd:session): session closed for user core May 7 23:48:51.657567 systemd[1]: sshd@18-172.31.28.15:22-147.75.109.163:55022.service: Deactivated successfully. May 7 23:48:51.662646 systemd[1]: session-18.scope: Deactivated successfully. May 7 23:48:51.664922 systemd-logind[1936]: Session 18 logged out. Waiting for processes to exit. May 7 23:48:51.667347 systemd-logind[1936]: Removed session 18. May 7 23:48:51.692864 systemd[1]: Started sshd@19-172.31.28.15:22-147.75.109.163:55036.service - OpenSSH per-connection server daemon (147.75.109.163:55036). May 7 23:48:51.885067 sshd[4951]: Accepted publickey for core from 147.75.109.163 port 55036 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:51.887528 sshd-session[4951]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:51.896458 systemd-logind[1936]: New session 19 of user core. May 7 23:48:51.915498 systemd[1]: Started session-19.scope - Session 19 of User core. May 7 23:48:53.074389 sshd[4953]: Connection closed by 147.75.109.163 port 55036 May 7 23:48:53.075249 sshd-session[4951]: pam_unix(sshd:session): session closed for user core May 7 23:48:53.088972 systemd[1]: sshd@19-172.31.28.15:22-147.75.109.163:55036.service: Deactivated successfully. May 7 23:48:53.098987 systemd[1]: session-19.scope: Deactivated successfully. May 7 23:48:53.110270 systemd-logind[1936]: Session 19 logged out. Waiting for processes to exit. May 7 23:48:53.140711 systemd[1]: Started sshd@20-172.31.28.15:22-147.75.109.163:55050.service - OpenSSH per-connection server daemon (147.75.109.163:55050). May 7 23:48:53.141609 systemd-logind[1936]: Removed session 19. May 7 23:48:53.338336 sshd[4970]: Accepted publickey for core from 147.75.109.163 port 55050 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:53.340415 sshd-session[4970]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:53.349890 systemd-logind[1936]: New session 20 of user core. May 7 23:48:53.355372 systemd[1]: Started session-20.scope - Session 20 of User core. May 7 23:48:53.841934 sshd[4973]: Connection closed by 147.75.109.163 port 55050 May 7 23:48:53.844043 sshd-session[4970]: pam_unix(sshd:session): session closed for user core May 7 23:48:53.850042 systemd[1]: sshd@20-172.31.28.15:22-147.75.109.163:55050.service: Deactivated successfully. May 7 23:48:53.857308 systemd[1]: session-20.scope: Deactivated successfully. May 7 23:48:53.860362 systemd-logind[1936]: Session 20 logged out. Waiting for processes to exit. May 7 23:48:53.862611 systemd-logind[1936]: Removed session 20. May 7 23:48:53.884749 systemd[1]: Started sshd@21-172.31.28.15:22-147.75.109.163:55062.service - OpenSSH per-connection server daemon (147.75.109.163:55062). May 7 23:48:54.071056 sshd[4983]: Accepted publickey for core from 147.75.109.163 port 55062 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:54.073578 sshd-session[4983]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:54.082663 systemd-logind[1936]: New session 21 of user core. May 7 23:48:54.101375 systemd[1]: Started session-21.scope - Session 21 of User core. May 7 23:48:54.338966 sshd[4985]: Connection closed by 147.75.109.163 port 55062 May 7 23:48:54.339854 sshd-session[4983]: pam_unix(sshd:session): session closed for user core May 7 23:48:54.346100 systemd-logind[1936]: Session 21 logged out. Waiting for processes to exit. May 7 23:48:54.347374 systemd[1]: sshd@21-172.31.28.15:22-147.75.109.163:55062.service: Deactivated successfully. May 7 23:48:54.350921 systemd[1]: session-21.scope: Deactivated successfully. May 7 23:48:54.354787 systemd-logind[1936]: Removed session 21. May 7 23:48:57.337654 systemd[1]: Started sshd@22-172.31.28.15:22-114.251.240.252:34282.service - OpenSSH per-connection server daemon (114.251.240.252:34282). May 7 23:48:59.233008 sshd[4996]: Invalid user wqmarlduiqkmgs from 114.251.240.252 port 34282 May 7 23:48:59.235295 sshd[4996]: userauth_pubkey: parse publickey packet: incomplete message [preauth] May 7 23:48:59.240933 systemd[1]: sshd@22-172.31.28.15:22-114.251.240.252:34282.service: Deactivated successfully. May 7 23:48:59.387631 systemd[1]: Started sshd@23-172.31.28.15:22-147.75.109.163:54742.service - OpenSSH per-connection server daemon (147.75.109.163:54742). May 7 23:48:59.574330 sshd[5001]: Accepted publickey for core from 147.75.109.163 port 54742 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:48:59.576245 sshd-session[5001]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:48:59.588435 systemd-logind[1936]: New session 22 of user core. May 7 23:48:59.606380 systemd[1]: Started session-22.scope - Session 22 of User core. May 7 23:48:59.866583 sshd[5003]: Connection closed by 147.75.109.163 port 54742 May 7 23:48:59.866248 sshd-session[5001]: pam_unix(sshd:session): session closed for user core May 7 23:48:59.876855 systemd-logind[1936]: Session 22 logged out. Waiting for processes to exit. May 7 23:48:59.879240 systemd[1]: sshd@23-172.31.28.15:22-147.75.109.163:54742.service: Deactivated successfully. May 7 23:48:59.885285 systemd[1]: session-22.scope: Deactivated successfully. May 7 23:48:59.887027 systemd-logind[1936]: Removed session 22. May 7 23:49:04.907656 systemd[1]: Started sshd@24-172.31.28.15:22-147.75.109.163:54744.service - OpenSSH per-connection server daemon (147.75.109.163:54744). May 7 23:49:05.101159 sshd[5018]: Accepted publickey for core from 147.75.109.163 port 54744 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:49:05.103736 sshd-session[5018]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:49:05.113082 systemd-logind[1936]: New session 23 of user core. May 7 23:49:05.120390 systemd[1]: Started session-23.scope - Session 23 of User core. May 7 23:49:05.363741 sshd[5020]: Connection closed by 147.75.109.163 port 54744 May 7 23:49:05.364592 sshd-session[5018]: pam_unix(sshd:session): session closed for user core May 7 23:49:05.370810 systemd-logind[1936]: Session 23 logged out. Waiting for processes to exit. May 7 23:49:05.372631 systemd[1]: sshd@24-172.31.28.15:22-147.75.109.163:54744.service: Deactivated successfully. May 7 23:49:05.378759 systemd[1]: session-23.scope: Deactivated successfully. May 7 23:49:05.381485 systemd-logind[1936]: Removed session 23. May 7 23:49:10.409106 systemd[1]: Started sshd@25-172.31.28.15:22-147.75.109.163:37850.service - OpenSSH per-connection server daemon (147.75.109.163:37850). May 7 23:49:10.600293 sshd[5032]: Accepted publickey for core from 147.75.109.163 port 37850 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:49:10.602853 sshd-session[5032]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:49:10.611482 systemd-logind[1936]: New session 24 of user core. May 7 23:49:10.619380 systemd[1]: Started session-24.scope - Session 24 of User core. May 7 23:49:10.871925 sshd[5034]: Connection closed by 147.75.109.163 port 37850 May 7 23:49:10.872856 sshd-session[5032]: pam_unix(sshd:session): session closed for user core May 7 23:49:10.879771 systemd[1]: sshd@25-172.31.28.15:22-147.75.109.163:37850.service: Deactivated successfully. May 7 23:49:10.883857 systemd[1]: session-24.scope: Deactivated successfully. May 7 23:49:10.886549 systemd-logind[1936]: Session 24 logged out. Waiting for processes to exit. May 7 23:49:10.889569 systemd-logind[1936]: Removed session 24. May 7 23:49:15.915665 systemd[1]: Started sshd@26-172.31.28.15:22-147.75.109.163:37858.service - OpenSSH per-connection server daemon (147.75.109.163:37858). May 7 23:49:16.102372 sshd[5046]: Accepted publickey for core from 147.75.109.163 port 37858 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:49:16.104967 sshd-session[5046]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:49:16.113356 systemd-logind[1936]: New session 25 of user core. May 7 23:49:16.119376 systemd[1]: Started session-25.scope - Session 25 of User core. May 7 23:49:16.363007 sshd[5048]: Connection closed by 147.75.109.163 port 37858 May 7 23:49:16.363977 sshd-session[5046]: pam_unix(sshd:session): session closed for user core May 7 23:49:16.371209 systemd[1]: sshd@26-172.31.28.15:22-147.75.109.163:37858.service: Deactivated successfully. May 7 23:49:16.375908 systemd[1]: session-25.scope: Deactivated successfully. May 7 23:49:16.378102 systemd-logind[1936]: Session 25 logged out. Waiting for processes to exit. May 7 23:49:16.380701 systemd-logind[1936]: Removed session 25. May 7 23:49:16.408589 systemd[1]: Started sshd@27-172.31.28.15:22-147.75.109.163:37862.service - OpenSSH per-connection server daemon (147.75.109.163:37862). May 7 23:49:16.589804 sshd[5060]: Accepted publickey for core from 147.75.109.163 port 37862 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:49:16.592350 sshd-session[5060]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:49:16.601384 systemd-logind[1936]: New session 26 of user core. May 7 23:49:16.616418 systemd[1]: Started session-26.scope - Session 26 of User core. May 7 23:49:20.018879 kubelet[3220]: I0507 23:49:20.018752 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-rkdcb" podStartSLOduration=111.018725078 podStartE2EDuration="1m51.018725078s" podCreationTimestamp="2025-05-07 23:47:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-07 23:48:00.054878037 +0000 UTC m=+35.647240424" watchObservedRunningTime="2025-05-07 23:49:20.018725078 +0000 UTC m=+115.611087477" May 7 23:49:20.047161 containerd[1963]: time="2025-05-07T23:49:20.046370942Z" level=info msg="StopContainer for \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\" with timeout 30 (s)" May 7 23:49:20.050216 containerd[1963]: time="2025-05-07T23:49:20.048128462Z" level=info msg="Stop container \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\" with signal terminated" May 7 23:49:20.083909 systemd[1]: cri-containerd-4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7.scope: Deactivated successfully. May 7 23:49:20.104475 containerd[1963]: time="2025-05-07T23:49:20.103409523Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 7 23:49:20.121987 containerd[1963]: time="2025-05-07T23:49:20.121900995Z" level=info msg="StopContainer for \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\" with timeout 2 (s)" May 7 23:49:20.123431 containerd[1963]: time="2025-05-07T23:49:20.123291807Z" level=info msg="Stop container \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\" with signal terminated" May 7 23:49:20.139645 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7-rootfs.mount: Deactivated successfully. May 7 23:49:20.145301 systemd-networkd[1866]: lxc_health: Link DOWN May 7 23:49:20.145322 systemd-networkd[1866]: lxc_health: Lost carrier May 7 23:49:20.171242 containerd[1963]: time="2025-05-07T23:49:20.170940987Z" level=info msg="shim disconnected" id=4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7 namespace=k8s.io May 7 23:49:20.171242 containerd[1963]: time="2025-05-07T23:49:20.171019659Z" level=warning msg="cleaning up after shim disconnected" id=4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7 namespace=k8s.io May 7 23:49:20.171242 containerd[1963]: time="2025-05-07T23:49:20.171043755Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:20.184649 systemd[1]: cri-containerd-4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367.scope: Deactivated successfully. May 7 23:49:20.187226 systemd[1]: cri-containerd-4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367.scope: Consumed 14.470s CPU time, 128.3M memory peak, 136K read from disk, 12.9M written to disk. May 7 23:49:20.215840 containerd[1963]: time="2025-05-07T23:49:20.215638395Z" level=info msg="StopContainer for \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\" returns successfully" May 7 23:49:20.217906 containerd[1963]: time="2025-05-07T23:49:20.217654827Z" level=info msg="StopPodSandbox for \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\"" May 7 23:49:20.217906 containerd[1963]: time="2025-05-07T23:49:20.217735575Z" level=info msg="Container to stop \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 7 23:49:20.222901 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b-shm.mount: Deactivated successfully. May 7 23:49:20.239583 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367-rootfs.mount: Deactivated successfully. May 7 23:49:20.244772 systemd[1]: cri-containerd-de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b.scope: Deactivated successfully. May 7 23:49:20.248994 containerd[1963]: time="2025-05-07T23:49:20.248703627Z" level=info msg="shim disconnected" id=4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367 namespace=k8s.io May 7 23:49:20.248994 containerd[1963]: time="2025-05-07T23:49:20.248781747Z" level=warning msg="cleaning up after shim disconnected" id=4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367 namespace=k8s.io May 7 23:49:20.248994 containerd[1963]: time="2025-05-07T23:49:20.248801667Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:20.283151 containerd[1963]: time="2025-05-07T23:49:20.282595360Z" level=info msg="StopContainer for \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\" returns successfully" May 7 23:49:20.283986 containerd[1963]: time="2025-05-07T23:49:20.283850320Z" level=info msg="StopPodSandbox for \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\"" May 7 23:49:20.283986 containerd[1963]: time="2025-05-07T23:49:20.283936360Z" level=info msg="Container to stop \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 7 23:49:20.284729 containerd[1963]: time="2025-05-07T23:49:20.284655208Z" level=info msg="Container to stop \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 7 23:49:20.285591 containerd[1963]: time="2025-05-07T23:49:20.284869384Z" level=info msg="Container to stop \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 7 23:49:20.285935 containerd[1963]: time="2025-05-07T23:49:20.285882328Z" level=info msg="Container to stop \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 7 23:49:20.286087 containerd[1963]: time="2025-05-07T23:49:20.286057768Z" level=info msg="Container to stop \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 7 23:49:20.302698 systemd[1]: cri-containerd-8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a.scope: Deactivated successfully. May 7 23:49:20.310894 containerd[1963]: time="2025-05-07T23:49:20.310406452Z" level=info msg="shim disconnected" id=de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b namespace=k8s.io May 7 23:49:20.310894 containerd[1963]: time="2025-05-07T23:49:20.310636468Z" level=warning msg="cleaning up after shim disconnected" id=de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b namespace=k8s.io May 7 23:49:20.310894 containerd[1963]: time="2025-05-07T23:49:20.310775392Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:20.340475 containerd[1963]: time="2025-05-07T23:49:20.339953800Z" level=info msg="TearDown network for sandbox \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" successfully" May 7 23:49:20.340475 containerd[1963]: time="2025-05-07T23:49:20.340015360Z" level=info msg="StopPodSandbox for \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" returns successfully" May 7 23:49:20.359802 containerd[1963]: time="2025-05-07T23:49:20.357277144Z" level=info msg="shim disconnected" id=8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a namespace=k8s.io May 7 23:49:20.359802 containerd[1963]: time="2025-05-07T23:49:20.357647632Z" level=warning msg="cleaning up after shim disconnected" id=8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a namespace=k8s.io May 7 23:49:20.359802 containerd[1963]: time="2025-05-07T23:49:20.357673516Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:20.385802 containerd[1963]: time="2025-05-07T23:49:20.385743388Z" level=warning msg="cleanup warnings time=\"2025-05-07T23:49:20Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io May 7 23:49:20.388409 containerd[1963]: time="2025-05-07T23:49:20.388339432Z" level=info msg="TearDown network for sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" successfully" May 7 23:49:20.388409 containerd[1963]: time="2025-05-07T23:49:20.388407976Z" level=info msg="StopPodSandbox for \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" returns successfully" May 7 23:49:20.443144 kubelet[3220]: I0507 23:49:20.443073 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/00cd9891-bcda-4389-9118-a75d01bcd984-cilium-config-path\") pod \"00cd9891-bcda-4389-9118-a75d01bcd984\" (UID: \"00cd9891-bcda-4389-9118-a75d01bcd984\") " May 7 23:49:20.443343 kubelet[3220]: I0507 23:49:20.443173 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnf2r\" (UniqueName: \"kubernetes.io/projected/00cd9891-bcda-4389-9118-a75d01bcd984-kube-api-access-bnf2r\") pod \"00cd9891-bcda-4389-9118-a75d01bcd984\" (UID: \"00cd9891-bcda-4389-9118-a75d01bcd984\") " May 7 23:49:20.447799 kubelet[3220]: I0507 23:49:20.447739 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00cd9891-bcda-4389-9118-a75d01bcd984-kube-api-access-bnf2r" (OuterVolumeSpecName: "kube-api-access-bnf2r") pod "00cd9891-bcda-4389-9118-a75d01bcd984" (UID: "00cd9891-bcda-4389-9118-a75d01bcd984"). InnerVolumeSpecName "kube-api-access-bnf2r". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 7 23:49:20.450789 kubelet[3220]: I0507 23:49:20.450747 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00cd9891-bcda-4389-9118-a75d01bcd984-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "00cd9891-bcda-4389-9118-a75d01bcd984" (UID: "00cd9891-bcda-4389-9118-a75d01bcd984"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 7 23:49:20.543833 kubelet[3220]: I0507 23:49:20.543699 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cni-path\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.544477 kubelet[3220]: I0507 23:49:20.544449 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hubble-tls\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.545134 kubelet[3220]: I0507 23:49:20.544870 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-kernel\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.545134 kubelet[3220]: I0507 23:49:20.544913 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-bpf-maps\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.546866 kubelet[3220]: I0507 23:49:20.545534 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-etc-cni-netd\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.546866 kubelet[3220]: I0507 23:49:20.545589 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-cgroup\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.546866 kubelet[3220]: I0507 23:49:20.545640 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-config-path\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.546866 kubelet[3220]: I0507 23:49:20.545682 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8krcx\" (UniqueName: \"kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-kube-api-access-8krcx\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.546866 kubelet[3220]: I0507 23:49:20.545726 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hostproc\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.546866 kubelet[3220]: I0507 23:49:20.545760 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-net\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.547321 kubelet[3220]: I0507 23:49:20.545792 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-run\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.547321 kubelet[3220]: I0507 23:49:20.545832 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-clustermesh-secrets\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.547321 kubelet[3220]: I0507 23:49:20.545871 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-lib-modules\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.547321 kubelet[3220]: I0507 23:49:20.545907 3220 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-xtables-lock\") pod \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\" (UID: \"e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c\") " May 7 23:49:20.547321 kubelet[3220]: I0507 23:49:20.545982 3220 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/00cd9891-bcda-4389-9118-a75d01bcd984-cilium-config-path\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.547321 kubelet[3220]: I0507 23:49:20.546006 3220 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bnf2r\" (UniqueName: \"kubernetes.io/projected/00cd9891-bcda-4389-9118-a75d01bcd984-kube-api-access-bnf2r\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.558151 kubelet[3220]: I0507 23:49:20.544212 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cni-path" (OuterVolumeSpecName: "cni-path") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.558151 kubelet[3220]: I0507 23:49:20.546055 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.558151 kubelet[3220]: I0507 23:49:20.546083 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.558151 kubelet[3220]: I0507 23:49:20.549049 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.558151 kubelet[3220]: I0507 23:49:20.549147 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.558526 kubelet[3220]: I0507 23:49:20.549176 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.558526 kubelet[3220]: I0507 23:49:20.552994 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 7 23:49:20.558526 kubelet[3220]: I0507 23:49:20.553223 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.561155 kubelet[3220]: I0507 23:49:20.558950 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-kube-api-access-8krcx" (OuterVolumeSpecName: "kube-api-access-8krcx") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "kube-api-access-8krcx". PluginName "kubernetes.io/projected", VolumeGIDValue "" May 7 23:49:20.561155 kubelet[3220]: I0507 23:49:20.559041 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hostproc" (OuterVolumeSpecName: "hostproc") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.561423 kubelet[3220]: I0507 23:49:20.561361 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" May 7 23:49:20.561506 kubelet[3220]: I0507 23:49:20.561457 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.561561 kubelet[3220]: I0507 23:49:20.561501 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" May 7 23:49:20.565566 kubelet[3220]: I0507 23:49:20.565467 3220 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" (UID: "e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" May 7 23:49:20.647257 kubelet[3220]: I0507 23:49:20.647196 3220 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cni-path\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647257 kubelet[3220]: I0507 23:49:20.647248 3220 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hubble-tls\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647273 3220 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-bpf-maps\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647299 3220 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-kernel\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647322 3220 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8krcx\" (UniqueName: \"kubernetes.io/projected/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-kube-api-access-8krcx\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647343 3220 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-etc-cni-netd\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647367 3220 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-cgroup\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647388 3220 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-config-path\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647410 3220 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-cilium-run\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647497 kubelet[3220]: I0507 23:49:20.647434 3220 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-clustermesh-secrets\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647898 kubelet[3220]: I0507 23:49:20.647455 3220 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-hostproc\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647898 kubelet[3220]: I0507 23:49:20.647496 3220 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-host-proc-sys-net\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647898 kubelet[3220]: I0507 23:49:20.647523 3220 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-lib-modules\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.647898 kubelet[3220]: I0507 23:49:20.647543 3220 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c-xtables-lock\") on node \"ip-172-31-28-15\" DevicePath \"\"" May 7 23:49:20.726638 systemd[1]: Removed slice kubepods-burstable-pode62d43b6_116a_4d7f_bd3f_0bbbd8bcaa8c.slice - libcontainer container kubepods-burstable-pode62d43b6_116a_4d7f_bd3f_0bbbd8bcaa8c.slice. May 7 23:49:20.727136 systemd[1]: kubepods-burstable-pode62d43b6_116a_4d7f_bd3f_0bbbd8bcaa8c.slice: Consumed 14.630s CPU time, 128.8M memory peak, 136K read from disk, 12.9M written to disk. May 7 23:49:20.730874 systemd[1]: Removed slice kubepods-besteffort-pod00cd9891_bcda_4389_9118_a75d01bcd984.slice - libcontainer container kubepods-besteffort-pod00cd9891_bcda_4389_9118_a75d01bcd984.slice. May 7 23:49:21.063711 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b-rootfs.mount: Deactivated successfully. May 7 23:49:21.063882 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a-rootfs.mount: Deactivated successfully. May 7 23:49:21.064016 systemd[1]: var-lib-kubelet-pods-00cd9891\x2dbcda\x2d4389\x2d9118\x2da75d01bcd984-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dbnf2r.mount: Deactivated successfully. May 7 23:49:21.064567 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a-shm.mount: Deactivated successfully. May 7 23:49:21.064891 systemd[1]: var-lib-kubelet-pods-e62d43b6\x2d116a\x2d4d7f\x2dbd3f\x2d0bbbd8bcaa8c-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d8krcx.mount: Deactivated successfully. May 7 23:49:21.065033 systemd[1]: var-lib-kubelet-pods-e62d43b6\x2d116a\x2d4d7f\x2dbd3f\x2d0bbbd8bcaa8c-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 7 23:49:21.065216 systemd[1]: var-lib-kubelet-pods-e62d43b6\x2d116a\x2d4d7f\x2dbd3f\x2d0bbbd8bcaa8c-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 7 23:49:21.195243 kubelet[3220]: I0507 23:49:21.195196 3220 scope.go:117] "RemoveContainer" containerID="4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7" May 7 23:49:21.204317 containerd[1963]: time="2025-05-07T23:49:21.202239988Z" level=info msg="RemoveContainer for \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\"" May 7 23:49:21.216311 containerd[1963]: time="2025-05-07T23:49:21.216094276Z" level=info msg="RemoveContainer for \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\" returns successfully" May 7 23:49:21.216855 kubelet[3220]: I0507 23:49:21.216715 3220 scope.go:117] "RemoveContainer" containerID="4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7" May 7 23:49:21.217286 containerd[1963]: time="2025-05-07T23:49:21.217089700Z" level=error msg="ContainerStatus for \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\": not found" May 7 23:49:21.219135 kubelet[3220]: E0507 23:49:21.217645 3220 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\": not found" containerID="4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7" May 7 23:49:21.219302 kubelet[3220]: I0507 23:49:21.219058 3220 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7"} err="failed to get container status \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\": rpc error: code = NotFound desc = an error occurred when try to find container \"4547545513f22a3b09aad84cbaa407ceda0068f40a1f5c6bafd3bbfd2fdaebf7\": not found" May 7 23:49:21.219302 kubelet[3220]: I0507 23:49:21.219207 3220 scope.go:117] "RemoveContainer" containerID="4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367" May 7 23:49:21.226060 containerd[1963]: time="2025-05-07T23:49:21.225739576Z" level=info msg="RemoveContainer for \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\"" May 7 23:49:21.233051 containerd[1963]: time="2025-05-07T23:49:21.232999240Z" level=info msg="RemoveContainer for \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\" returns successfully" May 7 23:49:21.235138 kubelet[3220]: I0507 23:49:21.234158 3220 scope.go:117] "RemoveContainer" containerID="449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf" May 7 23:49:21.238457 containerd[1963]: time="2025-05-07T23:49:21.238381216Z" level=info msg="RemoveContainer for \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\"" May 7 23:49:21.247303 containerd[1963]: time="2025-05-07T23:49:21.247229548Z" level=info msg="RemoveContainer for \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\" returns successfully" May 7 23:49:21.247647 kubelet[3220]: I0507 23:49:21.247580 3220 scope.go:117] "RemoveContainer" containerID="bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6" May 7 23:49:21.251165 containerd[1963]: time="2025-05-07T23:49:21.250173904Z" level=info msg="RemoveContainer for \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\"" May 7 23:49:21.257337 containerd[1963]: time="2025-05-07T23:49:21.257279764Z" level=info msg="RemoveContainer for \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\" returns successfully" May 7 23:49:21.257859 kubelet[3220]: I0507 23:49:21.257621 3220 scope.go:117] "RemoveContainer" containerID="af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e" May 7 23:49:21.260004 containerd[1963]: time="2025-05-07T23:49:21.259820920Z" level=info msg="RemoveContainer for \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\"" May 7 23:49:21.266478 containerd[1963]: time="2025-05-07T23:49:21.266407901Z" level=info msg="RemoveContainer for \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\" returns successfully" May 7 23:49:21.266906 kubelet[3220]: I0507 23:49:21.266826 3220 scope.go:117] "RemoveContainer" containerID="b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436" May 7 23:49:21.269172 containerd[1963]: time="2025-05-07T23:49:21.268868525Z" level=info msg="RemoveContainer for \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\"" May 7 23:49:21.275342 containerd[1963]: time="2025-05-07T23:49:21.275295125Z" level=info msg="RemoveContainer for \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\" returns successfully" May 7 23:49:21.276039 kubelet[3220]: I0507 23:49:21.275787 3220 scope.go:117] "RemoveContainer" containerID="4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367" May 7 23:49:21.276227 containerd[1963]: time="2025-05-07T23:49:21.276170681Z" level=error msg="ContainerStatus for \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\": not found" May 7 23:49:21.276662 kubelet[3220]: E0507 23:49:21.276471 3220 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\": not found" containerID="4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367" May 7 23:49:21.276662 kubelet[3220]: I0507 23:49:21.276517 3220 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367"} err="failed to get container status \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\": rpc error: code = NotFound desc = an error occurred when try to find container \"4a67d371479618bc80e59eaf8c78dd3dadcfc6d6d9354370268d059ece88c367\": not found" May 7 23:49:21.276662 kubelet[3220]: I0507 23:49:21.276552 3220 scope.go:117] "RemoveContainer" containerID="449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf" May 7 23:49:21.276915 containerd[1963]: time="2025-05-07T23:49:21.276828869Z" level=error msg="ContainerStatus for \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\": not found" May 7 23:49:21.277361 kubelet[3220]: E0507 23:49:21.277281 3220 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\": not found" containerID="449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf" May 7 23:49:21.277525 kubelet[3220]: I0507 23:49:21.277350 3220 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf"} err="failed to get container status \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\": rpc error: code = NotFound desc = an error occurred when try to find container \"449c8fb7e9e2c4b2df93c3f7060f9153a77540b21871da5d44cca52aada2bfbf\": not found" May 7 23:49:21.277525 kubelet[3220]: I0507 23:49:21.277384 3220 scope.go:117] "RemoveContainer" containerID="bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6" May 7 23:49:21.277918 containerd[1963]: time="2025-05-07T23:49:21.277838537Z" level=error msg="ContainerStatus for \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\": not found" May 7 23:49:21.278315 kubelet[3220]: E0507 23:49:21.278088 3220 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\": not found" containerID="bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6" May 7 23:49:21.278315 kubelet[3220]: I0507 23:49:21.278152 3220 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6"} err="failed to get container status \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\": rpc error: code = NotFound desc = an error occurred when try to find container \"bec0f05f94ce0e747d0f974f940a773586acda00ccf4f6149a7270cefdb07ea6\": not found" May 7 23:49:21.278315 kubelet[3220]: I0507 23:49:21.278185 3220 scope.go:117] "RemoveContainer" containerID="af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e" May 7 23:49:21.278808 containerd[1963]: time="2025-05-07T23:49:21.278695865Z" level=error msg="ContainerStatus for \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\": not found" May 7 23:49:21.279002 kubelet[3220]: E0507 23:49:21.278948 3220 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\": not found" containerID="af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e" May 7 23:49:21.279081 kubelet[3220]: I0507 23:49:21.279014 3220 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e"} err="failed to get container status \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\": rpc error: code = NotFound desc = an error occurred when try to find container \"af08f508dfc0aaeb898b71e45b4273aaa2ab8fdefab77a2ebd1369cbc9c7634e\": not found" May 7 23:49:21.279081 kubelet[3220]: I0507 23:49:21.279049 3220 scope.go:117] "RemoveContainer" containerID="b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436" May 7 23:49:21.280016 containerd[1963]: time="2025-05-07T23:49:21.279941705Z" level=error msg="ContainerStatus for \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\": not found" May 7 23:49:21.280382 kubelet[3220]: E0507 23:49:21.280335 3220 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\": not found" containerID="b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436" May 7 23:49:21.280473 kubelet[3220]: I0507 23:49:21.280386 3220 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436"} err="failed to get container status \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\": rpc error: code = NotFound desc = an error occurred when try to find container \"b45f00b1236410249d8644a23a6727718197429d821fa1aff43c2d4965388436\": not found" May 7 23:49:21.976439 sshd[5062]: Connection closed by 147.75.109.163 port 37862 May 7 23:49:21.977444 sshd-session[5060]: pam_unix(sshd:session): session closed for user core May 7 23:49:21.982906 systemd[1]: sshd@27-172.31.28.15:22-147.75.109.163:37862.service: Deactivated successfully. May 7 23:49:21.986720 systemd[1]: session-26.scope: Deactivated successfully. May 7 23:49:21.987722 systemd[1]: session-26.scope: Consumed 2.659s CPU time, 25.8M memory peak. May 7 23:49:21.990372 systemd-logind[1936]: Session 26 logged out. Waiting for processes to exit. May 7 23:49:21.993225 systemd-logind[1936]: Removed session 26. May 7 23:49:22.017631 systemd[1]: Started sshd@28-172.31.28.15:22-147.75.109.163:37938.service - OpenSSH per-connection server daemon (147.75.109.163:37938). May 7 23:49:22.204270 sshd[5219]: Accepted publickey for core from 147.75.109.163 port 37938 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:49:22.206696 sshd-session[5219]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:49:22.216079 systemd-logind[1936]: New session 27 of user core. May 7 23:49:22.226422 systemd[1]: Started session-27.scope - Session 27 of User core. May 7 23:49:22.710692 kubelet[3220]: I0507 23:49:22.710468 3220 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00cd9891-bcda-4389-9118-a75d01bcd984" path="/var/lib/kubelet/pods/00cd9891-bcda-4389-9118-a75d01bcd984/volumes" May 7 23:49:22.713139 kubelet[3220]: I0507 23:49:22.712262 3220 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" path="/var/lib/kubelet/pods/e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c/volumes" May 7 23:49:22.885525 ntpd[1928]: Deleting interface #11 lxc_health, fe80::c3:56ff:fec5:d642%8#123, interface stats: received=0, sent=0, dropped=0, active_time=88 secs May 7 23:49:22.886004 ntpd[1928]: 7 May 23:49:22 ntpd[1928]: Deleting interface #11 lxc_health, fe80::c3:56ff:fec5:d642%8#123, interface stats: received=0, sent=0, dropped=0, active_time=88 secs May 7 23:49:24.004175 sshd[5221]: Connection closed by 147.75.109.163 port 37938 May 7 23:49:24.005202 sshd-session[5219]: pam_unix(sshd:session): session closed for user core May 7 23:49:24.017456 systemd-logind[1936]: Session 27 logged out. Waiting for processes to exit. May 7 23:49:24.019725 systemd[1]: sshd@28-172.31.28.15:22-147.75.109.163:37938.service: Deactivated successfully. May 7 23:49:24.025602 systemd[1]: session-27.scope: Deactivated successfully. May 7 23:49:24.027801 systemd[1]: session-27.scope: Consumed 1.588s CPU time, 24.2M memory peak. May 7 23:49:24.051291 systemd-logind[1936]: Removed session 27. May 7 23:49:24.064645 systemd[1]: Started sshd@29-172.31.28.15:22-147.75.109.163:37952.service - OpenSSH per-connection server daemon (147.75.109.163:37952). May 7 23:49:24.111150 kubelet[3220]: I0507 23:49:24.107735 3220 memory_manager.go:355] "RemoveStaleState removing state" podUID="00cd9891-bcda-4389-9118-a75d01bcd984" containerName="cilium-operator" May 7 23:49:24.111150 kubelet[3220]: I0507 23:49:24.107785 3220 memory_manager.go:355] "RemoveStaleState removing state" podUID="e62d43b6-116a-4d7f-bd3f-0bbbd8bcaa8c" containerName="cilium-agent" May 7 23:49:24.133810 systemd[1]: Created slice kubepods-burstable-pod14fbcc0e_9757_4aad_9198_120f5bcbb1a2.slice - libcontainer container kubepods-burstable-pod14fbcc0e_9757_4aad_9198_120f5bcbb1a2.slice. May 7 23:49:24.168639 kubelet[3220]: I0507 23:49:24.168581 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-cilium-cgroup\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.168948 kubelet[3220]: I0507 23:49:24.168919 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-bpf-maps\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.169154 kubelet[3220]: I0507 23:49:24.169125 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-etc-cni-netd\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.169330 kubelet[3220]: I0507 23:49:24.169302 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-lib-modules\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170000 kubelet[3220]: I0507 23:49:24.169451 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-cilium-ipsec-secrets\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170000 kubelet[3220]: I0507 23:49:24.169499 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-host-proc-sys-kernel\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170000 kubelet[3220]: I0507 23:49:24.169570 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-cilium-config-path\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170000 kubelet[3220]: I0507 23:49:24.169639 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-cilium-run\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170000 kubelet[3220]: I0507 23:49:24.169680 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-hostproc\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170349 kubelet[3220]: I0507 23:49:24.169730 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-clustermesh-secrets\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170349 kubelet[3220]: I0507 23:49:24.169782 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mq9c\" (UniqueName: \"kubernetes.io/projected/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-kube-api-access-9mq9c\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170349 kubelet[3220]: I0507 23:49:24.169824 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-xtables-lock\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170349 kubelet[3220]: I0507 23:49:24.169858 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-host-proc-sys-net\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170349 kubelet[3220]: I0507 23:49:24.169900 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-hubble-tls\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.170349 kubelet[3220]: I0507 23:49:24.169938 3220 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/14fbcc0e-9757-4aad-9198-120f5bcbb1a2-cni-path\") pod \"cilium-sd7cc\" (UID: \"14fbcc0e-9757-4aad-9198-120f5bcbb1a2\") " pod="kube-system/cilium-sd7cc" May 7 23:49:24.304301 sshd[5232]: Accepted publickey for core from 147.75.109.163 port 37952 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:49:24.310733 sshd-session[5232]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:49:24.348968 systemd-logind[1936]: New session 28 of user core. May 7 23:49:24.353437 systemd[1]: Started session-28.scope - Session 28 of User core. May 7 23:49:24.446178 containerd[1963]: time="2025-05-07T23:49:24.446042156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-sd7cc,Uid:14fbcc0e-9757-4aad-9198-120f5bcbb1a2,Namespace:kube-system,Attempt:0,}" May 7 23:49:24.489897 sshd[5238]: Connection closed by 147.75.109.163 port 37952 May 7 23:49:24.493346 containerd[1963]: time="2025-05-07T23:49:24.491791665Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 7 23:49:24.493346 containerd[1963]: time="2025-05-07T23:49:24.491895201Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 7 23:49:24.493346 containerd[1963]: time="2025-05-07T23:49:24.491990565Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:49:24.492255 sshd-session[5232]: pam_unix(sshd:session): session closed for user core May 7 23:49:24.495139 containerd[1963]: time="2025-05-07T23:49:24.493039377Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 7 23:49:24.501088 systemd-logind[1936]: Session 28 logged out. Waiting for processes to exit. May 7 23:49:24.503844 systemd[1]: sshd@29-172.31.28.15:22-147.75.109.163:37952.service: Deactivated successfully. May 7 23:49:24.508958 systemd[1]: session-28.scope: Deactivated successfully. May 7 23:49:24.524734 systemd-logind[1936]: Removed session 28. May 7 23:49:24.536445 systemd[1]: Started cri-containerd-5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d.scope - libcontainer container 5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d. May 7 23:49:24.540960 systemd[1]: Started sshd@30-172.31.28.15:22-147.75.109.163:37966.service - OpenSSH per-connection server daemon (147.75.109.163:37966). May 7 23:49:24.603778 containerd[1963]: time="2025-05-07T23:49:24.602964945Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-sd7cc,Uid:14fbcc0e-9757-4aad-9198-120f5bcbb1a2,Namespace:kube-system,Attempt:0,} returns sandbox id \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\"" May 7 23:49:24.610413 containerd[1963]: time="2025-05-07T23:49:24.610325277Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 7 23:49:24.641895 containerd[1963]: time="2025-05-07T23:49:24.641735913Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44\"" May 7 23:49:24.644286 containerd[1963]: time="2025-05-07T23:49:24.644080017Z" level=info msg="StartContainer for \"c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44\"" May 7 23:49:24.697419 systemd[1]: Started cri-containerd-c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44.scope - libcontainer container c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44. May 7 23:49:24.715274 containerd[1963]: time="2025-05-07T23:49:24.715212850Z" level=info msg="StopPodSandbox for \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\"" May 7 23:49:24.715649 containerd[1963]: time="2025-05-07T23:49:24.715358950Z" level=info msg="TearDown network for sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" successfully" May 7 23:49:24.715649 containerd[1963]: time="2025-05-07T23:49:24.715381978Z" level=info msg="StopPodSandbox for \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" returns successfully" May 7 23:49:24.716587 containerd[1963]: time="2025-05-07T23:49:24.716166550Z" level=info msg="RemovePodSandbox for \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\"" May 7 23:49:24.716587 containerd[1963]: time="2025-05-07T23:49:24.716228890Z" level=info msg="Forcibly stopping sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\"" May 7 23:49:24.716587 containerd[1963]: time="2025-05-07T23:49:24.716323054Z" level=info msg="TearDown network for sandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" successfully" May 7 23:49:24.725670 containerd[1963]: time="2025-05-07T23:49:24.725611738Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." May 7 23:49:24.725822 containerd[1963]: time="2025-05-07T23:49:24.725698078Z" level=info msg="RemovePodSandbox \"8de2d2ed5f42bbb303de9a57dee53c8004a5f2f1237a2f94321e7b94c5c04e5a\" returns successfully" May 7 23:49:24.729926 containerd[1963]: time="2025-05-07T23:49:24.729637690Z" level=info msg="StopPodSandbox for \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\"" May 7 23:49:24.729926 containerd[1963]: time="2025-05-07T23:49:24.729798922Z" level=info msg="TearDown network for sandbox \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" successfully" May 7 23:49:24.729926 containerd[1963]: time="2025-05-07T23:49:24.729821518Z" level=info msg="StopPodSandbox for \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" returns successfully" May 7 23:49:24.731356 containerd[1963]: time="2025-05-07T23:49:24.731008150Z" level=info msg="RemovePodSandbox for \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\"" May 7 23:49:24.731356 containerd[1963]: time="2025-05-07T23:49:24.731061442Z" level=info msg="Forcibly stopping sandbox \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\"" May 7 23:49:24.732015 containerd[1963]: time="2025-05-07T23:49:24.731950438Z" level=info msg="TearDown network for sandbox \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" successfully" May 7 23:49:24.741046 containerd[1963]: time="2025-05-07T23:49:24.740606086Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." May 7 23:49:24.741046 containerd[1963]: time="2025-05-07T23:49:24.740686018Z" level=info msg="RemovePodSandbox \"de3eeda258218bdca02842c33175575a1494d6b32630ec68f304ec2756429c9b\" returns successfully" May 7 23:49:24.766653 sshd[5274]: Accepted publickey for core from 147.75.109.163 port 37966 ssh2: RSA SHA256:kQP1JwyMe/WwD6o95f0kuF0WNYd/0mECzU0K15pTcJg May 7 23:49:24.767664 containerd[1963]: time="2025-05-07T23:49:24.766564138Z" level=info msg="StartContainer for \"c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44\" returns successfully" May 7 23:49:24.773560 sshd-session[5274]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 7 23:49:24.788444 systemd[1]: cri-containerd-c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44.scope: Deactivated successfully. May 7 23:49:24.788553 systemd-logind[1936]: New session 29 of user core. May 7 23:49:24.797752 systemd[1]: Started session-29.scope - Session 29 of User core. May 7 23:49:24.865522 containerd[1963]: time="2025-05-07T23:49:24.865304134Z" level=info msg="shim disconnected" id=c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44 namespace=k8s.io May 7 23:49:24.865522 containerd[1963]: time="2025-05-07T23:49:24.865408114Z" level=warning msg="cleaning up after shim disconnected" id=c1f3b0972b6f5bd58c4dc11477ac8595d49596ba43fc2ec1b364302592058e44 namespace=k8s.io May 7 23:49:24.865522 containerd[1963]: time="2025-05-07T23:49:24.865444126Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:24.938151 kubelet[3220]: E0507 23:49:24.937788 3220 kubelet.go:3008] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 7 23:49:25.239969 containerd[1963]: time="2025-05-07T23:49:25.239804420Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 7 23:49:25.269164 containerd[1963]: time="2025-05-07T23:49:25.265624208Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78\"" May 7 23:49:25.273577 containerd[1963]: time="2025-05-07T23:49:25.273511940Z" level=info msg="StartContainer for \"0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78\"" May 7 23:49:25.342417 systemd[1]: Started cri-containerd-0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78.scope - libcontainer container 0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78. May 7 23:49:25.394266 containerd[1963]: time="2025-05-07T23:49:25.394064745Z" level=info msg="StartContainer for \"0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78\" returns successfully" May 7 23:49:25.408252 systemd[1]: cri-containerd-0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78.scope: Deactivated successfully. May 7 23:49:25.445247 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78-rootfs.mount: Deactivated successfully. May 7 23:49:25.455616 containerd[1963]: time="2025-05-07T23:49:25.455531517Z" level=info msg="shim disconnected" id=0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78 namespace=k8s.io May 7 23:49:25.455616 containerd[1963]: time="2025-05-07T23:49:25.455604933Z" level=warning msg="cleaning up after shim disconnected" id=0415afd0e472b5fd8da2c17831bdb07d227dab8a0d06436c318d50dc3ae52c78 namespace=k8s.io May 7 23:49:25.456580 containerd[1963]: time="2025-05-07T23:49:25.455626317Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:26.243782 containerd[1963]: time="2025-05-07T23:49:26.243678213Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 7 23:49:26.278764 containerd[1963]: time="2025-05-07T23:49:26.278680953Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5\"" May 7 23:49:26.281152 containerd[1963]: time="2025-05-07T23:49:26.279549753Z" level=info msg="StartContainer for \"b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5\"" May 7 23:49:26.350414 systemd[1]: Started cri-containerd-b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5.scope - libcontainer container b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5. May 7 23:49:26.410345 systemd[1]: cri-containerd-b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5.scope: Deactivated successfully. May 7 23:49:26.410541 containerd[1963]: time="2025-05-07T23:49:26.410424394Z" level=info msg="StartContainer for \"b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5\" returns successfully" May 7 23:49:26.452395 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5-rootfs.mount: Deactivated successfully. May 7 23:49:26.464563 containerd[1963]: time="2025-05-07T23:49:26.464427994Z" level=info msg="shim disconnected" id=b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5 namespace=k8s.io May 7 23:49:26.465241 containerd[1963]: time="2025-05-07T23:49:26.464581414Z" level=warning msg="cleaning up after shim disconnected" id=b0c5645d8a5a604cafac51d4c9eebde13f237f89ee718044384f1623870800b5 namespace=k8s.io May 7 23:49:26.465241 containerd[1963]: time="2025-05-07T23:49:26.464604550Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:27.249792 containerd[1963]: time="2025-05-07T23:49:27.249536614Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 7 23:49:27.275262 containerd[1963]: time="2025-05-07T23:49:27.275051734Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef\"" May 7 23:49:27.279174 containerd[1963]: time="2025-05-07T23:49:27.277737838Z" level=info msg="StartContainer for \"8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef\"" May 7 23:49:27.340498 systemd[1]: Started cri-containerd-8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef.scope - libcontainer container 8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef. May 7 23:49:27.382298 kubelet[3220]: I0507 23:49:27.381392 3220 setters.go:602] "Node became not ready" node="ip-172-31-28-15" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-07T23:49:27Z","lastTransitionTime":"2025-05-07T23:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 7 23:49:27.388147 systemd[1]: cri-containerd-8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef.scope: Deactivated successfully. May 7 23:49:27.392732 containerd[1963]: time="2025-05-07T23:49:27.392500523Z" level=info msg="StartContainer for \"8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef\" returns successfully" May 7 23:49:27.442053 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef-rootfs.mount: Deactivated successfully. May 7 23:49:27.450310 containerd[1963]: time="2025-05-07T23:49:27.450099599Z" level=info msg="shim disconnected" id=8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef namespace=k8s.io May 7 23:49:27.450310 containerd[1963]: time="2025-05-07T23:49:27.450216131Z" level=warning msg="cleaning up after shim disconnected" id=8df66ca9f3523c1d7bccf96fe83ef10a959b1c499eb0b8a72c2350ba6f6de9ef namespace=k8s.io May 7 23:49:27.450310 containerd[1963]: time="2025-05-07T23:49:27.450236051Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 7 23:49:28.257900 containerd[1963]: time="2025-05-07T23:49:28.257514539Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 7 23:49:28.292898 containerd[1963]: time="2025-05-07T23:49:28.292818215Z" level=info msg="CreateContainer within sandbox \"5ebad153881a27a32ee1a480490b55b197594025fa06146a6f67baf4131e8b4d\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"83837fdd828baddc07eee36642586837690b2963e8fc42a79e7d172bd4ab5632\"" May 7 23:49:28.295365 containerd[1963]: time="2025-05-07T23:49:28.294151067Z" level=info msg="StartContainer for \"83837fdd828baddc07eee36642586837690b2963e8fc42a79e7d172bd4ab5632\"" May 7 23:49:28.351404 systemd[1]: Started cri-containerd-83837fdd828baddc07eee36642586837690b2963e8fc42a79e7d172bd4ab5632.scope - libcontainer container 83837fdd828baddc07eee36642586837690b2963e8fc42a79e7d172bd4ab5632. May 7 23:49:28.411352 containerd[1963]: time="2025-05-07T23:49:28.411028440Z" level=info msg="StartContainer for \"83837fdd828baddc07eee36642586837690b2963e8fc42a79e7d172bd4ab5632\" returns successfully" May 7 23:49:29.244258 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) May 7 23:49:29.299783 kubelet[3220]: I0507 23:49:29.299260 3220 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-sd7cc" podStartSLOduration=6.299237844 podStartE2EDuration="6.299237844s" podCreationTimestamp="2025-05-07 23:49:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-07 23:49:29.299085924 +0000 UTC m=+124.891448311" watchObservedRunningTime="2025-05-07 23:49:29.299237844 +0000 UTC m=+124.891600255" May 7 23:49:33.442039 (udev-worker)[6068]: Network interface NamePolicy= disabled on kernel command line. May 7 23:49:33.446548 systemd-networkd[1866]: lxc_health: Link UP May 7 23:49:33.448936 systemd-networkd[1866]: lxc_health: Gained carrier May 7 23:49:35.134487 systemd-networkd[1866]: lxc_health: Gained IPv6LL May 7 23:49:37.885266 ntpd[1928]: Listen normally on 14 lxc_health [fe80::acd9:8eff:fe3d:ed54%14]:123 May 7 23:49:37.886615 ntpd[1928]: 7 May 23:49:37 ntpd[1928]: Listen normally on 14 lxc_health [fe80::acd9:8eff:fe3d:ed54%14]:123 May 7 23:49:38.335338 systemd[1]: run-containerd-runc-k8s.io-83837fdd828baddc07eee36642586837690b2963e8fc42a79e7d172bd4ab5632-runc.ZN9Cj8.mount: Deactivated successfully. May 7 23:49:38.454939 sshd[5334]: Connection closed by 147.75.109.163 port 37966 May 7 23:49:38.455847 sshd-session[5274]: pam_unix(sshd:session): session closed for user core May 7 23:49:38.465675 systemd-logind[1936]: Session 29 logged out. Waiting for processes to exit. May 7 23:49:38.468511 systemd[1]: sshd@30-172.31.28.15:22-147.75.109.163:37966.service: Deactivated successfully. May 7 23:49:38.474969 systemd[1]: session-29.scope: Deactivated successfully. May 7 23:49:38.477970 systemd-logind[1936]: Removed session 29.