Feb 13 19:00:37.216689 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd083] Feb 13 19:00:37.216735 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT Thu Feb 13 17:39:57 -00 2025 Feb 13 19:00:37.216761 kernel: KASLR disabled due to lack of seed Feb 13 19:00:37.216778 kernel: efi: EFI v2.7 by EDK II Feb 13 19:00:37.216793 kernel: efi: SMBIOS=0x7bed0000 SMBIOS 3.0=0x7beb0000 ACPI=0x786e0000 ACPI 2.0=0x786e0014 MEMATTR=0x7a736a98 MEMRESERVE=0x78557598 Feb 13 19:00:37.216809 kernel: secureboot: Secure boot disabled Feb 13 19:00:37.216827 kernel: ACPI: Early table checksum verification disabled Feb 13 19:00:37.216842 kernel: ACPI: RSDP 0x00000000786E0014 000024 (v02 AMAZON) Feb 13 19:00:37.216858 kernel: ACPI: XSDT 0x00000000786D00E8 000064 (v01 AMAZON AMZNFACP 00000001 01000013) Feb 13 19:00:37.216874 kernel: ACPI: FACP 0x00000000786B0000 000114 (v06 AMAZON AMZNFACP 00000001 AMZN 00000001) Feb 13 19:00:37.216894 kernel: ACPI: DSDT 0x0000000078640000 00159D (v02 AMAZON AMZNDSDT 00000001 INTL 20160527) Feb 13 19:00:37.216910 kernel: ACPI: APIC 0x00000000786C0000 000108 (v04 AMAZON AMZNAPIC 00000001 AMZN 00000001) Feb 13 19:00:37.216925 kernel: ACPI: SPCR 0x00000000786A0000 000050 (v02 AMAZON AMZNSPCR 00000001 AMZN 00000001) Feb 13 19:00:37.216941 kernel: ACPI: GTDT 0x0000000078690000 000060 (v02 AMAZON AMZNGTDT 00000001 AMZN 00000001) Feb 13 19:00:37.216959 kernel: ACPI: MCFG 0x0000000078680000 00003C (v02 AMAZON AMZNMCFG 00000001 AMZN 00000001) Feb 13 19:00:37.216981 kernel: ACPI: SLIT 0x0000000078670000 00002D (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Feb 13 19:00:37.216998 kernel: ACPI: IORT 0x0000000078660000 000078 (v01 AMAZON AMZNIORT 00000001 AMZN 00000001) Feb 13 19:00:37.217015 kernel: ACPI: PPTT 0x0000000078650000 0000EC (v01 AMAZON AMZNPPTT 00000001 AMZN 00000001) Feb 13 19:00:37.217031 kernel: ACPI: SPCR: console: uart,mmio,0x90a0000,115200 Feb 13 19:00:37.217047 kernel: earlycon: uart0 at MMIO 0x00000000090a0000 (options '115200') Feb 13 19:00:37.217064 kernel: printk: bootconsole [uart0] enabled Feb 13 19:00:37.217081 kernel: NUMA: Failed to initialise from firmware Feb 13 19:00:37.217098 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000004b5ffffff] Feb 13 19:00:37.219162 kernel: NUMA: NODE_DATA [mem 0x4b583f800-0x4b5844fff] Feb 13 19:00:37.219209 kernel: Zone ranges: Feb 13 19:00:37.219227 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Feb 13 19:00:37.219256 kernel: DMA32 empty Feb 13 19:00:37.219273 kernel: Normal [mem 0x0000000100000000-0x00000004b5ffffff] Feb 13 19:00:37.219290 kernel: Movable zone start for each node Feb 13 19:00:37.219306 kernel: Early memory node ranges Feb 13 19:00:37.219322 kernel: node 0: [mem 0x0000000040000000-0x000000007862ffff] Feb 13 19:00:37.219339 kernel: node 0: [mem 0x0000000078630000-0x000000007863ffff] Feb 13 19:00:37.219355 kernel: node 0: [mem 0x0000000078640000-0x00000000786effff] Feb 13 19:00:37.219372 kernel: node 0: [mem 0x00000000786f0000-0x000000007872ffff] Feb 13 19:00:37.219388 kernel: node 0: [mem 0x0000000078730000-0x000000007bbfffff] Feb 13 19:00:37.219404 kernel: node 0: [mem 0x000000007bc00000-0x000000007bfdffff] Feb 13 19:00:37.219420 kernel: node 0: [mem 0x000000007bfe0000-0x000000007fffffff] Feb 13 19:00:37.219436 kernel: node 0: [mem 0x0000000400000000-0x00000004b5ffffff] Feb 13 19:00:37.219457 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000004b5ffffff] Feb 13 19:00:37.219475 kernel: On node 0, zone Normal: 8192 pages in unavailable ranges Feb 13 19:00:37.219521 kernel: psci: probing for conduit method from ACPI. Feb 13 19:00:37.219539 kernel: psci: PSCIv1.0 detected in firmware. Feb 13 19:00:37.219557 kernel: psci: Using standard PSCI v0.2 function IDs Feb 13 19:00:37.219579 kernel: psci: Trusted OS migration not required Feb 13 19:00:37.219596 kernel: psci: SMC Calling Convention v1.1 Feb 13 19:00:37.219614 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Feb 13 19:00:37.219631 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Feb 13 19:00:37.219649 kernel: pcpu-alloc: [0] 0 [0] 1 Feb 13 19:00:37.219666 kernel: Detected PIPT I-cache on CPU0 Feb 13 19:00:37.219684 kernel: CPU features: detected: GIC system register CPU interface Feb 13 19:00:37.219701 kernel: CPU features: detected: Spectre-v2 Feb 13 19:00:37.219718 kernel: CPU features: detected: Spectre-v3a Feb 13 19:00:37.219735 kernel: CPU features: detected: Spectre-BHB Feb 13 19:00:37.219752 kernel: CPU features: detected: ARM erratum 1742098 Feb 13 19:00:37.219770 kernel: CPU features: detected: ARM errata 1165522, 1319367, or 1530923 Feb 13 19:00:37.219791 kernel: alternatives: applying boot alternatives Feb 13 19:00:37.219811 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=f06bad36699a22ae88c1968cd72b62b3503d97da521712e50a4b744320b1ba33 Feb 13 19:00:37.219830 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 19:00:37.219848 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 19:00:37.219865 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 19:00:37.219882 kernel: Fallback order for Node 0: 0 Feb 13 19:00:37.219900 kernel: Built 1 zonelists, mobility grouping on. Total pages: 991872 Feb 13 19:00:37.219917 kernel: Policy zone: Normal Feb 13 19:00:37.219934 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 19:00:37.219951 kernel: software IO TLB: area num 2. Feb 13 19:00:37.219973 kernel: software IO TLB: mapped [mem 0x000000007c000000-0x0000000080000000] (64MB) Feb 13 19:00:37.219990 kernel: Memory: 3821240K/4030464K available (10304K kernel code, 2186K rwdata, 8092K rodata, 38336K init, 897K bss, 209224K reserved, 0K cma-reserved) Feb 13 19:00:37.220008 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 13 19:00:37.220025 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 19:00:37.220043 kernel: rcu: RCU event tracing is enabled. Feb 13 19:00:37.220062 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 13 19:00:37.220079 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 19:00:37.220097 kernel: Tracing variant of Tasks RCU enabled. Feb 13 19:00:37.220134 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 19:00:37.220157 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 13 19:00:37.220174 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Feb 13 19:00:37.220198 kernel: GICv3: 96 SPIs implemented Feb 13 19:00:37.220215 kernel: GICv3: 0 Extended SPIs implemented Feb 13 19:00:37.220232 kernel: Root IRQ handler: gic_handle_irq Feb 13 19:00:37.220249 kernel: GICv3: GICv3 features: 16 PPIs Feb 13 19:00:37.220266 kernel: GICv3: CPU0: found redistributor 0 region 0:0x0000000010200000 Feb 13 19:00:37.220283 kernel: ITS [mem 0x10080000-0x1009ffff] Feb 13 19:00:37.220301 kernel: ITS@0x0000000010080000: allocated 8192 Devices @4000b0000 (indirect, esz 8, psz 64K, shr 1) Feb 13 19:00:37.220319 kernel: ITS@0x0000000010080000: allocated 8192 Interrupt Collections @4000c0000 (flat, esz 8, psz 64K, shr 1) Feb 13 19:00:37.220336 kernel: GICv3: using LPI property table @0x00000004000d0000 Feb 13 19:00:37.220353 kernel: ITS: Using hypervisor restricted LPI range [128] Feb 13 19:00:37.220370 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000004000e0000 Feb 13 19:00:37.220387 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 19:00:37.220410 kernel: arch_timer: cp15 timer(s) running at 83.33MHz (virt). Feb 13 19:00:37.220427 kernel: clocksource: arch_sys_counter: mask: 0x1ffffffffffffff max_cycles: 0x13381ebeec, max_idle_ns: 440795203145 ns Feb 13 19:00:37.220445 kernel: sched_clock: 57 bits at 83MHz, resolution 12ns, wraps every 4398046511100ns Feb 13 19:00:37.220462 kernel: Console: colour dummy device 80x25 Feb 13 19:00:37.220480 kernel: printk: console [tty1] enabled Feb 13 19:00:37.220498 kernel: ACPI: Core revision 20230628 Feb 13 19:00:37.220516 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 166.66 BogoMIPS (lpj=83333) Feb 13 19:00:37.220533 kernel: pid_max: default: 32768 minimum: 301 Feb 13 19:00:37.220551 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 19:00:37.220569 kernel: landlock: Up and running. Feb 13 19:00:37.220592 kernel: SELinux: Initializing. Feb 13 19:00:37.220610 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:00:37.220627 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:00:37.220645 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 19:00:37.220664 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 19:00:37.220681 kernel: rcu: Hierarchical SRCU implementation. Feb 13 19:00:37.220699 kernel: rcu: Max phase no-delay instances is 400. Feb 13 19:00:37.220717 kernel: Platform MSI: ITS@0x10080000 domain created Feb 13 19:00:37.220739 kernel: PCI/MSI: ITS@0x10080000 domain created Feb 13 19:00:37.220757 kernel: Remapping and enabling EFI services. Feb 13 19:00:37.220775 kernel: smp: Bringing up secondary CPUs ... Feb 13 19:00:37.220793 kernel: Detected PIPT I-cache on CPU1 Feb 13 19:00:37.220811 kernel: GICv3: CPU1: found redistributor 1 region 0:0x0000000010220000 Feb 13 19:00:37.220828 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000004000f0000 Feb 13 19:00:37.220846 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd083] Feb 13 19:00:37.220863 kernel: smp: Brought up 1 node, 2 CPUs Feb 13 19:00:37.220881 kernel: SMP: Total of 2 processors activated. Feb 13 19:00:37.220898 kernel: CPU features: detected: 32-bit EL0 Support Feb 13 19:00:37.220921 kernel: CPU features: detected: 32-bit EL1 Support Feb 13 19:00:37.220940 kernel: CPU features: detected: CRC32 instructions Feb 13 19:00:37.220970 kernel: CPU: All CPU(s) started at EL1 Feb 13 19:00:37.220992 kernel: alternatives: applying system-wide alternatives Feb 13 19:00:37.221010 kernel: devtmpfs: initialized Feb 13 19:00:37.221029 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 19:00:37.221047 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 13 19:00:37.221066 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 19:00:37.221084 kernel: SMBIOS 3.0.0 present. Feb 13 19:00:37.221107 kernel: DMI: Amazon EC2 a1.large/, BIOS 1.0 11/1/2018 Feb 13 19:00:37.224090 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 19:00:37.224136 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Feb 13 19:00:37.224160 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Feb 13 19:00:37.224180 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Feb 13 19:00:37.224199 kernel: audit: initializing netlink subsys (disabled) Feb 13 19:00:37.224218 kernel: audit: type=2000 audit(0.221:1): state=initialized audit_enabled=0 res=1 Feb 13 19:00:37.224246 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 19:00:37.224265 kernel: cpuidle: using governor menu Feb 13 19:00:37.224284 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Feb 13 19:00:37.224303 kernel: ASID allocator initialised with 65536 entries Feb 13 19:00:37.224322 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 19:00:37.224340 kernel: Serial: AMBA PL011 UART driver Feb 13 19:00:37.224360 kernel: Modules: 17760 pages in range for non-PLT usage Feb 13 19:00:37.224379 kernel: Modules: 509280 pages in range for PLT usage Feb 13 19:00:37.224397 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 19:00:37.224421 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 19:00:37.224440 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Feb 13 19:00:37.224458 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Feb 13 19:00:37.224477 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 19:00:37.224496 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 19:00:37.224515 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Feb 13 19:00:37.224534 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Feb 13 19:00:37.224553 kernel: ACPI: Added _OSI(Module Device) Feb 13 19:00:37.224571 kernel: ACPI: Added _OSI(Processor Device) Feb 13 19:00:37.224594 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 19:00:37.224613 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 19:00:37.224631 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 19:00:37.224650 kernel: ACPI: Interpreter enabled Feb 13 19:00:37.224669 kernel: ACPI: Using GIC for interrupt routing Feb 13 19:00:37.224688 kernel: ACPI: MCFG table detected, 1 entries Feb 13 19:00:37.224706 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-0f]) Feb 13 19:00:37.225015 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Feb 13 19:00:37.225276 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Feb 13 19:00:37.225520 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Feb 13 19:00:37.225728 kernel: acpi PNP0A08:00: ECAM area [mem 0x20000000-0x20ffffff] reserved by PNP0C02:00 Feb 13 19:00:37.225929 kernel: acpi PNP0A08:00: ECAM at [mem 0x20000000-0x20ffffff] for [bus 00-0f] Feb 13 19:00:37.225955 kernel: ACPI: Remapped I/O 0x000000001fff0000 to [io 0x0000-0xffff window] Feb 13 19:00:37.225974 kernel: acpiphp: Slot [1] registered Feb 13 19:00:37.225993 kernel: acpiphp: Slot [2] registered Feb 13 19:00:37.226011 kernel: acpiphp: Slot [3] registered Feb 13 19:00:37.226038 kernel: acpiphp: Slot [4] registered Feb 13 19:00:37.226057 kernel: acpiphp: Slot [5] registered Feb 13 19:00:37.226075 kernel: acpiphp: Slot [6] registered Feb 13 19:00:37.226093 kernel: acpiphp: Slot [7] registered Feb 13 19:00:37.226111 kernel: acpiphp: Slot [8] registered Feb 13 19:00:37.226170 kernel: acpiphp: Slot [9] registered Feb 13 19:00:37.226190 kernel: acpiphp: Slot [10] registered Feb 13 19:00:37.226208 kernel: acpiphp: Slot [11] registered Feb 13 19:00:37.226227 kernel: acpiphp: Slot [12] registered Feb 13 19:00:37.226245 kernel: acpiphp: Slot [13] registered Feb 13 19:00:37.226270 kernel: acpiphp: Slot [14] registered Feb 13 19:00:37.226289 kernel: acpiphp: Slot [15] registered Feb 13 19:00:37.226309 kernel: acpiphp: Slot [16] registered Feb 13 19:00:37.226327 kernel: acpiphp: Slot [17] registered Feb 13 19:00:37.226345 kernel: acpiphp: Slot [18] registered Feb 13 19:00:37.226363 kernel: acpiphp: Slot [19] registered Feb 13 19:00:37.226382 kernel: acpiphp: Slot [20] registered Feb 13 19:00:37.226400 kernel: acpiphp: Slot [21] registered Feb 13 19:00:37.226418 kernel: acpiphp: Slot [22] registered Feb 13 19:00:37.226441 kernel: acpiphp: Slot [23] registered Feb 13 19:00:37.226460 kernel: acpiphp: Slot [24] registered Feb 13 19:00:37.226478 kernel: acpiphp: Slot [25] registered Feb 13 19:00:37.226496 kernel: acpiphp: Slot [26] registered Feb 13 19:00:37.226514 kernel: acpiphp: Slot [27] registered Feb 13 19:00:37.226533 kernel: acpiphp: Slot [28] registered Feb 13 19:00:37.226551 kernel: acpiphp: Slot [29] registered Feb 13 19:00:37.226569 kernel: acpiphp: Slot [30] registered Feb 13 19:00:37.226587 kernel: acpiphp: Slot [31] registered Feb 13 19:00:37.226606 kernel: PCI host bridge to bus 0000:00 Feb 13 19:00:37.226850 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xffffffff window] Feb 13 19:00:37.227048 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Feb 13 19:00:37.227281 kernel: pci_bus 0000:00: root bus resource [mem 0x400000000000-0x407fffffffff window] Feb 13 19:00:37.227503 kernel: pci_bus 0000:00: root bus resource [bus 00-0f] Feb 13 19:00:37.227756 kernel: pci 0000:00:00.0: [1d0f:0200] type 00 class 0x060000 Feb 13 19:00:37.228006 kernel: pci 0000:00:01.0: [1d0f:8250] type 00 class 0x070003 Feb 13 19:00:37.230392 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80118000-0x80118fff] Feb 13 19:00:37.230645 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Feb 13 19:00:37.230860 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80114000-0x80117fff] Feb 13 19:00:37.231072 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D2 D3hot D3cold Feb 13 19:00:37.231372 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Feb 13 19:00:37.231616 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80110000-0x80113fff] Feb 13 19:00:37.231828 kernel: pci 0000:00:05.0: reg 0x18: [mem 0x80000000-0x800fffff pref] Feb 13 19:00:37.232047 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x80100000-0x8010ffff] Feb 13 19:00:37.233749 kernel: pci 0000:00:05.0: PME# supported from D0 D1 D2 D3hot D3cold Feb 13 19:00:37.233986 kernel: pci 0000:00:05.0: BAR 2: assigned [mem 0x80000000-0x800fffff pref] Feb 13 19:00:37.234258 kernel: pci 0000:00:05.0: BAR 4: assigned [mem 0x80100000-0x8010ffff] Feb 13 19:00:37.234477 kernel: pci 0000:00:04.0: BAR 0: assigned [mem 0x80110000-0x80113fff] Feb 13 19:00:37.234685 kernel: pci 0000:00:05.0: BAR 0: assigned [mem 0x80114000-0x80117fff] Feb 13 19:00:37.234895 kernel: pci 0000:00:01.0: BAR 0: assigned [mem 0x80118000-0x80118fff] Feb 13 19:00:37.235096 kernel: pci_bus 0000:00: resource 4 [mem 0x80000000-0xffffffff window] Feb 13 19:00:37.236295 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Feb 13 19:00:37.236495 kernel: pci_bus 0000:00: resource 6 [mem 0x400000000000-0x407fffffffff window] Feb 13 19:00:37.236522 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Feb 13 19:00:37.236542 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Feb 13 19:00:37.236560 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Feb 13 19:00:37.236579 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Feb 13 19:00:37.236598 kernel: iommu: Default domain type: Translated Feb 13 19:00:37.236627 kernel: iommu: DMA domain TLB invalidation policy: strict mode Feb 13 19:00:37.236645 kernel: efivars: Registered efivars operations Feb 13 19:00:37.236664 kernel: vgaarb: loaded Feb 13 19:00:37.236682 kernel: clocksource: Switched to clocksource arch_sys_counter Feb 13 19:00:37.236700 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 19:00:37.236719 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 19:00:37.236737 kernel: pnp: PnP ACPI init Feb 13 19:00:37.236946 kernel: system 00:00: [mem 0x20000000-0x2fffffff] could not be reserved Feb 13 19:00:37.236979 kernel: pnp: PnP ACPI: found 1 devices Feb 13 19:00:37.236998 kernel: NET: Registered PF_INET protocol family Feb 13 19:00:37.237016 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 19:00:37.237035 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Feb 13 19:00:37.237054 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 19:00:37.237072 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 19:00:37.237090 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Feb 13 19:00:37.237109 kernel: TCP: Hash tables configured (established 32768 bind 32768) Feb 13 19:00:37.237149 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:00:37.237174 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:00:37.237194 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 19:00:37.237212 kernel: PCI: CLS 0 bytes, default 64 Feb 13 19:00:37.237230 kernel: kvm [1]: HYP mode not available Feb 13 19:00:37.237249 kernel: Initialise system trusted keyrings Feb 13 19:00:37.237267 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Feb 13 19:00:37.237286 kernel: Key type asymmetric registered Feb 13 19:00:37.237304 kernel: Asymmetric key parser 'x509' registered Feb 13 19:00:37.237322 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Feb 13 19:00:37.237345 kernel: io scheduler mq-deadline registered Feb 13 19:00:37.237364 kernel: io scheduler kyber registered Feb 13 19:00:37.237382 kernel: io scheduler bfq registered Feb 13 19:00:37.237605 kernel: pl061_gpio ARMH0061:00: PL061 GPIO chip registered Feb 13 19:00:37.237632 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Feb 13 19:00:37.237651 kernel: ACPI: button: Power Button [PWRB] Feb 13 19:00:37.237670 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input1 Feb 13 19:00:37.237688 kernel: ACPI: button: Sleep Button [SLPB] Feb 13 19:00:37.237712 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 19:00:37.237731 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Feb 13 19:00:37.237980 kernel: serial 0000:00:01.0: enabling device (0010 -> 0012) Feb 13 19:00:37.238009 kernel: printk: console [ttyS0] disabled Feb 13 19:00:37.238029 kernel: 0000:00:01.0: ttyS0 at MMIO 0x80118000 (irq = 14, base_baud = 115200) is a 16550A Feb 13 19:00:37.238048 kernel: printk: console [ttyS0] enabled Feb 13 19:00:37.238066 kernel: printk: bootconsole [uart0] disabled Feb 13 19:00:37.238084 kernel: thunder_xcv, ver 1.0 Feb 13 19:00:37.238103 kernel: thunder_bgx, ver 1.0 Feb 13 19:00:37.238232 kernel: nicpf, ver 1.0 Feb 13 19:00:37.238263 kernel: nicvf, ver 1.0 Feb 13 19:00:37.238505 kernel: rtc-efi rtc-efi.0: registered as rtc0 Feb 13 19:00:37.238700 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-02-13T19:00:36 UTC (1739473236) Feb 13 19:00:37.238726 kernel: hid: raw HID events driver (C) Jiri Kosina Feb 13 19:00:37.238746 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 3 counters available Feb 13 19:00:37.238766 kernel: watchdog: Delayed init of the lockup detector failed: -19 Feb 13 19:00:37.238784 kernel: watchdog: Hard watchdog permanently disabled Feb 13 19:00:37.238808 kernel: NET: Registered PF_INET6 protocol family Feb 13 19:00:37.238827 kernel: Segment Routing with IPv6 Feb 13 19:00:37.238845 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 19:00:37.238864 kernel: NET: Registered PF_PACKET protocol family Feb 13 19:00:37.238882 kernel: Key type dns_resolver registered Feb 13 19:00:37.238901 kernel: registered taskstats version 1 Feb 13 19:00:37.238920 kernel: Loading compiled-in X.509 certificates Feb 13 19:00:37.238938 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 58bec1a0c6b8a133d1af4ea745973da0351f7027' Feb 13 19:00:37.238956 kernel: Key type .fscrypt registered Feb 13 19:00:37.238974 kernel: Key type fscrypt-provisioning registered Feb 13 19:00:37.238997 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 19:00:37.239016 kernel: ima: Allocated hash algorithm: sha1 Feb 13 19:00:37.239034 kernel: ima: No architecture policies found Feb 13 19:00:37.239053 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Feb 13 19:00:37.239071 kernel: clk: Disabling unused clocks Feb 13 19:00:37.239089 kernel: Freeing unused kernel memory: 38336K Feb 13 19:00:37.239108 kernel: Run /init as init process Feb 13 19:00:37.239170 kernel: with arguments: Feb 13 19:00:37.239189 kernel: /init Feb 13 19:00:37.239214 kernel: with environment: Feb 13 19:00:37.239233 kernel: HOME=/ Feb 13 19:00:37.239252 kernel: TERM=linux Feb 13 19:00:37.239270 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 19:00:37.239290 systemd[1]: Successfully made /usr/ read-only. Feb 13 19:00:37.239315 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:00:37.239337 systemd[1]: Detected virtualization amazon. Feb 13 19:00:37.239361 systemd[1]: Detected architecture arm64. Feb 13 19:00:37.239381 systemd[1]: Running in initrd. Feb 13 19:00:37.239400 systemd[1]: No hostname configured, using default hostname. Feb 13 19:00:37.239420 systemd[1]: Hostname set to . Feb 13 19:00:37.239440 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:00:37.239460 systemd[1]: Queued start job for default target initrd.target. Feb 13 19:00:37.239497 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:00:37.239521 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:00:37.239543 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 19:00:37.239569 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:00:37.239590 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 19:00:37.239612 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 19:00:37.239634 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 19:00:37.239654 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 19:00:37.239674 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:00:37.239699 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:00:37.239719 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:00:37.239739 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:00:37.239758 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:00:37.239778 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:00:37.239798 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:00:37.239818 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:00:37.239838 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 19:00:37.239858 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Feb 13 19:00:37.239883 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:00:37.239903 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:00:37.239922 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:00:37.239942 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:00:37.239962 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 19:00:37.239981 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:00:37.240001 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 19:00:37.240021 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 19:00:37.240045 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:00:37.240065 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:00:37.240085 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:00:37.240105 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 19:00:37.240145 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:00:37.240168 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 19:00:37.240195 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:00:37.240215 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:00:37.240235 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:00:37.240256 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:00:37.240315 systemd-journald[252]: Collecting audit messages is disabled. Feb 13 19:00:37.240363 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:00:37.240384 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 19:00:37.240403 kernel: Bridge firewalling registered Feb 13 19:00:37.240423 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:00:37.240443 systemd-journald[252]: Journal started Feb 13 19:00:37.240484 systemd-journald[252]: Runtime Journal (/run/log/journal/ec2710d9e4512ee9262a9abc383a0092) is 8M, max 75.3M, 67.3M free. Feb 13 19:00:37.184068 systemd-modules-load[253]: Inserted module 'overlay' Feb 13 19:00:37.235192 systemd-modules-load[253]: Inserted module 'br_netfilter' Feb 13 19:00:37.247766 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:00:37.251209 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:00:37.254912 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:00:37.270499 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 19:00:37.283401 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:00:37.289585 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:00:37.311399 dracut-cmdline[282]: dracut-dracut-053 Feb 13 19:00:37.327822 dracut-cmdline[282]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=f06bad36699a22ae88c1968cd72b62b3503d97da521712e50a4b744320b1ba33 Feb 13 19:00:37.340370 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:00:37.343737 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:00:37.360510 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:00:37.432533 systemd-resolved[305]: Positive Trust Anchors: Feb 13 19:00:37.432561 systemd-resolved[305]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:00:37.432622 systemd-resolved[305]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:00:37.475153 kernel: SCSI subsystem initialized Feb 13 19:00:37.482259 kernel: Loading iSCSI transport class v2.0-870. Feb 13 19:00:37.495251 kernel: iscsi: registered transport (tcp) Feb 13 19:00:37.516597 kernel: iscsi: registered transport (qla4xxx) Feb 13 19:00:37.516672 kernel: QLogic iSCSI HBA Driver Feb 13 19:00:37.634155 kernel: random: crng init done Feb 13 19:00:37.634577 systemd-resolved[305]: Defaulting to hostname 'linux'. Feb 13 19:00:37.638249 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:00:37.640604 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:00:37.660659 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 19:00:37.670418 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 19:00:37.709920 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 19:00:37.709996 kernel: device-mapper: uevent: version 1.0.3 Feb 13 19:00:37.711769 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 19:00:37.776176 kernel: raid6: neonx8 gen() 6583 MB/s Feb 13 19:00:37.793150 kernel: raid6: neonx4 gen() 6556 MB/s Feb 13 19:00:37.810147 kernel: raid6: neonx2 gen() 5460 MB/s Feb 13 19:00:37.827147 kernel: raid6: neonx1 gen() 3953 MB/s Feb 13 19:00:37.844148 kernel: raid6: int64x8 gen() 3625 MB/s Feb 13 19:00:37.861147 kernel: raid6: int64x4 gen() 3725 MB/s Feb 13 19:00:37.878147 kernel: raid6: int64x2 gen() 3610 MB/s Feb 13 19:00:37.895917 kernel: raid6: int64x1 gen() 2759 MB/s Feb 13 19:00:37.895956 kernel: raid6: using algorithm neonx8 gen() 6583 MB/s Feb 13 19:00:37.913904 kernel: raid6: .... xor() 4727 MB/s, rmw enabled Feb 13 19:00:37.913941 kernel: raid6: using neon recovery algorithm Feb 13 19:00:37.921151 kernel: xor: measuring software checksum speed Feb 13 19:00:37.921203 kernel: 8regs : 11934 MB/sec Feb 13 19:00:37.923150 kernel: 32regs : 12015 MB/sec Feb 13 19:00:37.925127 kernel: arm64_neon : 8943 MB/sec Feb 13 19:00:37.925170 kernel: xor: using function: 32regs (12015 MB/sec) Feb 13 19:00:38.007162 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 19:00:38.025433 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:00:38.035434 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:00:38.079979 systemd-udevd[473]: Using default interface naming scheme 'v255'. Feb 13 19:00:38.090231 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:00:38.100406 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 19:00:38.131318 dracut-pre-trigger[478]: rd.md=0: removing MD RAID activation Feb 13 19:00:38.185279 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:00:38.196405 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:00:38.310268 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:00:38.321927 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 19:00:38.356779 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 19:00:38.373859 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:00:38.376667 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:00:38.391607 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:00:38.403945 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 19:00:38.438739 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:00:38.488237 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Feb 13 19:00:38.488300 kernel: ena 0000:00:05.0: enabling device (0010 -> 0012) Feb 13 19:00:38.524238 kernel: ena 0000:00:05.0: ENA device version: 0.10 Feb 13 19:00:38.524503 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Feb 13 19:00:38.524736 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80114000, mac addr 06:f5:6a:ca:5f:7d Feb 13 19:00:38.528959 (udev-worker)[541]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:00:38.541055 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:00:38.542553 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:00:38.550441 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:00:38.555585 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:00:38.555873 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:00:38.579500 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Feb 13 19:00:38.579538 kernel: nvme nvme0: pci function 0000:00:04.0 Feb 13 19:00:38.558614 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:00:38.592286 kernel: nvme nvme0: 2/0/0 default/read/poll queues Feb 13 19:00:38.592193 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:00:38.603969 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 19:00:38.604036 kernel: GPT:9289727 != 16777215 Feb 13 19:00:38.606967 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 19:00:38.607039 kernel: GPT:9289727 != 16777215 Feb 13 19:00:38.607066 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 19:00:38.608326 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:00:38.618188 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:00:38.629496 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:00:38.669333 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:00:38.750192 kernel: BTRFS: device fsid 4fff035f-dd55-45d8-9bb7-2a61f21b22d5 devid 1 transid 38 /dev/nvme0n1p3 scanned by (udev-worker) (540) Feb 13 19:00:38.762208 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by (udev-worker) (524) Feb 13 19:00:38.837076 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Feb 13 19:00:38.879718 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 19:00:38.905611 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Feb 13 19:00:38.927770 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Feb 13 19:00:38.933358 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Feb 13 19:00:38.944429 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 19:00:38.957442 disk-uuid[663]: Primary Header is updated. Feb 13 19:00:38.957442 disk-uuid[663]: Secondary Entries is updated. Feb 13 19:00:38.957442 disk-uuid[663]: Secondary Header is updated. Feb 13 19:00:38.968479 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:00:39.989391 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:00:39.989463 disk-uuid[664]: The operation has completed successfully. Feb 13 19:00:40.191849 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 19:00:40.193187 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 19:00:40.261370 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 19:00:40.267314 sh[924]: Success Feb 13 19:00:40.284150 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Feb 13 19:00:40.396603 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 19:00:40.401925 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 19:00:40.405575 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 19:00:40.442715 kernel: BTRFS info (device dm-0): first mount of filesystem 4fff035f-dd55-45d8-9bb7-2a61f21b22d5 Feb 13 19:00:40.442778 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:00:40.442804 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 19:00:40.444414 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 19:00:40.445650 kernel: BTRFS info (device dm-0): using free space tree Feb 13 19:00:40.570161 kernel: BTRFS info (device dm-0): enabling ssd optimizations Feb 13 19:00:40.585493 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 19:00:40.589360 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 19:00:40.600361 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 19:00:40.607806 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 19:00:40.637730 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:00:40.637810 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:00:40.639412 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 19:00:40.650144 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 19:00:40.666486 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 19:00:40.669184 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:00:40.681205 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 19:00:40.688503 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 19:00:40.784967 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:00:40.805409 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:00:40.858341 systemd-networkd[1117]: lo: Link UP Feb 13 19:00:40.858356 systemd-networkd[1117]: lo: Gained carrier Feb 13 19:00:40.863564 systemd-networkd[1117]: Enumeration completed Feb 13 19:00:40.864223 systemd-networkd[1117]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:00:40.864231 systemd-networkd[1117]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:00:40.864455 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:00:40.867132 systemd[1]: Reached target network.target - Network. Feb 13 19:00:40.875328 systemd-networkd[1117]: eth0: Link UP Feb 13 19:00:40.875336 systemd-networkd[1117]: eth0: Gained carrier Feb 13 19:00:40.875354 systemd-networkd[1117]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:00:40.897195 systemd-networkd[1117]: eth0: DHCPv4 address 172.31.18.134/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 19:00:41.081850 ignition[1034]: Ignition 2.20.0 Feb 13 19:00:41.081878 ignition[1034]: Stage: fetch-offline Feb 13 19:00:41.082368 ignition[1034]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:00:41.083383 ignition[1034]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:00:41.086575 ignition[1034]: Ignition finished successfully Feb 13 19:00:41.092023 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:00:41.108842 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Feb 13 19:00:41.130100 ignition[1128]: Ignition 2.20.0 Feb 13 19:00:41.130157 ignition[1128]: Stage: fetch Feb 13 19:00:41.131040 ignition[1128]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:00:41.131066 ignition[1128]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:00:41.131382 ignition[1128]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:00:41.142870 ignition[1128]: PUT result: OK Feb 13 19:00:41.152226 ignition[1128]: parsed url from cmdline: "" Feb 13 19:00:41.152291 ignition[1128]: no config URL provided Feb 13 19:00:41.152307 ignition[1128]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 19:00:41.152333 ignition[1128]: no config at "/usr/lib/ignition/user.ign" Feb 13 19:00:41.152366 ignition[1128]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:00:41.157738 ignition[1128]: PUT result: OK Feb 13 19:00:41.157820 ignition[1128]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Feb 13 19:00:41.161228 ignition[1128]: GET result: OK Feb 13 19:00:41.161352 ignition[1128]: parsing config with SHA512: 8bd284dba63375495700061ee651441a67fa2543c6f0c0ec97b9cccf67d1311465f03059aef5dac371d2fbfdecfc21d381ea51b7a5422690cdb8a950be02750f Feb 13 19:00:41.169617 unknown[1128]: fetched base config from "system" Feb 13 19:00:41.169635 unknown[1128]: fetched base config from "system" Feb 13 19:00:41.170660 ignition[1128]: fetch: fetch complete Feb 13 19:00:41.169648 unknown[1128]: fetched user config from "aws" Feb 13 19:00:41.170672 ignition[1128]: fetch: fetch passed Feb 13 19:00:41.170761 ignition[1128]: Ignition finished successfully Feb 13 19:00:41.182571 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Feb 13 19:00:41.208474 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 19:00:41.234909 ignition[1134]: Ignition 2.20.0 Feb 13 19:00:41.234941 ignition[1134]: Stage: kargs Feb 13 19:00:41.236586 ignition[1134]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:00:41.236615 ignition[1134]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:00:41.236784 ignition[1134]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:00:41.244223 ignition[1134]: PUT result: OK Feb 13 19:00:41.248112 ignition[1134]: kargs: kargs passed Feb 13 19:00:41.248540 ignition[1134]: Ignition finished successfully Feb 13 19:00:41.253606 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 19:00:41.265425 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 19:00:41.295487 ignition[1141]: Ignition 2.20.0 Feb 13 19:00:41.295511 ignition[1141]: Stage: disks Feb 13 19:00:41.296071 ignition[1141]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:00:41.296095 ignition[1141]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:00:41.296283 ignition[1141]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:00:41.299549 ignition[1141]: PUT result: OK Feb 13 19:00:41.309439 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 19:00:41.305985 ignition[1141]: disks: disks passed Feb 13 19:00:41.313000 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 19:00:41.306082 ignition[1141]: Ignition finished successfully Feb 13 19:00:41.315917 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 19:00:41.318309 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:00:41.320236 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:00:41.322185 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:00:41.338868 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 19:00:41.389063 systemd-fsck[1150]: ROOT: clean, 14/553520 files, 52654/553472 blocks Feb 13 19:00:41.393233 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 19:00:41.443341 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 19:00:41.535151 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 24882d04-b1a5-4a27-95f1-925956e69b18 r/w with ordered data mode. Quota mode: none. Feb 13 19:00:41.536006 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 19:00:41.539217 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 19:00:41.554309 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:00:41.566554 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 19:00:41.570773 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Feb 13 19:00:41.570862 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 19:00:41.570909 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:00:41.579233 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 19:00:41.592476 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 19:00:41.612221 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 scanned by mount (1169) Feb 13 19:00:41.617880 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:00:41.617936 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:00:41.617974 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 19:00:41.632150 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 19:00:41.633390 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:00:42.042104 initrd-setup-root[1194]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 19:00:42.050924 initrd-setup-root[1201]: cut: /sysroot/etc/group: No such file or directory Feb 13 19:00:42.059542 initrd-setup-root[1208]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 19:00:42.067914 initrd-setup-root[1215]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 19:00:42.407293 systemd-networkd[1117]: eth0: Gained IPv6LL Feb 13 19:00:42.410369 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 19:00:42.424460 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 19:00:42.434017 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 19:00:42.447406 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:00:42.448058 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 19:00:42.501230 ignition[1282]: INFO : Ignition 2.20.0 Feb 13 19:00:42.501230 ignition[1282]: INFO : Stage: mount Feb 13 19:00:42.501230 ignition[1282]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:00:42.501230 ignition[1282]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:00:42.512595 ignition[1282]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:00:42.512595 ignition[1282]: INFO : PUT result: OK Feb 13 19:00:42.503004 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 19:00:42.522198 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 19:00:42.528463 ignition[1282]: INFO : mount: mount passed Feb 13 19:00:42.528463 ignition[1282]: INFO : Ignition finished successfully Feb 13 19:00:42.536366 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 19:00:42.564308 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:00:42.589934 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/nvme0n1p6 scanned by mount (1294) Feb 13 19:00:42.589996 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:00:42.590023 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:00:42.592949 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 19:00:42.600155 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 19:00:42.601240 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:00:42.634178 ignition[1311]: INFO : Ignition 2.20.0 Feb 13 19:00:42.634178 ignition[1311]: INFO : Stage: files Feb 13 19:00:42.637485 ignition[1311]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:00:42.637485 ignition[1311]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:00:42.637485 ignition[1311]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:00:42.643872 ignition[1311]: INFO : PUT result: OK Feb 13 19:00:42.648709 ignition[1311]: DEBUG : files: compiled without relabeling support, skipping Feb 13 19:00:42.661993 ignition[1311]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 19:00:42.661993 ignition[1311]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 19:00:42.693678 ignition[1311]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 19:00:42.696349 ignition[1311]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 19:00:42.696349 ignition[1311]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 19:00:42.695002 unknown[1311]: wrote ssh authorized keys file for user: core Feb 13 19:00:42.708171 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Feb 13 19:00:42.711784 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 Feb 13 19:00:42.794564 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 19:00:43.762812 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:00:43.767071 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-arm64.raw: attempt #1 Feb 13 19:00:44.260849 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 13 19:00:44.652406 ignition[1311]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:00:44.652406 ignition[1311]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:00:44.660071 ignition[1311]: INFO : files: files passed Feb 13 19:00:44.660071 ignition[1311]: INFO : Ignition finished successfully Feb 13 19:00:44.676748 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 19:00:44.690545 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 19:00:44.702453 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 19:00:44.713968 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 19:00:44.715079 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 19:00:44.731777 initrd-setup-root-after-ignition[1340]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:00:44.731777 initrd-setup-root-after-ignition[1340]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:00:44.738258 initrd-setup-root-after-ignition[1344]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:00:44.745233 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:00:44.748995 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 19:00:44.772449 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 19:00:44.823345 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 19:00:44.823730 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 19:00:44.831271 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 19:00:44.833274 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 19:00:44.835480 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 19:00:44.843384 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 19:00:44.883192 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:00:44.893478 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 19:00:44.919854 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:00:44.922666 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:00:44.926703 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 19:00:44.931806 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 19:00:44.933802 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:00:44.938600 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 19:00:44.940819 systemd[1]: Stopped target basic.target - Basic System. Feb 13 19:00:44.945705 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 19:00:44.947967 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:00:44.954056 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 19:00:44.956384 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 19:00:44.958454 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:00:44.960964 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 19:00:44.963958 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 19:00:44.975084 systemd[1]: Stopped target swap.target - Swaps. Feb 13 19:00:44.976822 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 19:00:44.977053 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:00:44.979679 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:00:44.989605 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:00:44.991917 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 19:00:44.995953 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:00:45.001548 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 19:00:45.001774 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 19:00:45.004993 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 19:00:45.005531 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:00:45.010810 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 19:00:45.013135 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 19:00:45.028531 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 19:00:45.032347 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 19:00:45.032626 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:00:45.040649 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 19:00:45.045640 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 19:00:45.047668 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:00:45.057759 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 19:00:45.057993 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:00:45.081380 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 19:00:45.081607 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 19:00:45.090776 ignition[1364]: INFO : Ignition 2.20.0 Feb 13 19:00:45.090776 ignition[1364]: INFO : Stage: umount Feb 13 19:00:45.090776 ignition[1364]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:00:45.090776 ignition[1364]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:00:45.090776 ignition[1364]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:00:45.100585 ignition[1364]: INFO : PUT result: OK Feb 13 19:00:45.104669 ignition[1364]: INFO : umount: umount passed Feb 13 19:00:45.106350 ignition[1364]: INFO : Ignition finished successfully Feb 13 19:00:45.113187 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 19:00:45.114070 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 19:00:45.119998 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 19:00:45.123509 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 19:00:45.123690 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 19:00:45.133730 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 19:00:45.133861 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 19:00:45.136942 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 13 19:00:45.137031 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Feb 13 19:00:45.144641 systemd[1]: Stopped target network.target - Network. Feb 13 19:00:45.147003 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 19:00:45.147105 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:00:45.149404 systemd[1]: Stopped target paths.target - Path Units. Feb 13 19:00:45.151249 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 19:00:45.154742 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:00:45.157205 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 19:00:45.158961 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 19:00:45.160798 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 19:00:45.160875 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:00:45.165151 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 19:00:45.165231 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:00:45.183471 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 19:00:45.183587 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 19:00:45.187314 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 19:00:45.187398 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 19:00:45.195249 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 19:00:45.197242 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 19:00:45.199647 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 19:00:45.200231 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 19:00:45.211478 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 19:00:45.211661 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 19:00:45.222272 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 19:00:45.223418 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 19:00:45.231582 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Feb 13 19:00:45.232947 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 19:00:45.233658 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 19:00:45.243604 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Feb 13 19:00:45.247443 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 19:00:45.247562 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:00:45.261352 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 19:00:45.264721 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 19:00:45.264835 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:00:45.275478 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 19:00:45.275589 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:00:45.279746 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 19:00:45.279831 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 19:00:45.288079 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 19:00:45.288213 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:00:45.292247 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:00:45.301054 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Feb 13 19:00:45.302669 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:00:45.322494 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 19:00:45.322779 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:00:45.328789 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 19:00:45.328940 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 19:00:45.334256 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 19:00:45.334333 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:00:45.344043 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 19:00:45.344167 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:00:45.346739 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 19:00:45.346825 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 19:00:45.350427 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:00:45.350523 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:00:45.371520 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 19:00:45.375646 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 19:00:45.375766 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:00:45.378650 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:00:45.378737 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:00:45.396342 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Feb 13 19:00:45.396469 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:00:45.405337 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 19:00:45.406568 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 19:00:45.421607 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 19:00:45.422012 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 19:00:45.426796 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 19:00:45.444475 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 19:00:45.460921 systemd[1]: Switching root. Feb 13 19:00:45.509974 systemd-journald[252]: Journal stopped Feb 13 19:00:48.008325 systemd-journald[252]: Received SIGTERM from PID 1 (systemd). Feb 13 19:00:48.008469 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 19:00:48.008512 kernel: SELinux: policy capability open_perms=1 Feb 13 19:00:48.008543 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 19:00:48.008573 kernel: SELinux: policy capability always_check_network=0 Feb 13 19:00:48.008603 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 19:00:48.008639 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 19:00:48.008669 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 19:00:48.008704 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 19:00:48.008733 kernel: audit: type=1403 audit(1739473246.077:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 19:00:48.008765 systemd[1]: Successfully loaded SELinux policy in 82.035ms. Feb 13 19:00:48.008811 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 24.326ms. Feb 13 19:00:48.008845 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:00:48.008878 systemd[1]: Detected virtualization amazon. Feb 13 19:00:48.008909 systemd[1]: Detected architecture arm64. Feb 13 19:00:48.008939 systemd[1]: Detected first boot. Feb 13 19:00:48.008969 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:00:48.009004 zram_generator::config[1409]: No configuration found. Feb 13 19:00:48.009037 kernel: NET: Registered PF_VSOCK protocol family Feb 13 19:00:48.009068 systemd[1]: Populated /etc with preset unit settings. Feb 13 19:00:48.009102 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Feb 13 19:00:48.009159 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 19:00:48.009193 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 19:00:48.009225 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 19:00:48.009257 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 19:00:48.009302 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 19:00:48.009331 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 19:00:48.009360 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 19:00:48.009389 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 19:00:48.009429 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 19:00:48.009459 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 19:00:48.009489 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 19:00:48.009520 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:00:48.009550 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:00:48.009584 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 19:00:48.009613 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 19:00:48.009643 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 19:00:48.009675 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:00:48.009703 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Feb 13 19:00:48.009734 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:00:48.009764 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 19:00:48.009796 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 19:00:48.009825 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 19:00:48.009854 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 19:00:48.009886 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:00:48.009917 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:00:48.009952 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:00:48.009982 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:00:48.010011 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 19:00:48.010043 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 19:00:48.010078 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Feb 13 19:00:48.010107 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:00:48.011206 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:00:48.011247 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:00:48.011279 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 19:00:48.011308 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 19:00:48.011339 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 19:00:48.011369 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 19:00:48.011400 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 19:00:48.011470 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 19:00:48.011509 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 19:00:48.011541 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 19:00:48.011571 systemd[1]: Reached target machines.target - Containers. Feb 13 19:00:48.011603 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 19:00:48.011637 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:00:48.011668 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:00:48.011696 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 19:00:48.011727 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:00:48.011762 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:00:48.011793 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:00:48.011821 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 19:00:48.011850 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:00:48.011878 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 19:00:48.011910 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 19:00:48.011938 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 19:00:48.011966 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 19:00:48.011998 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 19:00:48.012030 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:00:48.012059 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:00:48.012087 kernel: fuse: init (API version 7.39) Feb 13 19:00:48.014153 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:00:48.014219 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 19:00:48.014250 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 19:00:48.014288 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Feb 13 19:00:48.014324 kernel: loop: module loaded Feb 13 19:00:48.014353 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:00:48.014385 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 19:00:48.014415 systemd[1]: Stopped verity-setup.service. Feb 13 19:00:48.014446 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 19:00:48.014475 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 19:00:48.014509 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 19:00:48.014538 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 19:00:48.014566 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 19:00:48.014596 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 19:00:48.014624 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:00:48.014658 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 19:00:48.014687 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 19:00:48.014716 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:00:48.014749 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:00:48.014780 kernel: ACPI: bus type drm_connector registered Feb 13 19:00:48.014862 systemd-journald[1488]: Collecting audit messages is disabled. Feb 13 19:00:48.014922 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:00:48.014951 systemd-journald[1488]: Journal started Feb 13 19:00:48.015000 systemd-journald[1488]: Runtime Journal (/run/log/journal/ec2710d9e4512ee9262a9abc383a0092) is 8M, max 75.3M, 67.3M free. Feb 13 19:00:47.477742 systemd[1]: Queued start job for default target multi-user.target. Feb 13 19:00:47.488439 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Feb 13 19:00:47.489299 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 19:00:48.026499 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:00:48.039170 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:00:48.040931 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:00:48.041403 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:00:48.045980 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 19:00:48.046556 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 19:00:48.050998 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:00:48.051654 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:00:48.054977 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:00:48.059313 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 19:00:48.062914 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Feb 13 19:00:48.076636 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 19:00:48.092194 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 19:00:48.114270 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 19:00:48.126409 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 19:00:48.138848 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 19:00:48.142322 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 19:00:48.142420 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:00:48.149097 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Feb 13 19:00:48.163819 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 19:00:48.169461 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 19:00:48.171691 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:00:48.182883 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 19:00:48.188418 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 19:00:48.191351 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:00:48.198478 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 19:00:48.201367 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:00:48.204475 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:00:48.212449 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 19:00:48.221492 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 19:00:48.228423 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 19:00:48.231010 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 19:00:48.257768 systemd-journald[1488]: Time spent on flushing to /var/log/journal/ec2710d9e4512ee9262a9abc383a0092 is 127.495ms for 916 entries. Feb 13 19:00:48.257768 systemd-journald[1488]: System Journal (/var/log/journal/ec2710d9e4512ee9262a9abc383a0092) is 8M, max 195.6M, 187.6M free. Feb 13 19:00:48.418826 systemd-journald[1488]: Received client request to flush runtime journal. Feb 13 19:00:48.418907 kernel: loop0: detected capacity change from 0 to 53784 Feb 13 19:00:48.259935 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 19:00:48.305775 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 19:00:48.308569 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 19:00:48.328696 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Feb 13 19:00:48.367209 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:00:48.386820 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 19:00:48.412275 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:00:48.426473 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 19:00:48.435581 udevadm[1555]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Feb 13 19:00:48.448860 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Feb 13 19:00:48.466234 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 19:00:48.475431 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:00:48.493030 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 19:00:48.530162 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 19:00:48.556678 systemd-tmpfiles[1561]: ACLs are not supported, ignoring. Feb 13 19:00:48.556710 systemd-tmpfiles[1561]: ACLs are not supported, ignoring. Feb 13 19:00:48.567826 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:00:48.573165 kernel: loop1: detected capacity change from 0 to 113512 Feb 13 19:00:48.715247 kernel: loop2: detected capacity change from 0 to 123192 Feb 13 19:00:48.857522 kernel: loop3: detected capacity change from 0 to 201592 Feb 13 19:00:48.919178 kernel: loop4: detected capacity change from 0 to 53784 Feb 13 19:00:48.933404 kernel: loop5: detected capacity change from 0 to 113512 Feb 13 19:00:48.950052 kernel: loop6: detected capacity change from 0 to 123192 Feb 13 19:00:48.969218 kernel: loop7: detected capacity change from 0 to 201592 Feb 13 19:00:49.000378 (sd-merge)[1569]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Feb 13 19:00:49.001456 (sd-merge)[1569]: Merged extensions into '/usr'. Feb 13 19:00:49.010064 systemd[1]: Reload requested from client PID 1544 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 19:00:49.010213 systemd[1]: Reloading... Feb 13 19:00:49.165151 zram_generator::config[1597]: No configuration found. Feb 13 19:00:49.548999 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:00:49.698382 systemd[1]: Reloading finished in 687 ms. Feb 13 19:00:49.717769 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 19:00:49.720874 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 19:00:49.741446 systemd[1]: Starting ensure-sysext.service... Feb 13 19:00:49.754411 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:00:49.761485 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:00:49.793258 systemd[1]: Reload requested from client PID 1649 ('systemctl') (unit ensure-sysext.service)... Feb 13 19:00:49.793298 systemd[1]: Reloading... Feb 13 19:00:49.848593 systemd-tmpfiles[1650]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 19:00:49.852866 systemd-tmpfiles[1650]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 19:00:49.854883 systemd-tmpfiles[1650]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 19:00:49.856259 systemd-tmpfiles[1650]: ACLs are not supported, ignoring. Feb 13 19:00:49.856408 systemd-tmpfiles[1650]: ACLs are not supported, ignoring. Feb 13 19:00:49.869079 systemd-tmpfiles[1650]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:00:49.869108 systemd-tmpfiles[1650]: Skipping /boot Feb 13 19:00:49.889196 systemd-udevd[1651]: Using default interface naming scheme 'v255'. Feb 13 19:00:49.946112 systemd-tmpfiles[1650]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:00:49.952720 systemd-tmpfiles[1650]: Skipping /boot Feb 13 19:00:49.980081 ldconfig[1539]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 19:00:49.991154 zram_generator::config[1683]: No configuration found. Feb 13 19:00:50.231175 (udev-worker)[1697]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:00:50.393135 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:00:50.401190 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (1741) Feb 13 19:00:50.611012 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Feb 13 19:00:50.611161 systemd[1]: Reloading finished in 817 ms. Feb 13 19:00:50.630362 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:00:50.636180 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 19:00:50.639014 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:00:50.723759 systemd[1]: Finished ensure-sysext.service. Feb 13 19:00:50.756100 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 19:00:50.779028 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 19:00:50.787464 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:00:50.801509 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 19:00:50.803979 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:00:50.808350 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 19:00:50.822729 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:00:50.826699 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:00:50.847537 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:00:50.856777 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:00:50.859049 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:00:50.866493 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 19:00:50.868772 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:00:50.872536 lvm[1851]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:00:50.874465 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 19:00:50.887585 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:00:50.899506 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:00:50.901514 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 19:00:50.907706 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 19:00:50.926474 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:00:50.931022 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:00:50.935219 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:00:50.938263 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:00:50.939230 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:00:50.942197 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:00:50.943207 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:00:50.951462 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:00:50.984522 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 19:00:50.993440 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 19:00:50.996935 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:00:50.998062 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:00:51.007358 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:00:51.021906 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 19:00:51.024323 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:00:51.026759 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 19:00:51.051055 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 19:00:51.065820 lvm[1887]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:00:51.103201 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 19:00:51.106755 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 19:00:51.109875 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 19:00:51.117969 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 19:00:51.130541 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 19:00:51.134991 augenrules[1898]: No rules Feb 13 19:00:51.139887 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:00:51.141070 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:00:51.172208 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 19:00:51.183837 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:00:51.196049 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 19:00:51.325533 systemd-resolved[1867]: Positive Trust Anchors: Feb 13 19:00:51.325568 systemd-resolved[1867]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:00:51.325631 systemd-resolved[1867]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:00:51.331828 systemd-networkd[1865]: lo: Link UP Feb 13 19:00:51.332487 systemd-networkd[1865]: lo: Gained carrier Feb 13 19:00:51.333243 systemd-resolved[1867]: Defaulting to hostname 'linux'. Feb 13 19:00:51.336068 systemd-networkd[1865]: Enumeration completed Feb 13 19:00:51.336222 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:00:51.339199 systemd-networkd[1865]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:00:51.339214 systemd-networkd[1865]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:00:51.339758 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:00:51.342233 systemd[1]: Reached target network.target - Network. Feb 13 19:00:51.343976 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:00:51.346306 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:00:51.348479 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 19:00:51.348548 systemd-networkd[1865]: eth0: Link UP Feb 13 19:00:51.348866 systemd-networkd[1865]: eth0: Gained carrier Feb 13 19:00:51.348900 systemd-networkd[1865]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:00:51.351675 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 19:00:51.354351 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 19:00:51.356618 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 19:00:51.359845 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 19:00:51.362640 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 19:00:51.362712 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:00:51.364589 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:00:51.367328 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 19:00:51.372111 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 19:00:51.378631 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Feb 13 19:00:51.381436 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Feb 13 19:00:51.382066 systemd-networkd[1865]: eth0: DHCPv4 address 172.31.18.134/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 19:00:51.384062 systemd[1]: Reached target ssh-access.target - SSH Access Available. Feb 13 19:00:51.392010 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 19:00:51.395567 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Feb 13 19:00:51.407420 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Feb 13 19:00:51.412032 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 19:00:51.415357 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 19:00:51.417925 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:00:51.419918 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:00:51.421819 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:00:51.421868 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:00:51.430350 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 19:00:51.443213 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Feb 13 19:00:51.451357 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 19:00:51.457371 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 19:00:51.466101 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 19:00:51.468202 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 19:00:51.475469 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 19:00:51.482949 systemd[1]: Started ntpd.service - Network Time Service. Feb 13 19:00:51.496424 jq[1923]: false Feb 13 19:00:51.497304 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 19:00:51.502347 systemd[1]: Starting setup-oem.service - Setup OEM... Feb 13 19:00:51.506448 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 19:00:51.515582 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 19:00:51.537541 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 19:00:51.542070 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 19:00:51.543020 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 19:00:51.546466 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 19:00:51.567321 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 19:00:51.581348 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Feb 13 19:00:51.589833 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 19:00:51.592222 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 19:00:51.610846 jq[1936]: true Feb 13 19:00:51.647684 update_engine[1935]: I20250213 19:00:51.646407 1935 main.cc:92] Flatcar Update Engine starting Feb 13 19:00:51.663633 jq[1947]: true Feb 13 19:00:51.668822 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 19:00:51.669276 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 19:00:51.687479 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 19:00:51.688018 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 19:00:51.710722 extend-filesystems[1924]: Found loop4 Feb 13 19:00:51.710722 extend-filesystems[1924]: Found loop5 Feb 13 19:00:51.710722 extend-filesystems[1924]: Found loop6 Feb 13 19:00:51.710722 extend-filesystems[1924]: Found loop7 Feb 13 19:00:51.710722 extend-filesystems[1924]: Found nvme0n1 Feb 13 19:00:51.710722 extend-filesystems[1924]: Found nvme0n1p1 Feb 13 19:00:51.710722 extend-filesystems[1924]: Found nvme0n1p2 Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 17:02:48 UTC 2025 (1): Starting Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: ---------------------------------------------------- Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: ntp-4 is maintained by Network Time Foundation, Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: corporation. Support and training for ntp-4 are Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: available at https://www.nwtime.org/support Feb 13 19:00:51.757425 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: ---------------------------------------------------- Feb 13 19:00:51.763913 extend-filesystems[1924]: Found nvme0n1p3 Feb 13 19:00:51.763913 extend-filesystems[1924]: Found usr Feb 13 19:00:51.763913 extend-filesystems[1924]: Found nvme0n1p4 Feb 13 19:00:51.763913 extend-filesystems[1924]: Found nvme0n1p6 Feb 13 19:00:51.763913 extend-filesystems[1924]: Found nvme0n1p7 Feb 13 19:00:51.763913 extend-filesystems[1924]: Found nvme0n1p9 Feb 13 19:00:51.763913 extend-filesystems[1924]: Checking size of /dev/nvme0n1p9 Feb 13 19:00:51.733384 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 19:00:51.756532 ntpd[1926]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 17:02:48 UTC 2025 (1): Starting Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: proto: precision = 0.096 usec (-23) Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: basedate set to 2025-02-01 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: gps base set to 2025-02-02 (week 2352) Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Listen normally on 3 eth0 172.31.18.134:123 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Listen normally on 4 lo [::1]:123 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: bind(21) AF_INET6 fe80::4f5:6aff:feca:5f7d%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: unable to create socket on eth0 (5) for fe80::4f5:6aff:feca:5f7d%2#123 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: failed to init interface for address fe80::4f5:6aff:feca:5f7d%2 Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: Listening on routing socket on fd #21 for interface updates Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:00:51.812174 ntpd[1926]: 13 Feb 19:00:51 ntpd[1926]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:00:51.775208 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 19:00:51.756578 ntpd[1926]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 19:00:51.786787 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 19:00:51.756597 ntpd[1926]: ---------------------------------------------------- Feb 13 19:00:51.833915 tar[1941]: linux-arm64/LICENSE Feb 13 19:00:51.833915 tar[1941]: linux-arm64/helm Feb 13 19:00:51.786828 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 19:00:51.756616 ntpd[1926]: ntp-4 is maintained by Network Time Foundation, Feb 13 19:00:51.852670 update_engine[1935]: I20250213 19:00:51.837822 1935 update_check_scheduler.cc:74] Next update check in 3m36s Feb 13 19:00:51.789294 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 19:00:51.756634 ntpd[1926]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 19:00:51.789331 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 19:00:51.756653 ntpd[1926]: corporation. Support and training for ntp-4 are Feb 13 19:00:51.808080 (ntainerd)[1960]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 19:00:51.756671 ntpd[1926]: available at https://www.nwtime.org/support Feb 13 19:00:51.840399 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Feb 13 19:00:51.756688 ntpd[1926]: ---------------------------------------------------- Feb 13 19:00:51.850608 systemd[1]: Started update-engine.service - Update Engine. Feb 13 19:00:51.762905 ntpd[1926]: proto: precision = 0.096 usec (-23) Feb 13 19:00:51.862407 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 19:00:51.769012 ntpd[1926]: basedate set to 2025-02-01 Feb 13 19:00:51.769048 ntpd[1926]: gps base set to 2025-02-02 (week 2352) Feb 13 19:00:51.774860 dbus-daemon[1922]: [system] SELinux support is enabled Feb 13 19:00:51.777255 ntpd[1926]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 19:00:51.777338 ntpd[1926]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 19:00:51.777621 ntpd[1926]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 19:00:51.777684 ntpd[1926]: Listen normally on 3 eth0 172.31.18.134:123 Feb 13 19:00:51.777749 ntpd[1926]: Listen normally on 4 lo [::1]:123 Feb 13 19:00:51.777826 ntpd[1926]: bind(21) AF_INET6 fe80::4f5:6aff:feca:5f7d%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 19:00:51.777864 ntpd[1926]: unable to create socket on eth0 (5) for fe80::4f5:6aff:feca:5f7d%2#123 Feb 13 19:00:51.777892 ntpd[1926]: failed to init interface for address fe80::4f5:6aff:feca:5f7d%2 Feb 13 19:00:51.777944 ntpd[1926]: Listening on routing socket on fd #21 for interface updates Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Fetch successful Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Fetch successful Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Fetch successful Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.879 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.884 INFO Fetch successful Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.884 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.886 INFO Fetch failed with 404: resource not found Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.886 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.888 INFO Fetch successful Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.888 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.888 INFO Fetch successful Feb 13 19:00:51.888498 coreos-metadata[1921]: Feb 13 19:00:51.888 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Feb 13 19:00:51.806484 ntpd[1926]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:00:51.893063 coreos-metadata[1921]: Feb 13 19:00:51.889 INFO Fetch successful Feb 13 19:00:51.893063 coreos-metadata[1921]: Feb 13 19:00:51.889 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Feb 13 19:00:51.893063 coreos-metadata[1921]: Feb 13 19:00:51.889 INFO Fetch successful Feb 13 19:00:51.893063 coreos-metadata[1921]: Feb 13 19:00:51.889 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Feb 13 19:00:51.893063 coreos-metadata[1921]: Feb 13 19:00:51.892 INFO Fetch successful Feb 13 19:00:51.806539 ntpd[1926]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:00:51.815340 dbus-daemon[1922]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.1' (uid=244 pid=1865 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Feb 13 19:00:51.910244 extend-filesystems[1924]: Resized partition /dev/nvme0n1p9 Feb 13 19:00:51.921928 extend-filesystems[1984]: resize2fs 1.47.1 (20-May-2024) Feb 13 19:00:51.943152 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Feb 13 19:00:51.954098 systemd[1]: Finished setup-oem.service - Setup OEM. Feb 13 19:00:52.022975 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Feb 13 19:00:52.026414 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 19:00:52.095728 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Feb 13 19:00:52.118196 extend-filesystems[1984]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Feb 13 19:00:52.118196 extend-filesystems[1984]: old_desc_blocks = 1, new_desc_blocks = 1 Feb 13 19:00:52.118196 extend-filesystems[1984]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Feb 13 19:00:52.134454 bash[2000]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:00:52.125835 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 19:00:52.151673 extend-filesystems[1924]: Resized filesystem in /dev/nvme0n1p9 Feb 13 19:00:52.156646 systemd[1]: Starting sshkeys.service... Feb 13 19:00:52.158652 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 19:00:52.162287 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 19:00:52.213910 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Feb 13 19:00:52.224723 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Feb 13 19:00:52.253736 systemd-logind[1932]: Watching system buttons on /dev/input/event0 (Power Button) Feb 13 19:00:52.253788 systemd-logind[1932]: Watching system buttons on /dev/input/event1 (Sleep Button) Feb 13 19:00:52.255295 systemd-logind[1932]: New seat seat0. Feb 13 19:00:52.257168 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 19:00:52.398742 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Feb 13 19:00:52.411310 dbus-daemon[1922]: [system] Successfully activated service 'org.freedesktop.hostname1' Feb 13 19:00:52.416012 dbus-daemon[1922]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=1970 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Feb 13 19:00:52.458269 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (1740) Feb 13 19:00:52.479873 systemd[1]: Starting polkit.service - Authorization Manager... Feb 13 19:00:52.509530 polkitd[2040]: Started polkitd version 121 Feb 13 19:00:52.552583 polkitd[2040]: Loading rules from directory /etc/polkit-1/rules.d Feb 13 19:00:52.552717 polkitd[2040]: Loading rules from directory /usr/share/polkit-1/rules.d Feb 13 19:00:52.559030 polkitd[2040]: Finished loading, compiling and executing 2 rules Feb 13 19:00:52.561863 dbus-daemon[1922]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Feb 13 19:00:52.562879 systemd[1]: Started polkit.service - Authorization Manager. Feb 13 19:00:52.569417 polkitd[2040]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Feb 13 19:00:52.590865 coreos-metadata[2013]: Feb 13 19:00:52.590 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 19:00:52.593071 coreos-metadata[2013]: Feb 13 19:00:52.592 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Feb 13 19:00:52.594273 coreos-metadata[2013]: Feb 13 19:00:52.593 INFO Fetch successful Feb 13 19:00:52.595841 coreos-metadata[2013]: Feb 13 19:00:52.594 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Feb 13 19:00:52.598328 coreos-metadata[2013]: Feb 13 19:00:52.598 INFO Fetch successful Feb 13 19:00:52.612962 unknown[2013]: wrote ssh authorized keys file for user: core Feb 13 19:00:52.668318 containerd[1960]: time="2025-02-13T19:00:52.668102016Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Feb 13 19:00:52.676066 systemd-hostnamed[1970]: Hostname set to (transient) Feb 13 19:00:52.676799 systemd-resolved[1867]: System hostname changed to 'ip-172-31-18-134'. Feb 13 19:00:52.685155 update-ssh-keys[2094]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:00:52.685929 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Feb 13 19:00:52.693254 systemd[1]: Finished sshkeys.service. Feb 13 19:00:52.700370 locksmithd[1972]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 19:00:52.758506 ntpd[1926]: bind(24) AF_INET6 fe80::4f5:6aff:feca:5f7d%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 19:00:52.761632 ntpd[1926]: 13 Feb 19:00:52 ntpd[1926]: bind(24) AF_INET6 fe80::4f5:6aff:feca:5f7d%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 19:00:52.761632 ntpd[1926]: 13 Feb 19:00:52 ntpd[1926]: unable to create socket on eth0 (6) for fe80::4f5:6aff:feca:5f7d%2#123 Feb 13 19:00:52.761632 ntpd[1926]: 13 Feb 19:00:52 ntpd[1926]: failed to init interface for address fe80::4f5:6aff:feca:5f7d%2 Feb 13 19:00:52.758575 ntpd[1926]: unable to create socket on eth0 (6) for fe80::4f5:6aff:feca:5f7d%2#123 Feb 13 19:00:52.758603 ntpd[1926]: failed to init interface for address fe80::4f5:6aff:feca:5f7d%2 Feb 13 19:00:52.863791 containerd[1960]: time="2025-02-13T19:00:52.863734141Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:00:52.867199 containerd[1960]: time="2025-02-13T19:00:52.867084109Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:00:52.867199 containerd[1960]: time="2025-02-13T19:00:52.867191005Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 19:00:52.867357 containerd[1960]: time="2025-02-13T19:00:52.867228577Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 19:00:52.867608 containerd[1960]: time="2025-02-13T19:00:52.867566953Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 19:00:52.867700 containerd[1960]: time="2025-02-13T19:00:52.867615757Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 19:00:52.867808 containerd[1960]: time="2025-02-13T19:00:52.867762661Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:00:52.867861 containerd[1960]: time="2025-02-13T19:00:52.867803641Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:00:52.868272 containerd[1960]: time="2025-02-13T19:00:52.868222285Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:00:52.868335 containerd[1960]: time="2025-02-13T19:00:52.868267369Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 19:00:52.868335 containerd[1960]: time="2025-02-13T19:00:52.868300093Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:00:52.868335 containerd[1960]: time="2025-02-13T19:00:52.868327129Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 19:00:52.868549 containerd[1960]: time="2025-02-13T19:00:52.868508221Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:00:52.868956 containerd[1960]: time="2025-02-13T19:00:52.868913413Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:00:52.871775 containerd[1960]: time="2025-02-13T19:00:52.871717525Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:00:52.871775 containerd[1960]: time="2025-02-13T19:00:52.871768333Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 19:00:52.872021 containerd[1960]: time="2025-02-13T19:00:52.871982005Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 19:00:52.874145 containerd[1960]: time="2025-02-13T19:00:52.872097037Z" level=info msg="metadata content store policy set" policy=shared Feb 13 19:00:52.878067 containerd[1960]: time="2025-02-13T19:00:52.878011117Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 19:00:52.880214 containerd[1960]: time="2025-02-13T19:00:52.880163593Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 19:00:52.880320 containerd[1960]: time="2025-02-13T19:00:52.880281013Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 19:00:52.880374 containerd[1960]: time="2025-02-13T19:00:52.880333717Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 19:00:52.880421 containerd[1960]: time="2025-02-13T19:00:52.880369441Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 19:00:52.880693 containerd[1960]: time="2025-02-13T19:00:52.880652437Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 19:00:52.881900 containerd[1960]: time="2025-02-13T19:00:52.881844265Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 19:00:52.882477 containerd[1960]: time="2025-02-13T19:00:52.882100621Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 19:00:52.882557 containerd[1960]: time="2025-02-13T19:00:52.882495133Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 19:00:52.882557 containerd[1960]: time="2025-02-13T19:00:52.882536341Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 19:00:52.882640 containerd[1960]: time="2025-02-13T19:00:52.882570337Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882640 containerd[1960]: time="2025-02-13T19:00:52.882599713Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882640 containerd[1960]: time="2025-02-13T19:00:52.882630865Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882785 containerd[1960]: time="2025-02-13T19:00:52.882662557Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882785 containerd[1960]: time="2025-02-13T19:00:52.882694033Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882785 containerd[1960]: time="2025-02-13T19:00:52.882724225Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882785 containerd[1960]: time="2025-02-13T19:00:52.882752017Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882785 containerd[1960]: time="2025-02-13T19:00:52.882778609Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 19:00:52.882980 containerd[1960]: time="2025-02-13T19:00:52.882820453Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.882980 containerd[1960]: time="2025-02-13T19:00:52.882851617Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.882980 containerd[1960]: time="2025-02-13T19:00:52.882879469Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.882980 containerd[1960]: time="2025-02-13T19:00:52.882909037Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.882980 containerd[1960]: time="2025-02-13T19:00:52.882937321Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.883235 containerd[1960]: time="2025-02-13T19:00:52.882990457Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.883235 containerd[1960]: time="2025-02-13T19:00:52.883023613Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.883235 containerd[1960]: time="2025-02-13T19:00:52.883056829Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.883235 containerd[1960]: time="2025-02-13T19:00:52.883085965Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.883637749Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.883691929Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.883729705Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.883761385Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.884341333Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.884403577Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.884436433Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886145 containerd[1960]: time="2025-02-13T19:00:52.884463589Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 19:00:52.886537 containerd[1960]: time="2025-02-13T19:00:52.886286821Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 19:00:52.886537 containerd[1960]: time="2025-02-13T19:00:52.886447753Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 19:00:52.886537 containerd[1960]: time="2025-02-13T19:00:52.886475713Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 19:00:52.886537 containerd[1960]: time="2025-02-13T19:00:52.886508641Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 19:00:52.886537 containerd[1960]: time="2025-02-13T19:00:52.886532641Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.886737 containerd[1960]: time="2025-02-13T19:00:52.886561237Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 19:00:52.886737 containerd[1960]: time="2025-02-13T19:00:52.886585057Z" level=info msg="NRI interface is disabled by configuration." Feb 13 19:00:52.886737 containerd[1960]: time="2025-02-13T19:00:52.886609393Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 19:00:52.889154 containerd[1960]: time="2025-02-13T19:00:52.887472265Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 19:00:52.889154 containerd[1960]: time="2025-02-13T19:00:52.887574421Z" level=info msg="Connect containerd service" Feb 13 19:00:52.889154 containerd[1960]: time="2025-02-13T19:00:52.887642857Z" level=info msg="using legacy CRI server" Feb 13 19:00:52.889154 containerd[1960]: time="2025-02-13T19:00:52.887660377Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 19:00:52.889154 containerd[1960]: time="2025-02-13T19:00:52.887905441Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 19:00:52.890004 containerd[1960]: time="2025-02-13T19:00:52.889945597Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:00:52.890420 containerd[1960]: time="2025-02-13T19:00:52.890357773Z" level=info msg="Start subscribing containerd event" Feb 13 19:00:52.890501 containerd[1960]: time="2025-02-13T19:00:52.890439697Z" level=info msg="Start recovering state" Feb 13 19:00:52.890594 containerd[1960]: time="2025-02-13T19:00:52.890557957Z" level=info msg="Start event monitor" Feb 13 19:00:52.890649 containerd[1960]: time="2025-02-13T19:00:52.890592565Z" level=info msg="Start snapshots syncer" Feb 13 19:00:52.890649 containerd[1960]: time="2025-02-13T19:00:52.890617861Z" level=info msg="Start cni network conf syncer for default" Feb 13 19:00:52.890649 containerd[1960]: time="2025-02-13T19:00:52.890639821Z" level=info msg="Start streaming server" Feb 13 19:00:52.891384 containerd[1960]: time="2025-02-13T19:00:52.891338533Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 19:00:52.891846 containerd[1960]: time="2025-02-13T19:00:52.891458953Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 19:00:52.891846 containerd[1960]: time="2025-02-13T19:00:52.891571345Z" level=info msg="containerd successfully booted in 0.231281s" Feb 13 19:00:52.891696 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 19:00:53.287297 systemd-networkd[1865]: eth0: Gained IPv6LL Feb 13 19:00:53.296201 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 19:00:53.300330 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 19:00:53.313796 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Feb 13 19:00:53.325487 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:00:53.330827 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: Initializing new seelog logger Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: New Seelog Logger Creation Complete Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 processing appconfig overrides Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 processing appconfig overrides Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 processing appconfig overrides Feb 13 19:00:53.418261 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO Proxy environment variables: Feb 13 19:00:53.424139 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.424139 amazon-ssm-agent[2129]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:00:53.424139 amazon-ssm-agent[2129]: 2025/02/13 19:00:53 processing appconfig overrides Feb 13 19:00:53.463910 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 19:00:53.520201 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO no_proxy: Feb 13 19:00:53.620411 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO https_proxy: Feb 13 19:00:53.669571 tar[1941]: linux-arm64/README.md Feb 13 19:00:53.711817 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 19:00:53.720599 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO http_proxy: Feb 13 19:00:53.817387 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO Checking if agent identity type OnPrem can be assumed Feb 13 19:00:53.915816 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO Checking if agent identity type EC2 can be assumed Feb 13 19:00:54.015920 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO Agent will take identity from EC2 Feb 13 19:00:54.118061 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 19:00:54.214264 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 19:00:54.314533 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 19:00:54.415216 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Feb 13 19:00:54.515830 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [amazon-ssm-agent] OS: linux, Arch: arm64 Feb 13 19:00:54.617020 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [amazon-ssm-agent] Starting Core Agent Feb 13 19:00:54.717459 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [amazon-ssm-agent] registrar detected. Attempting registration Feb 13 19:00:54.820191 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [Registrar] Starting registrar module Feb 13 19:00:54.920136 amazon-ssm-agent[2129]: 2025-02-13 19:00:53 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Feb 13 19:00:55.246581 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:00:55.252312 amazon-ssm-agent[2129]: 2025-02-13 19:00:55 INFO [EC2Identity] EC2 registration was successful. Feb 13 19:00:55.253038 (kubelet)[2156]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:00:55.291849 amazon-ssm-agent[2129]: 2025-02-13 19:00:55 INFO [CredentialRefresher] credentialRefresher has started Feb 13 19:00:55.291849 amazon-ssm-agent[2129]: 2025-02-13 19:00:55 INFO [CredentialRefresher] Starting credentials refresher loop Feb 13 19:00:55.291849 amazon-ssm-agent[2129]: 2025-02-13 19:00:55 INFO EC2RoleProvider Successfully connected with instance profile role credentials Feb 13 19:00:55.353514 amazon-ssm-agent[2129]: 2025-02-13 19:00:55 INFO [CredentialRefresher] Next credential rotation will be in 32.31663604926667 minutes Feb 13 19:00:55.572871 sshd_keygen[1959]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 19:00:55.615465 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 19:00:55.628953 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 19:00:55.635903 systemd[1]: Started sshd@0-172.31.18.134:22-139.178.89.65:50070.service - OpenSSH per-connection server daemon (139.178.89.65:50070). Feb 13 19:00:55.646457 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 19:00:55.648251 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 19:00:55.662470 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 19:00:55.702032 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 19:00:55.715774 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 19:00:55.721720 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Feb 13 19:00:55.726201 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 19:00:55.729475 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 19:00:55.732213 systemd[1]: Startup finished in 1.075s (kernel) + 9.265s (initrd) + 9.734s (userspace) = 20.076s. Feb 13 19:00:55.757548 ntpd[1926]: Listen normally on 7 eth0 [fe80::4f5:6aff:feca:5f7d%2]:123 Feb 13 19:00:55.758007 ntpd[1926]: 13 Feb 19:00:55 ntpd[1926]: Listen normally on 7 eth0 [fe80::4f5:6aff:feca:5f7d%2]:123 Feb 13 19:00:55.931218 sshd[2170]: Accepted publickey for core from 139.178.89.65 port 50070 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:00:55.936617 sshd-session[2170]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:00:55.963361 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 19:00:55.971760 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 19:00:55.978399 systemd-logind[1932]: New session 1 of user core. Feb 13 19:00:56.012802 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 19:00:56.024372 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 19:00:56.038637 (systemd)[2186]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 19:00:56.044422 systemd-logind[1932]: New session c1 of user core. Feb 13 19:00:56.235635 kubelet[2156]: E0213 19:00:56.235485 2156 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:00:56.239028 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:00:56.239374 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:00:56.239945 systemd[1]: kubelet.service: Consumed 1.286s CPU time, 251.8M memory peak. Feb 13 19:00:56.317182 amazon-ssm-agent[2129]: 2025-02-13 19:00:56 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Feb 13 19:00:56.350213 systemd[2186]: Queued start job for default target default.target. Feb 13 19:00:56.359584 systemd[2186]: Created slice app.slice - User Application Slice. Feb 13 19:00:56.359647 systemd[2186]: Reached target paths.target - Paths. Feb 13 19:00:56.359751 systemd[2186]: Reached target timers.target - Timers. Feb 13 19:00:56.368387 systemd[2186]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 19:00:56.386820 systemd[2186]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 19:00:56.386942 systemd[2186]: Reached target sockets.target - Sockets. Feb 13 19:00:56.387049 systemd[2186]: Reached target basic.target - Basic System. Feb 13 19:00:56.387161 systemd[2186]: Reached target default.target - Main User Target. Feb 13 19:00:56.387225 systemd[2186]: Startup finished in 329ms. Feb 13 19:00:56.387987 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 19:00:56.396658 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 19:00:56.418025 amazon-ssm-agent[2129]: 2025-02-13 19:00:56 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2196) started Feb 13 19:00:56.519065 amazon-ssm-agent[2129]: 2025-02-13 19:00:56 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Feb 13 19:00:56.561092 systemd[1]: Started sshd@1-172.31.18.134:22-139.178.89.65:49532.service - OpenSSH per-connection server daemon (139.178.89.65:49532). Feb 13 19:00:56.751592 sshd[2211]: Accepted publickey for core from 139.178.89.65 port 49532 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:00:56.754083 sshd-session[2211]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:00:56.763215 systemd-logind[1932]: New session 2 of user core. Feb 13 19:00:56.776382 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 19:00:56.903279 sshd[2213]: Connection closed by 139.178.89.65 port 49532 Feb 13 19:00:56.903034 sshd-session[2211]: pam_unix(sshd:session): session closed for user core Feb 13 19:00:56.910264 systemd[1]: sshd@1-172.31.18.134:22-139.178.89.65:49532.service: Deactivated successfully. Feb 13 19:00:56.914537 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 19:00:56.916731 systemd-logind[1932]: Session 2 logged out. Waiting for processes to exit. Feb 13 19:00:56.918630 systemd-logind[1932]: Removed session 2. Feb 13 19:00:56.944464 systemd[1]: Started sshd@2-172.31.18.134:22-139.178.89.65:49536.service - OpenSSH per-connection server daemon (139.178.89.65:49536). Feb 13 19:00:57.124960 sshd[2219]: Accepted publickey for core from 139.178.89.65 port 49536 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:00:57.127333 sshd-session[2219]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:00:57.136992 systemd-logind[1932]: New session 3 of user core. Feb 13 19:00:57.142709 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 19:00:57.264090 sshd[2221]: Connection closed by 139.178.89.65 port 49536 Feb 13 19:00:57.263878 sshd-session[2219]: pam_unix(sshd:session): session closed for user core Feb 13 19:00:57.269623 systemd-logind[1932]: Session 3 logged out. Waiting for processes to exit. Feb 13 19:00:57.270425 systemd[1]: sshd@2-172.31.18.134:22-139.178.89.65:49536.service: Deactivated successfully. Feb 13 19:00:57.274902 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 19:00:57.278207 systemd-logind[1932]: Removed session 3. Feb 13 19:00:57.303663 systemd[1]: Started sshd@3-172.31.18.134:22-139.178.89.65:49546.service - OpenSSH per-connection server daemon (139.178.89.65:49546). Feb 13 19:00:57.495635 sshd[2227]: Accepted publickey for core from 139.178.89.65 port 49546 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:00:57.497526 sshd-session[2227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:00:57.506312 systemd-logind[1932]: New session 4 of user core. Feb 13 19:00:57.517375 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 19:00:57.648156 sshd[2229]: Connection closed by 139.178.89.65 port 49546 Feb 13 19:00:57.647307 sshd-session[2227]: pam_unix(sshd:session): session closed for user core Feb 13 19:00:57.652571 systemd[1]: sshd@3-172.31.18.134:22-139.178.89.65:49546.service: Deactivated successfully. Feb 13 19:00:57.656111 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 19:00:57.659179 systemd-logind[1932]: Session 4 logged out. Waiting for processes to exit. Feb 13 19:00:57.661188 systemd-logind[1932]: Removed session 4. Feb 13 19:00:57.688584 systemd[1]: Started sshd@4-172.31.18.134:22-139.178.89.65:49548.service - OpenSSH per-connection server daemon (139.178.89.65:49548). Feb 13 19:00:57.871928 sshd[2235]: Accepted publickey for core from 139.178.89.65 port 49548 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:00:57.873864 sshd-session[2235]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:00:57.882464 systemd-logind[1932]: New session 5 of user core. Feb 13 19:00:57.892418 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 19:00:58.031001 sudo[2238]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 19:00:58.032316 sudo[2238]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:00:58.263061 systemd-resolved[1867]: Clock change detected. Flushing caches. Feb 13 19:00:58.542219 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 19:00:58.552203 (dockerd)[2254]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 19:00:59.028064 dockerd[2254]: time="2025-02-13T19:00:59.027881075Z" level=info msg="Starting up" Feb 13 19:00:59.245874 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport76918103-merged.mount: Deactivated successfully. Feb 13 19:00:59.529159 dockerd[2254]: time="2025-02-13T19:00:59.529110590Z" level=info msg="Loading containers: start." Feb 13 19:00:59.803747 kernel: Initializing XFRM netlink socket Feb 13 19:00:59.849740 (udev-worker)[2280]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:00:59.951514 systemd-networkd[1865]: docker0: Link UP Feb 13 19:00:59.986051 dockerd[2254]: time="2025-02-13T19:00:59.985981468Z" level=info msg="Loading containers: done." Feb 13 19:01:00.015746 dockerd[2254]: time="2025-02-13T19:01:00.015363852Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 19:01:00.015746 dockerd[2254]: time="2025-02-13T19:01:00.015501768Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Feb 13 19:01:00.016018 dockerd[2254]: time="2025-02-13T19:01:00.015750288Z" level=info msg="Daemon has completed initialization" Feb 13 19:01:00.074719 dockerd[2254]: time="2025-02-13T19:01:00.074397444Z" level=info msg="API listen on /run/docker.sock" Feb 13 19:01:00.075286 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 19:01:00.236219 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck635425267-merged.mount: Deactivated successfully. Feb 13 19:01:01.002041 containerd[1960]: time="2025-02-13T19:01:01.001979869Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\"" Feb 13 19:01:01.669171 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount768729057.mount: Deactivated successfully. Feb 13 19:01:03.453016 containerd[1960]: time="2025-02-13T19:01:03.452939261Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:03.455032 containerd[1960]: time="2025-02-13T19:01:03.454966865Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.2: active requests=0, bytes read=26218236" Feb 13 19:01:03.456017 containerd[1960]: time="2025-02-13T19:01:03.455936249Z" level=info msg="ImageCreate event name:\"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:03.461664 containerd[1960]: time="2025-02-13T19:01:03.461581889Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:03.464206 containerd[1960]: time="2025-02-13T19:01:03.463933997Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.2\" with image id \"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\", size \"26215036\" in 2.461293696s" Feb 13 19:01:03.464206 containerd[1960]: time="2025-02-13T19:01:03.463988621Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\" returns image reference \"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\"" Feb 13 19:01:03.465155 containerd[1960]: time="2025-02-13T19:01:03.464967773Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\"" Feb 13 19:01:05.624062 containerd[1960]: time="2025-02-13T19:01:05.623995616Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:05.626063 containerd[1960]: time="2025-02-13T19:01:05.625996964Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.2: active requests=0, bytes read=22528145" Feb 13 19:01:05.626784 containerd[1960]: time="2025-02-13T19:01:05.626704232Z" level=info msg="ImageCreate event name:\"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:05.632170 containerd[1960]: time="2025-02-13T19:01:05.632087624Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:05.634605 containerd[1960]: time="2025-02-13T19:01:05.634391636Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.2\" with image id \"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\", size \"23968941\" in 2.169081839s" Feb 13 19:01:05.634605 containerd[1960]: time="2025-02-13T19:01:05.634448960Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\" returns image reference \"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\"" Feb 13 19:01:05.635362 containerd[1960]: time="2025-02-13T19:01:05.635315000Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\"" Feb 13 19:01:05.950460 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 19:01:05.962024 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:01:06.456010 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:06.464446 (kubelet)[2511]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:01:06.543169 kubelet[2511]: E0213 19:01:06.542968 2511 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:01:06.552530 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:01:06.552913 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:01:06.554062 systemd[1]: kubelet.service: Consumed 287ms CPU time, 106.1M memory peak. Feb 13 19:01:07.667746 containerd[1960]: time="2025-02-13T19:01:07.666534430Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:07.669118 containerd[1960]: time="2025-02-13T19:01:07.669044410Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.2: active requests=0, bytes read=17480800" Feb 13 19:01:07.671210 containerd[1960]: time="2025-02-13T19:01:07.671157010Z" level=info msg="ImageCreate event name:\"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:07.678555 containerd[1960]: time="2025-02-13T19:01:07.678503626Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:07.680821 containerd[1960]: time="2025-02-13T19:01:07.680762566Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.2\" with image id \"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\", size \"18921614\" in 2.045388994s" Feb 13 19:01:07.680821 containerd[1960]: time="2025-02-13T19:01:07.680815894Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\" returns image reference \"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\"" Feb 13 19:01:07.681481 containerd[1960]: time="2025-02-13T19:01:07.681432046Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\"" Feb 13 19:01:09.074122 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3489958984.mount: Deactivated successfully. Feb 13 19:01:09.620980 containerd[1960]: time="2025-02-13T19:01:09.620899416Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:09.622297 containerd[1960]: time="2025-02-13T19:01:09.622211052Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.2: active requests=0, bytes read=27363382" Feb 13 19:01:09.623407 containerd[1960]: time="2025-02-13T19:01:09.623337144Z" level=info msg="ImageCreate event name:\"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:09.627000 containerd[1960]: time="2025-02-13T19:01:09.626913816Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:09.628982 containerd[1960]: time="2025-02-13T19:01:09.628407984Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.2\" with image id \"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\", repo tag \"registry.k8s.io/kube-proxy:v1.32.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\", size \"27362401\" in 1.94691979s" Feb 13 19:01:09.628982 containerd[1960]: time="2025-02-13T19:01:09.628465620Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\" returns image reference \"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\"" Feb 13 19:01:09.629586 containerd[1960]: time="2025-02-13T19:01:09.629408676Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Feb 13 19:01:10.259674 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2321158793.mount: Deactivated successfully. Feb 13 19:01:11.677545 containerd[1960]: time="2025-02-13T19:01:11.676720082Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:11.702701 containerd[1960]: time="2025-02-13T19:01:11.702590450Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951622" Feb 13 19:01:11.723959 containerd[1960]: time="2025-02-13T19:01:11.723874718Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:11.749431 containerd[1960]: time="2025-02-13T19:01:11.749330690Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:11.755657 containerd[1960]: time="2025-02-13T19:01:11.753265287Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 2.123795555s" Feb 13 19:01:11.756139 containerd[1960]: time="2025-02-13T19:01:11.753331863Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Feb 13 19:01:11.758958 containerd[1960]: time="2025-02-13T19:01:11.758877543Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Feb 13 19:01:12.238915 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2298774428.mount: Deactivated successfully. Feb 13 19:01:12.250390 containerd[1960]: time="2025-02-13T19:01:12.250057309Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:12.251422 containerd[1960]: time="2025-02-13T19:01:12.251334505Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Feb 13 19:01:12.252375 containerd[1960]: time="2025-02-13T19:01:12.252292309Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:12.259328 containerd[1960]: time="2025-02-13T19:01:12.259232065Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:12.261070 containerd[1960]: time="2025-02-13T19:01:12.260814877Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 501.878462ms" Feb 13 19:01:12.261070 containerd[1960]: time="2025-02-13T19:01:12.260868733Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Feb 13 19:01:12.261959 containerd[1960]: time="2025-02-13T19:01:12.261653761Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Feb 13 19:01:12.927138 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1465493803.mount: Deactivated successfully. Feb 13 19:01:16.652077 containerd[1960]: time="2025-02-13T19:01:16.651666583Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:16.654009 containerd[1960]: time="2025-02-13T19:01:16.653928211Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67812429" Feb 13 19:01:16.654578 containerd[1960]: time="2025-02-13T19:01:16.654453571Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:16.660639 containerd[1960]: time="2025-02-13T19:01:16.660543271Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:16.663229 containerd[1960]: time="2025-02-13T19:01:16.663182155Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 4.401461182s" Feb 13 19:01:16.663545 containerd[1960]: time="2025-02-13T19:01:16.663396607Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" Feb 13 19:01:16.700287 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 13 19:01:16.715789 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:01:17.054180 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:17.059485 (kubelet)[2660]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:01:17.150179 kubelet[2660]: E0213 19:01:17.149940 2660 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:01:17.153257 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:01:17.153556 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:01:17.157895 systemd[1]: kubelet.service: Consumed 286ms CPU time, 100.6M memory peak. Feb 13 19:01:22.200060 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Feb 13 19:01:24.407020 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:24.407372 systemd[1]: kubelet.service: Consumed 286ms CPU time, 100.6M memory peak. Feb 13 19:01:24.420169 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:01:24.490150 systemd[1]: Reload requested from client PID 2688 ('systemctl') (unit session-5.scope)... Feb 13 19:01:24.490373 systemd[1]: Reloading... Feb 13 19:01:24.728721 zram_generator::config[2736]: No configuration found. Feb 13 19:01:24.964755 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:01:25.189446 systemd[1]: Reloading finished in 698 ms. Feb 13 19:01:25.281985 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:25.292002 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:01:25.295095 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:01:25.295503 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:25.295567 systemd[1]: kubelet.service: Consumed 202ms CPU time, 90.1M memory peak. Feb 13 19:01:25.305174 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:01:25.785011 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:25.786537 (kubelet)[2798]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:01:25.862048 kubelet[2798]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:01:25.862048 kubelet[2798]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:01:25.862048 kubelet[2798]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:01:25.862613 kubelet[2798]: I0213 19:01:25.862182 2798 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:01:26.715854 kubelet[2798]: I0213 19:01:26.715785 2798 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:01:26.715854 kubelet[2798]: I0213 19:01:26.715837 2798 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:01:26.716467 kubelet[2798]: I0213 19:01:26.716420 2798 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:01:26.752712 kubelet[2798]: E0213 19:01:26.752623 2798 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.18.134:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:26.755571 kubelet[2798]: I0213 19:01:26.755365 2798 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:01:26.767907 kubelet[2798]: E0213 19:01:26.767839 2798 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:01:26.768426 kubelet[2798]: I0213 19:01:26.768097 2798 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:01:26.776511 kubelet[2798]: I0213 19:01:26.776426 2798 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:01:26.778203 kubelet[2798]: I0213 19:01:26.778108 2798 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:01:26.779614 kubelet[2798]: I0213 19:01:26.778204 2798 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-18-134","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:01:26.779614 kubelet[2798]: I0213 19:01:26.779230 2798 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:01:26.779614 kubelet[2798]: I0213 19:01:26.779254 2798 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:01:26.779614 kubelet[2798]: I0213 19:01:26.779495 2798 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:01:26.785340 kubelet[2798]: I0213 19:01:26.785280 2798 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:01:26.785533 kubelet[2798]: I0213 19:01:26.785491 2798 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:01:26.785600 kubelet[2798]: I0213 19:01:26.785543 2798 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:01:26.785600 kubelet[2798]: I0213 19:01:26.785565 2798 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:01:26.790422 kubelet[2798]: W0213 19:01:26.790314 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.18.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-18-134&limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:26.790563 kubelet[2798]: E0213 19:01:26.790431 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.18.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-18-134&limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:26.791370 kubelet[2798]: W0213 19:01:26.791061 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.18.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:26.791370 kubelet[2798]: E0213 19:01:26.791148 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.18.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:26.791370 kubelet[2798]: I0213 19:01:26.791385 2798 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:01:26.793737 kubelet[2798]: I0213 19:01:26.792168 2798 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:01:26.793737 kubelet[2798]: W0213 19:01:26.792291 2798 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 19:01:26.793982 kubelet[2798]: I0213 19:01:26.793936 2798 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:01:26.794046 kubelet[2798]: I0213 19:01:26.794002 2798 server.go:1287] "Started kubelet" Feb 13 19:01:26.802245 kubelet[2798]: I0213 19:01:26.802183 2798 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:01:26.807754 kubelet[2798]: E0213 19:01:26.807478 2798 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.18.134:6443/api/v1/namespaces/default/events\": dial tcp 172.31.18.134:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-18-134.1823d9bcfba2cc21 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-18-134,UID:ip-172-31-18-134,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-18-134,},FirstTimestamp:2025-02-13 19:01:26.793972769 +0000 UTC m=+0.999581694,LastTimestamp:2025-02-13 19:01:26.793972769 +0000 UTC m=+0.999581694,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-18-134,}" Feb 13 19:01:26.810789 kubelet[2798]: E0213 19:01:26.810751 2798 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:01:26.813371 kubelet[2798]: I0213 19:01:26.813323 2798 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:01:26.813832 kubelet[2798]: E0213 19:01:26.813791 2798 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-18-134\" not found" Feb 13 19:01:26.814110 kubelet[2798]: I0213 19:01:26.814065 2798 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:01:26.816304 kubelet[2798]: I0213 19:01:26.816263 2798 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:01:26.818054 kubelet[2798]: I0213 19:01:26.817978 2798 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:01:26.818501 kubelet[2798]: I0213 19:01:26.818475 2798 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:01:26.819014 kubelet[2798]: I0213 19:01:26.818980 2798 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:01:26.819851 kubelet[2798]: I0213 19:01:26.819805 2798 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:01:26.819952 kubelet[2798]: I0213 19:01:26.819900 2798 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:01:26.820431 kubelet[2798]: I0213 19:01:26.820401 2798 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:01:26.820769 kubelet[2798]: I0213 19:01:26.820737 2798 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:01:26.822506 kubelet[2798]: E0213 19:01:26.821390 2798 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.18.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-18-134?timeout=10s\": dial tcp 172.31.18.134:6443: connect: connection refused" interval="200ms" Feb 13 19:01:26.822506 kubelet[2798]: W0213 19:01:26.822386 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.18.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:26.822506 kubelet[2798]: E0213 19:01:26.822462 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.18.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:26.826859 kubelet[2798]: I0213 19:01:26.826313 2798 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:01:26.847057 kubelet[2798]: I0213 19:01:26.846995 2798 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:01:26.850715 kubelet[2798]: I0213 19:01:26.849246 2798 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:01:26.850715 kubelet[2798]: I0213 19:01:26.849306 2798 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:01:26.850715 kubelet[2798]: I0213 19:01:26.849397 2798 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:01:26.850715 kubelet[2798]: I0213 19:01:26.849414 2798 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:01:26.850715 kubelet[2798]: E0213 19:01:26.849498 2798 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:01:26.851230 kubelet[2798]: W0213 19:01:26.851148 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.18.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:26.851318 kubelet[2798]: E0213 19:01:26.851244 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.18.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:26.865881 kubelet[2798]: I0213 19:01:26.865842 2798 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:01:26.868880 kubelet[2798]: I0213 19:01:26.868838 2798 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:01:26.869433 kubelet[2798]: I0213 19:01:26.869093 2798 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:01:26.872522 kubelet[2798]: I0213 19:01:26.872406 2798 policy_none.go:49] "None policy: Start" Feb 13 19:01:26.872522 kubelet[2798]: I0213 19:01:26.872457 2798 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:01:26.872522 kubelet[2798]: I0213 19:01:26.872481 2798 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:01:26.882471 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 19:01:26.903115 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 19:01:26.910702 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 19:01:26.914936 kubelet[2798]: E0213 19:01:26.914869 2798 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-18-134\" not found" Feb 13 19:01:26.917406 kubelet[2798]: I0213 19:01:26.917362 2798 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:01:26.917809 kubelet[2798]: I0213 19:01:26.917649 2798 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:01:26.917809 kubelet[2798]: I0213 19:01:26.917697 2798 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:01:26.919653 kubelet[2798]: I0213 19:01:26.919420 2798 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:01:26.921330 kubelet[2798]: E0213 19:01:26.920828 2798 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:01:26.921330 kubelet[2798]: E0213 19:01:26.920943 2798 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-18-134\" not found" Feb 13 19:01:26.968752 systemd[1]: Created slice kubepods-burstable-pod341bfc7205109c3c0851cbc288f8f89e.slice - libcontainer container kubepods-burstable-pod341bfc7205109c3c0851cbc288f8f89e.slice. Feb 13 19:01:26.985231 kubelet[2798]: E0213 19:01:26.985177 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:26.992543 systemd[1]: Created slice kubepods-burstable-pod726ad03d2d1806d3cf7e153f842f8768.slice - libcontainer container kubepods-burstable-pod726ad03d2d1806d3cf7e153f842f8768.slice. Feb 13 19:01:27.003257 kubelet[2798]: E0213 19:01:27.003193 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:27.010272 systemd[1]: Created slice kubepods-burstable-pod8d30d2da5ef6b852f1bf3b4dabd56ad7.slice - libcontainer container kubepods-burstable-pod8d30d2da5ef6b852f1bf3b4dabd56ad7.slice. Feb 13 19:01:27.014104 kubelet[2798]: E0213 19:01:27.014062 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:27.020880 kubelet[2798]: I0213 19:01:27.020081 2798 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-18-134" Feb 13 19:01:27.020880 kubelet[2798]: I0213 19:01:27.020455 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/341bfc7205109c3c0851cbc288f8f89e-ca-certs\") pod \"kube-apiserver-ip-172-31-18-134\" (UID: \"341bfc7205109c3c0851cbc288f8f89e\") " pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:27.020880 kubelet[2798]: I0213 19:01:27.020501 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:27.020880 kubelet[2798]: I0213 19:01:27.020546 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-kubeconfig\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:27.020880 kubelet[2798]: I0213 19:01:27.020582 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/341bfc7205109c3c0851cbc288f8f89e-k8s-certs\") pod \"kube-apiserver-ip-172-31-18-134\" (UID: \"341bfc7205109c3c0851cbc288f8f89e\") " pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:27.020880 kubelet[2798]: I0213 19:01:27.020621 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/341bfc7205109c3c0851cbc288f8f89e-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-18-134\" (UID: \"341bfc7205109c3c0851cbc288f8f89e\") " pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:27.021395 kubelet[2798]: I0213 19:01:27.020660 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-ca-certs\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:27.021395 kubelet[2798]: E0213 19:01:27.020665 2798 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.18.134:6443/api/v1/nodes\": dial tcp 172.31.18.134:6443: connect: connection refused" node="ip-172-31-18-134" Feb 13 19:01:27.021395 kubelet[2798]: I0213 19:01:27.020740 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-k8s-certs\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:27.021395 kubelet[2798]: I0213 19:01:27.020787 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:27.021395 kubelet[2798]: I0213 19:01:27.020828 2798 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8d30d2da5ef6b852f1bf3b4dabd56ad7-kubeconfig\") pod \"kube-scheduler-ip-172-31-18-134\" (UID: \"8d30d2da5ef6b852f1bf3b4dabd56ad7\") " pod="kube-system/kube-scheduler-ip-172-31-18-134" Feb 13 19:01:27.022400 kubelet[2798]: E0213 19:01:27.022349 2798 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.18.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-18-134?timeout=10s\": dial tcp 172.31.18.134:6443: connect: connection refused" interval="400ms" Feb 13 19:01:27.224221 kubelet[2798]: I0213 19:01:27.223476 2798 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-18-134" Feb 13 19:01:27.224221 kubelet[2798]: E0213 19:01:27.223954 2798 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.18.134:6443/api/v1/nodes\": dial tcp 172.31.18.134:6443: connect: connection refused" node="ip-172-31-18-134" Feb 13 19:01:27.286987 containerd[1960]: time="2025-02-13T19:01:27.286908160Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-18-134,Uid:341bfc7205109c3c0851cbc288f8f89e,Namespace:kube-system,Attempt:0,}" Feb 13 19:01:27.305286 containerd[1960]: time="2025-02-13T19:01:27.305218240Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-18-134,Uid:726ad03d2d1806d3cf7e153f842f8768,Namespace:kube-system,Attempt:0,}" Feb 13 19:01:27.316612 containerd[1960]: time="2025-02-13T19:01:27.316532320Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-18-134,Uid:8d30d2da5ef6b852f1bf3b4dabd56ad7,Namespace:kube-system,Attempt:0,}" Feb 13 19:01:27.423549 kubelet[2798]: E0213 19:01:27.423489 2798 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.18.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-18-134?timeout=10s\": dial tcp 172.31.18.134:6443: connect: connection refused" interval="800ms" Feb 13 19:01:27.627293 kubelet[2798]: I0213 19:01:27.627070 2798 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-18-134" Feb 13 19:01:27.627946 kubelet[2798]: E0213 19:01:27.627887 2798 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.18.134:6443/api/v1/nodes\": dial tcp 172.31.18.134:6443: connect: connection refused" node="ip-172-31-18-134" Feb 13 19:01:27.738269 kubelet[2798]: W0213 19:01:27.738174 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.18.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-18-134&limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:27.738424 kubelet[2798]: E0213 19:01:27.738279 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.18.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-18-134&limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:27.787300 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount420340681.mount: Deactivated successfully. Feb 13 19:01:27.794346 containerd[1960]: time="2025-02-13T19:01:27.794267346Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:01:27.798962 containerd[1960]: time="2025-02-13T19:01:27.798874782Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Feb 13 19:01:27.801317 containerd[1960]: time="2025-02-13T19:01:27.801249258Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:01:27.803986 containerd[1960]: time="2025-02-13T19:01:27.803910222Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:01:27.806538 containerd[1960]: time="2025-02-13T19:01:27.806315694Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:01:27.806538 containerd[1960]: time="2025-02-13T19:01:27.806464434Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:01:27.806866 containerd[1960]: time="2025-02-13T19:01:27.806813286Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:01:27.812347 containerd[1960]: time="2025-02-13T19:01:27.812267898Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:01:27.815949 containerd[1960]: time="2025-02-13T19:01:27.815523498Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 528.500666ms" Feb 13 19:01:27.817376 containerd[1960]: time="2025-02-13T19:01:27.817316874Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 500.662814ms" Feb 13 19:01:27.822121 containerd[1960]: time="2025-02-13T19:01:27.822047502Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 516.71171ms" Feb 13 19:01:27.999101 containerd[1960]: time="2025-02-13T19:01:27.998878111Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:01:27.999101 containerd[1960]: time="2025-02-13T19:01:27.998989267Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:01:27.999101 containerd[1960]: time="2025-02-13T19:01:27.999015823Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:27.999534 containerd[1960]: time="2025-02-13T19:01:27.999148687Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:28.003807 containerd[1960]: time="2025-02-13T19:01:28.003595923Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:01:28.005104 containerd[1960]: time="2025-02-13T19:01:28.004993731Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:01:28.005467 containerd[1960]: time="2025-02-13T19:01:28.005397831Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:28.005876 containerd[1960]: time="2025-02-13T19:01:28.005806167Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:28.008834 containerd[1960]: time="2025-02-13T19:01:28.007320447Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:01:28.008834 containerd[1960]: time="2025-02-13T19:01:28.007428987Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:01:28.008834 containerd[1960]: time="2025-02-13T19:01:28.007464687Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:28.008834 containerd[1960]: time="2025-02-13T19:01:28.007606887Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:28.028108 kubelet[2798]: W0213 19:01:28.027912 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.18.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:28.028108 kubelet[2798]: E0213 19:01:28.028018 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.18.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:28.060049 systemd[1]: Started cri-containerd-6b087e01d9395e3ad8133445563511089c5a2d5621ed55de4f788e10e56d2e41.scope - libcontainer container 6b087e01d9395e3ad8133445563511089c5a2d5621ed55de4f788e10e56d2e41. Feb 13 19:01:28.070988 systemd[1]: Started cri-containerd-9dcdf7d091fe56e230053e1dd51d581c97b3b3efd01ec43e0dc46f784e0fffc5.scope - libcontainer container 9dcdf7d091fe56e230053e1dd51d581c97b3b3efd01ec43e0dc46f784e0fffc5. Feb 13 19:01:28.085021 systemd[1]: Started cri-containerd-9467c42bb09652a0a16192b7309ea87d656aa85b1dec0350ac09727e53b85134.scope - libcontainer container 9467c42bb09652a0a16192b7309ea87d656aa85b1dec0350ac09727e53b85134. Feb 13 19:01:28.090289 kubelet[2798]: E0213 19:01:28.090133 2798 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.18.134:6443/api/v1/namespaces/default/events\": dial tcp 172.31.18.134:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-18-134.1823d9bcfba2cc21 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-18-134,UID:ip-172-31-18-134,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-18-134,},FirstTimestamp:2025-02-13 19:01:26.793972769 +0000 UTC m=+0.999581694,LastTimestamp:2025-02-13 19:01:26.793972769 +0000 UTC m=+0.999581694,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-18-134,}" Feb 13 19:01:28.202380 containerd[1960]: time="2025-02-13T19:01:28.201815104Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-18-134,Uid:8d30d2da5ef6b852f1bf3b4dabd56ad7,Namespace:kube-system,Attempt:0,} returns sandbox id \"9467c42bb09652a0a16192b7309ea87d656aa85b1dec0350ac09727e53b85134\"" Feb 13 19:01:28.203153 containerd[1960]: time="2025-02-13T19:01:28.202560640Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-18-134,Uid:341bfc7205109c3c0851cbc288f8f89e,Namespace:kube-system,Attempt:0,} returns sandbox id \"6b087e01d9395e3ad8133445563511089c5a2d5621ed55de4f788e10e56d2e41\"" Feb 13 19:01:28.209098 containerd[1960]: time="2025-02-13T19:01:28.209033560Z" level=info msg="CreateContainer within sandbox \"6b087e01d9395e3ad8133445563511089c5a2d5621ed55de4f788e10e56d2e41\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 19:01:28.215297 containerd[1960]: time="2025-02-13T19:01:28.215104864Z" level=info msg="CreateContainer within sandbox \"9467c42bb09652a0a16192b7309ea87d656aa85b1dec0350ac09727e53b85134\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 19:01:28.216717 containerd[1960]: time="2025-02-13T19:01:28.216263908Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-18-134,Uid:726ad03d2d1806d3cf7e153f842f8768,Namespace:kube-system,Attempt:0,} returns sandbox id \"9dcdf7d091fe56e230053e1dd51d581c97b3b3efd01ec43e0dc46f784e0fffc5\"" Feb 13 19:01:28.224229 containerd[1960]: time="2025-02-13T19:01:28.224156560Z" level=info msg="CreateContainer within sandbox \"9dcdf7d091fe56e230053e1dd51d581c97b3b3efd01ec43e0dc46f784e0fffc5\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 19:01:28.224803 kubelet[2798]: E0213 19:01:28.224760 2798 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.18.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-18-134?timeout=10s\": dial tcp 172.31.18.134:6443: connect: connection refused" interval="1.6s" Feb 13 19:01:28.243495 containerd[1960]: time="2025-02-13T19:01:28.243364372Z" level=info msg="CreateContainer within sandbox \"6b087e01d9395e3ad8133445563511089c5a2d5621ed55de4f788e10e56d2e41\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"38537405aeedb6bfb79a99fb29974c25276ce3284fd582468cd22a038c33df7c\"" Feb 13 19:01:28.244678 containerd[1960]: time="2025-02-13T19:01:28.244624036Z" level=info msg="StartContainer for \"38537405aeedb6bfb79a99fb29974c25276ce3284fd582468cd22a038c33df7c\"" Feb 13 19:01:28.245398 containerd[1960]: time="2025-02-13T19:01:28.245225344Z" level=info msg="CreateContainer within sandbox \"9467c42bb09652a0a16192b7309ea87d656aa85b1dec0350ac09727e53b85134\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c\"" Feb 13 19:01:28.246212 containerd[1960]: time="2025-02-13T19:01:28.246151276Z" level=info msg="StartContainer for \"528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c\"" Feb 13 19:01:28.253999 containerd[1960]: time="2025-02-13T19:01:28.253822300Z" level=info msg="CreateContainer within sandbox \"9dcdf7d091fe56e230053e1dd51d581c97b3b3efd01ec43e0dc46f784e0fffc5\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3\"" Feb 13 19:01:28.256976 containerd[1960]: time="2025-02-13T19:01:28.256803448Z" level=info msg="StartContainer for \"5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3\"" Feb 13 19:01:28.265217 kubelet[2798]: W0213 19:01:28.265060 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.18.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:28.265509 kubelet[2798]: E0213 19:01:28.265185 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.18.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:28.319995 systemd[1]: Started cri-containerd-38537405aeedb6bfb79a99fb29974c25276ce3284fd582468cd22a038c33df7c.scope - libcontainer container 38537405aeedb6bfb79a99fb29974c25276ce3284fd582468cd22a038c33df7c. Feb 13 19:01:28.328019 systemd[1]: Started cri-containerd-528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c.scope - libcontainer container 528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c. Feb 13 19:01:28.345040 kubelet[2798]: W0213 19:01:28.344857 2798 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.18.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.18.134:6443: connect: connection refused Feb 13 19:01:28.345040 kubelet[2798]: E0213 19:01:28.344971 2798 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.18.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.18.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:01:28.347373 systemd[1]: Started cri-containerd-5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3.scope - libcontainer container 5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3. Feb 13 19:01:28.434637 kubelet[2798]: I0213 19:01:28.434342 2798 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-18-134" Feb 13 19:01:28.437736 kubelet[2798]: E0213 19:01:28.436808 2798 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.18.134:6443/api/v1/nodes\": dial tcp 172.31.18.134:6443: connect: connection refused" node="ip-172-31-18-134" Feb 13 19:01:28.453198 containerd[1960]: time="2025-02-13T19:01:28.453132677Z" level=info msg="StartContainer for \"528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c\" returns successfully" Feb 13 19:01:28.466742 containerd[1960]: time="2025-02-13T19:01:28.465242658Z" level=info msg="StartContainer for \"38537405aeedb6bfb79a99fb29974c25276ce3284fd582468cd22a038c33df7c\" returns successfully" Feb 13 19:01:28.490039 containerd[1960]: time="2025-02-13T19:01:28.489938262Z" level=info msg="StartContainer for \"5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3\" returns successfully" Feb 13 19:01:28.873462 kubelet[2798]: E0213 19:01:28.873392 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:28.886982 kubelet[2798]: E0213 19:01:28.886944 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:28.887566 kubelet[2798]: E0213 19:01:28.887532 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:29.895591 kubelet[2798]: E0213 19:01:29.895536 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:29.898211 kubelet[2798]: E0213 19:01:29.898162 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:30.041854 kubelet[2798]: I0213 19:01:30.041802 2798 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-18-134" Feb 13 19:01:31.753714 kubelet[2798]: E0213 19:01:31.753453 2798 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-18-134\" not found" node="ip-172-31-18-134" Feb 13 19:01:32.292571 kubelet[2798]: I0213 19:01:32.290368 2798 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-18-134" Feb 13 19:01:32.292571 kubelet[2798]: E0213 19:01:32.290422 2798 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"ip-172-31-18-134\": node \"ip-172-31-18-134\" not found" Feb 13 19:01:32.315153 kubelet[2798]: I0213 19:01:32.315010 2798 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-18-134" Feb 13 19:01:32.369310 kubelet[2798]: E0213 19:01:32.369233 2798 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-node-lease\" not found" interval="3.2s" Feb 13 19:01:32.384151 kubelet[2798]: E0213 19:01:32.383591 2798 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-18-134\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-18-134" Feb 13 19:01:32.384151 kubelet[2798]: I0213 19:01:32.383639 2798 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:32.392101 kubelet[2798]: E0213 19:01:32.392057 2798 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-18-134\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:32.392465 kubelet[2798]: I0213 19:01:32.392303 2798 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:32.396555 kubelet[2798]: E0213 19:01:32.396493 2798 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-18-134\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:32.794756 kubelet[2798]: I0213 19:01:32.794697 2798 apiserver.go:52] "Watching apiserver" Feb 13 19:01:32.820447 kubelet[2798]: I0213 19:01:32.820377 2798 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:01:33.154308 kubelet[2798]: I0213 19:01:33.154156 2798 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-18-134" Feb 13 19:01:34.794535 systemd[1]: Reload requested from client PID 3079 ('systemctl') (unit session-5.scope)... Feb 13 19:01:34.795057 systemd[1]: Reloading... Feb 13 19:01:35.078806 kubelet[2798]: I0213 19:01:35.078032 2798 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:35.116769 zram_generator::config[3125]: No configuration found. Feb 13 19:01:35.416663 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:01:35.679584 systemd[1]: Reloading finished in 883 ms. Feb 13 19:01:35.742181 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:01:35.761462 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:01:35.761996 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:35.762077 systemd[1]: kubelet.service: Consumed 1.698s CPU time, 124.6M memory peak. Feb 13 19:01:35.776210 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:01:36.162795 update_engine[1935]: I20250213 19:01:36.162727 1935 update_attempter.cc:509] Updating boot flags... Feb 13 19:01:36.196568 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:01:36.220346 (kubelet)[3189]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:01:36.416544 kubelet[3189]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:01:36.416544 kubelet[3189]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:01:36.416544 kubelet[3189]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:01:36.416544 kubelet[3189]: I0213 19:01:36.415058 3189 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:01:36.430737 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (3204) Feb 13 19:01:36.437738 kubelet[3189]: I0213 19:01:36.437026 3189 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:01:36.437738 kubelet[3189]: I0213 19:01:36.437095 3189 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:01:36.440215 kubelet[3189]: I0213 19:01:36.439379 3189 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:01:36.444537 kubelet[3189]: I0213 19:01:36.444473 3189 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 19:01:36.472610 kubelet[3189]: I0213 19:01:36.472541 3189 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:01:36.486854 kubelet[3189]: E0213 19:01:36.486182 3189 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:01:36.486854 kubelet[3189]: I0213 19:01:36.486246 3189 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:01:36.503661 kubelet[3189]: I0213 19:01:36.503610 3189 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:01:36.509216 kubelet[3189]: I0213 19:01:36.508223 3189 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:01:36.509216 kubelet[3189]: I0213 19:01:36.508296 3189 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-18-134","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:01:36.509216 kubelet[3189]: I0213 19:01:36.508618 3189 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:01:36.509216 kubelet[3189]: I0213 19:01:36.508638 3189 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:01:36.509601 kubelet[3189]: I0213 19:01:36.509190 3189 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:01:36.509601 kubelet[3189]: I0213 19:01:36.509523 3189 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:01:36.511774 kubelet[3189]: I0213 19:01:36.509825 3189 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:01:36.511774 kubelet[3189]: I0213 19:01:36.509863 3189 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:01:36.511774 kubelet[3189]: I0213 19:01:36.509883 3189 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:01:36.515300 kubelet[3189]: I0213 19:01:36.515249 3189 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:01:36.516132 kubelet[3189]: I0213 19:01:36.516078 3189 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:01:36.520007 kubelet[3189]: I0213 19:01:36.519951 3189 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:01:36.520805 kubelet[3189]: I0213 19:01:36.520046 3189 server.go:1287] "Started kubelet" Feb 13 19:01:36.540250 kubelet[3189]: I0213 19:01:36.539379 3189 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:01:36.555755 kubelet[3189]: I0213 19:01:36.553645 3189 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:01:36.559914 kubelet[3189]: I0213 19:01:36.558450 3189 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:01:36.559914 kubelet[3189]: I0213 19:01:36.558916 3189 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:01:36.559914 kubelet[3189]: I0213 19:01:36.559267 3189 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:01:36.573024 kubelet[3189]: I0213 19:01:36.571376 3189 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:01:36.584100 kubelet[3189]: I0213 19:01:36.576229 3189 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:01:36.616860 kubelet[3189]: I0213 19:01:36.576393 3189 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:01:36.678859 kubelet[3189]: E0213 19:01:36.577091 3189 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-18-134\" not found" Feb 13 19:01:36.678859 kubelet[3189]: I0213 19:01:36.616121 3189 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:01:36.775640 kubelet[3189]: E0213 19:01:36.775570 3189 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-18-134\" not found" Feb 13 19:01:36.778359 kubelet[3189]: I0213 19:01:36.777616 3189 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:01:36.832056 kubelet[3189]: E0213 19:01:36.829974 3189 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:01:36.832340 kubelet[3189]: I0213 19:01:36.832256 3189 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:01:36.832340 kubelet[3189]: I0213 19:01:36.832285 3189 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:01:36.861601 kubelet[3189]: I0213 19:01:36.861509 3189 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:01:36.915243 kubelet[3189]: I0213 19:01:36.913841 3189 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:01:36.915243 kubelet[3189]: I0213 19:01:36.913916 3189 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:01:36.915243 kubelet[3189]: I0213 19:01:36.913950 3189 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:01:36.915243 kubelet[3189]: I0213 19:01:36.913994 3189 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:01:36.915243 kubelet[3189]: E0213 19:01:36.914098 3189 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:01:37.016834 kubelet[3189]: E0213 19:01:37.016800 3189 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 13 19:01:37.062828 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (3207) Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.088640 3189 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.088671 3189 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.088724 3189 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.089013 3189 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.089037 3189 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.089082 3189 policy_none.go:49] "None policy: Start" Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.089102 3189 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.089121 3189 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:01:37.089787 kubelet[3189]: I0213 19:01:37.089295 3189 state_mem.go:75] "Updated machine memory state" Feb 13 19:01:37.111821 kubelet[3189]: I0213 19:01:37.107583 3189 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:01:37.117605 kubelet[3189]: I0213 19:01:37.115352 3189 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:01:37.117605 kubelet[3189]: I0213 19:01:37.115384 3189 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:01:37.117605 kubelet[3189]: I0213 19:01:37.115870 3189 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:01:37.132609 kubelet[3189]: E0213 19:01:37.132171 3189 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:01:37.217661 kubelet[3189]: I0213 19:01:37.217611 3189 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-18-134" Feb 13 19:01:37.222918 kubelet[3189]: I0213 19:01:37.221412 3189 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:37.227339 kubelet[3189]: I0213 19:01:37.226663 3189 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:37.241079 kubelet[3189]: E0213 19:01:37.241035 3189 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-18-134\" already exists" pod="kube-system/kube-scheduler-ip-172-31-18-134" Feb 13 19:01:37.242137 kubelet[3189]: E0213 19:01:37.241576 3189 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-18-134\" already exists" pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:37.259711 kubelet[3189]: I0213 19:01:37.259384 3189 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-18-134" Feb 13 19:01:37.295516 kubelet[3189]: I0213 19:01:37.293675 3189 kubelet_node_status.go:125] "Node was previously registered" node="ip-172-31-18-134" Feb 13 19:01:37.295516 kubelet[3189]: I0213 19:01:37.293811 3189 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-18-134" Feb 13 19:01:37.302125 kubelet[3189]: I0213 19:01:37.302079 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/341bfc7205109c3c0851cbc288f8f89e-ca-certs\") pod \"kube-apiserver-ip-172-31-18-134\" (UID: \"341bfc7205109c3c0851cbc288f8f89e\") " pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:37.304182 kubelet[3189]: I0213 19:01:37.303800 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-k8s-certs\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:37.304182 kubelet[3189]: I0213 19:01:37.303863 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-kubeconfig\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:37.304182 kubelet[3189]: I0213 19:01:37.303905 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/8d30d2da5ef6b852f1bf3b4dabd56ad7-kubeconfig\") pod \"kube-scheduler-ip-172-31-18-134\" (UID: \"8d30d2da5ef6b852f1bf3b4dabd56ad7\") " pod="kube-system/kube-scheduler-ip-172-31-18-134" Feb 13 19:01:37.304182 kubelet[3189]: I0213 19:01:37.303942 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/341bfc7205109c3c0851cbc288f8f89e-k8s-certs\") pod \"kube-apiserver-ip-172-31-18-134\" (UID: \"341bfc7205109c3c0851cbc288f8f89e\") " pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:37.304182 kubelet[3189]: I0213 19:01:37.303982 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/341bfc7205109c3c0851cbc288f8f89e-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-18-134\" (UID: \"341bfc7205109c3c0851cbc288f8f89e\") " pod="kube-system/kube-apiserver-ip-172-31-18-134" Feb 13 19:01:37.304552 kubelet[3189]: I0213 19:01:37.304027 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-ca-certs\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:37.304552 kubelet[3189]: I0213 19:01:37.304069 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:37.304552 kubelet[3189]: I0213 19:01:37.304109 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/726ad03d2d1806d3cf7e153f842f8768-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-18-134\" (UID: \"726ad03d2d1806d3cf7e153f842f8768\") " pod="kube-system/kube-controller-manager-ip-172-31-18-134" Feb 13 19:01:37.538048 kubelet[3189]: I0213 19:01:37.529276 3189 apiserver.go:52] "Watching apiserver" Feb 13 19:01:37.577845 kubelet[3189]: I0213 19:01:37.571129 3189 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:01:37.680507 kubelet[3189]: I0213 19:01:37.678853 3189 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-18-134" podStartSLOduration=4.678834315 podStartE2EDuration="4.678834315s" podCreationTimestamp="2025-02-13 19:01:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:01:37.678378135 +0000 UTC m=+1.443376040" watchObservedRunningTime="2025-02-13 19:01:37.678834315 +0000 UTC m=+1.443832208" Feb 13 19:01:37.736329 kubelet[3189]: I0213 19:01:37.736135 3189 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-18-134" podStartSLOduration=2.7361116 podStartE2EDuration="2.7361116s" podCreationTimestamp="2025-02-13 19:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:01:37.705104019 +0000 UTC m=+1.470101912" watchObservedRunningTime="2025-02-13 19:01:37.7361116 +0000 UTC m=+1.501109493" Feb 13 19:01:37.783094 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (3207) Feb 13 19:01:37.796379 kubelet[3189]: I0213 19:01:37.795364 3189 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-18-134" podStartSLOduration=0.795330028 podStartE2EDuration="795.330028ms" podCreationTimestamp="2025-02-13 19:01:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:01:37.738500608 +0000 UTC m=+1.503498501" watchObservedRunningTime="2025-02-13 19:01:37.795330028 +0000 UTC m=+1.560328017" Feb 13 19:01:38.691886 sudo[2238]: pam_unix(sudo:session): session closed for user root Feb 13 19:01:38.718299 sshd[2237]: Connection closed by 139.178.89.65 port 49548 Feb 13 19:01:38.718103 sshd-session[2235]: pam_unix(sshd:session): session closed for user core Feb 13 19:01:38.731980 systemd[1]: sshd@4-172.31.18.134:22-139.178.89.65:49548.service: Deactivated successfully. Feb 13 19:01:38.737821 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 19:01:38.738953 systemd[1]: session-5.scope: Consumed 9.872s CPU time, 220.3M memory peak. Feb 13 19:01:38.742228 systemd-logind[1932]: Session 5 logged out. Waiting for processes to exit. Feb 13 19:01:38.744799 systemd-logind[1932]: Removed session 5. Feb 13 19:01:38.934784 kubelet[3189]: I0213 19:01:38.934725 3189 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 19:01:38.935461 containerd[1960]: time="2025-02-13T19:01:38.935210178Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 19:01:38.936570 kubelet[3189]: I0213 19:01:38.936164 3189 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 19:01:39.655782 systemd[1]: Created slice kubepods-besteffort-pod08ce4e62_1425_414f_9ae7_e0adc33f1279.slice - libcontainer container kubepods-besteffort-pod08ce4e62_1425_414f_9ae7_e0adc33f1279.slice. Feb 13 19:01:39.684818 systemd[1]: Created slice kubepods-burstable-pode5d90c34_48e0_45ad_afd4_67b32c5a9049.slice - libcontainer container kubepods-burstable-pode5d90c34_48e0_45ad_afd4_67b32c5a9049.slice. Feb 13 19:01:39.735228 kubelet[3189]: I0213 19:01:39.735161 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/08ce4e62-1425-414f-9ae7-e0adc33f1279-lib-modules\") pod \"kube-proxy-jqgqk\" (UID: \"08ce4e62-1425-414f-9ae7-e0adc33f1279\") " pod="kube-system/kube-proxy-jqgqk" Feb 13 19:01:39.735393 kubelet[3189]: I0213 19:01:39.735233 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e5d90c34-48e0-45ad-afd4-67b32c5a9049-run\") pod \"kube-flannel-ds-p68sf\" (UID: \"e5d90c34-48e0-45ad-afd4-67b32c5a9049\") " pod="kube-flannel/kube-flannel-ds-p68sf" Feb 13 19:01:39.735393 kubelet[3189]: I0213 19:01:39.735278 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/e5d90c34-48e0-45ad-afd4-67b32c5a9049-flannel-cfg\") pod \"kube-flannel-ds-p68sf\" (UID: \"e5d90c34-48e0-45ad-afd4-67b32c5a9049\") " pod="kube-flannel/kube-flannel-ds-p68sf" Feb 13 19:01:39.735393 kubelet[3189]: I0213 19:01:39.735319 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/e5d90c34-48e0-45ad-afd4-67b32c5a9049-cni-plugin\") pod \"kube-flannel-ds-p68sf\" (UID: \"e5d90c34-48e0-45ad-afd4-67b32c5a9049\") " pod="kube-flannel/kube-flannel-ds-p68sf" Feb 13 19:01:39.735393 kubelet[3189]: I0213 19:01:39.735376 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/08ce4e62-1425-414f-9ae7-e0adc33f1279-kube-proxy\") pod \"kube-proxy-jqgqk\" (UID: \"08ce4e62-1425-414f-9ae7-e0adc33f1279\") " pod="kube-system/kube-proxy-jqgqk" Feb 13 19:01:39.735618 kubelet[3189]: I0213 19:01:39.735411 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/08ce4e62-1425-414f-9ae7-e0adc33f1279-xtables-lock\") pod \"kube-proxy-jqgqk\" (UID: \"08ce4e62-1425-414f-9ae7-e0adc33f1279\") " pod="kube-system/kube-proxy-jqgqk" Feb 13 19:01:39.735618 kubelet[3189]: I0213 19:01:39.735447 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/e5d90c34-48e0-45ad-afd4-67b32c5a9049-cni\") pod \"kube-flannel-ds-p68sf\" (UID: \"e5d90c34-48e0-45ad-afd4-67b32c5a9049\") " pod="kube-flannel/kube-flannel-ds-p68sf" Feb 13 19:01:39.735618 kubelet[3189]: I0213 19:01:39.735483 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e5d90c34-48e0-45ad-afd4-67b32c5a9049-xtables-lock\") pod \"kube-flannel-ds-p68sf\" (UID: \"e5d90c34-48e0-45ad-afd4-67b32c5a9049\") " pod="kube-flannel/kube-flannel-ds-p68sf" Feb 13 19:01:39.735618 kubelet[3189]: I0213 19:01:39.735522 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbrc7\" (UniqueName: \"kubernetes.io/projected/e5d90c34-48e0-45ad-afd4-67b32c5a9049-kube-api-access-xbrc7\") pod \"kube-flannel-ds-p68sf\" (UID: \"e5d90c34-48e0-45ad-afd4-67b32c5a9049\") " pod="kube-flannel/kube-flannel-ds-p68sf" Feb 13 19:01:39.735618 kubelet[3189]: I0213 19:01:39.735561 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8ps2\" (UniqueName: \"kubernetes.io/projected/08ce4e62-1425-414f-9ae7-e0adc33f1279-kube-api-access-k8ps2\") pod \"kube-proxy-jqgqk\" (UID: \"08ce4e62-1425-414f-9ae7-e0adc33f1279\") " pod="kube-system/kube-proxy-jqgqk" Feb 13 19:01:39.977980 containerd[1960]: time="2025-02-13T19:01:39.977912707Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jqgqk,Uid:08ce4e62-1425-414f-9ae7-e0adc33f1279,Namespace:kube-system,Attempt:0,}" Feb 13 19:01:39.993475 containerd[1960]: time="2025-02-13T19:01:39.993012043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-p68sf,Uid:e5d90c34-48e0-45ad-afd4-67b32c5a9049,Namespace:kube-flannel,Attempt:0,}" Feb 13 19:01:40.077870 containerd[1960]: time="2025-02-13T19:01:40.074754663Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:01:40.078900 containerd[1960]: time="2025-02-13T19:01:40.078055995Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:01:40.078900 containerd[1960]: time="2025-02-13T19:01:40.078095859Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:40.078900 containerd[1960]: time="2025-02-13T19:01:40.078259935Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:40.108072 containerd[1960]: time="2025-02-13T19:01:40.105655815Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:01:40.108072 containerd[1960]: time="2025-02-13T19:01:40.105796671Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:01:40.108072 containerd[1960]: time="2025-02-13T19:01:40.105835587Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:40.108072 containerd[1960]: time="2025-02-13T19:01:40.105995907Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:01:40.161524 systemd[1]: Started cri-containerd-1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47.scope - libcontainer container 1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47. Feb 13 19:01:40.184028 systemd[1]: Started cri-containerd-dec0fb606c44aac41512a53da5cf50103128fa2c05322f44cced28a6a07934de.scope - libcontainer container dec0fb606c44aac41512a53da5cf50103128fa2c05322f44cced28a6a07934de. Feb 13 19:01:40.245676 containerd[1960]: time="2025-02-13T19:01:40.244882276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jqgqk,Uid:08ce4e62-1425-414f-9ae7-e0adc33f1279,Namespace:kube-system,Attempt:0,} returns sandbox id \"dec0fb606c44aac41512a53da5cf50103128fa2c05322f44cced28a6a07934de\"" Feb 13 19:01:40.251446 containerd[1960]: time="2025-02-13T19:01:40.250891036Z" level=info msg="CreateContainer within sandbox \"dec0fb606c44aac41512a53da5cf50103128fa2c05322f44cced28a6a07934de\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 19:01:40.282194 containerd[1960]: time="2025-02-13T19:01:40.281932036Z" level=info msg="CreateContainer within sandbox \"dec0fb606c44aac41512a53da5cf50103128fa2c05322f44cced28a6a07934de\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"fd1fa62801f21b0a22e7a83a13841ca614ee16a93968da8a899fd24efc370d50\"" Feb 13 19:01:40.283705 containerd[1960]: time="2025-02-13T19:01:40.283265476Z" level=info msg="StartContainer for \"fd1fa62801f21b0a22e7a83a13841ca614ee16a93968da8a899fd24efc370d50\"" Feb 13 19:01:40.292740 containerd[1960]: time="2025-02-13T19:01:40.292097356Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-p68sf,Uid:e5d90c34-48e0-45ad-afd4-67b32c5a9049,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47\"" Feb 13 19:01:40.297670 containerd[1960]: time="2025-02-13T19:01:40.297616420Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\"" Feb 13 19:01:40.344005 systemd[1]: Started cri-containerd-fd1fa62801f21b0a22e7a83a13841ca614ee16a93968da8a899fd24efc370d50.scope - libcontainer container fd1fa62801f21b0a22e7a83a13841ca614ee16a93968da8a899fd24efc370d50. Feb 13 19:01:40.402777 containerd[1960]: time="2025-02-13T19:01:40.402648581Z" level=info msg="StartContainer for \"fd1fa62801f21b0a22e7a83a13841ca614ee16a93968da8a899fd24efc370d50\" returns successfully" Feb 13 19:01:42.581830 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2662847572.mount: Deactivated successfully. Feb 13 19:01:42.636156 containerd[1960]: time="2025-02-13T19:01:42.636091292Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:42.637794 containerd[1960]: time="2025-02-13T19:01:42.637728704Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3673532" Feb 13 19:01:42.639044 containerd[1960]: time="2025-02-13T19:01:42.638892704Z" level=info msg="ImageCreate event name:\"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:42.643241 containerd[1960]: time="2025-02-13T19:01:42.643141592Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:42.645017 containerd[1960]: time="2025-02-13T19:01:42.644803280Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3662650\" in 2.346892476s" Feb 13 19:01:42.645017 containerd[1960]: time="2025-02-13T19:01:42.644855912Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\"" Feb 13 19:01:42.650160 containerd[1960]: time="2025-02-13T19:01:42.649915124Z" level=info msg="CreateContainer within sandbox \"1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Feb 13 19:01:42.675297 containerd[1960]: time="2025-02-13T19:01:42.675244148Z" level=info msg="CreateContainer within sandbox \"1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952\"" Feb 13 19:01:42.678167 containerd[1960]: time="2025-02-13T19:01:42.676635524Z" level=info msg="StartContainer for \"9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952\"" Feb 13 19:01:42.731038 systemd[1]: Started cri-containerd-9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952.scope - libcontainer container 9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952. Feb 13 19:01:42.775897 containerd[1960]: time="2025-02-13T19:01:42.775823517Z" level=info msg="StartContainer for \"9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952\" returns successfully" Feb 13 19:01:42.779249 systemd[1]: cri-containerd-9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952.scope: Deactivated successfully. Feb 13 19:01:42.892998 containerd[1960]: time="2025-02-13T19:01:42.892798353Z" level=info msg="shim disconnected" id=9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952 namespace=k8s.io Feb 13 19:01:42.892998 containerd[1960]: time="2025-02-13T19:01:42.892871481Z" level=warning msg="cleaning up after shim disconnected" id=9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952 namespace=k8s.io Feb 13 19:01:42.892998 containerd[1960]: time="2025-02-13T19:01:42.892906929Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:01:43.043563 containerd[1960]: time="2025-02-13T19:01:43.043047318Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\"" Feb 13 19:01:43.064072 kubelet[3189]: I0213 19:01:43.063062 3189 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-jqgqk" podStartSLOduration=4.0630392220000004 podStartE2EDuration="4.063039222s" podCreationTimestamp="2025-02-13 19:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:01:41.071315224 +0000 UTC m=+4.836313141" watchObservedRunningTime="2025-02-13 19:01:43.063039222 +0000 UTC m=+6.828037115" Feb 13 19:01:43.428562 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9e0c2fc7dfa2d3a4fff385584f5f28a0fd151f729e0556c7aa2ca3c9c7872952-rootfs.mount: Deactivated successfully. Feb 13 19:01:45.221182 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1412312280.mount: Deactivated successfully. Feb 13 19:01:46.320816 containerd[1960]: time="2025-02-13T19:01:46.320753062Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:46.322888 containerd[1960]: time="2025-02-13T19:01:46.322820302Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26874261" Feb 13 19:01:46.323728 containerd[1960]: time="2025-02-13T19:01:46.323396818Z" level=info msg="ImageCreate event name:\"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:46.329131 containerd[1960]: time="2025-02-13T19:01:46.329031598Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:01:46.332710 containerd[1960]: time="2025-02-13T19:01:46.331378666Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26863435\" in 3.288268972s" Feb 13 19:01:46.332710 containerd[1960]: time="2025-02-13T19:01:46.331435486Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\"" Feb 13 19:01:46.336394 containerd[1960]: time="2025-02-13T19:01:46.336206242Z" level=info msg="CreateContainer within sandbox \"1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Feb 13 19:01:46.360366 containerd[1960]: time="2025-02-13T19:01:46.360290890Z" level=info msg="CreateContainer within sandbox \"1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa\"" Feb 13 19:01:46.362385 containerd[1960]: time="2025-02-13T19:01:46.362119402Z" level=info msg="StartContainer for \"33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa\"" Feb 13 19:01:46.421991 systemd[1]: Started cri-containerd-33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa.scope - libcontainer container 33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa. Feb 13 19:01:46.465611 systemd[1]: cri-containerd-33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa.scope: Deactivated successfully. Feb 13 19:01:46.468892 containerd[1960]: time="2025-02-13T19:01:46.468826223Z" level=info msg="StartContainer for \"33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa\" returns successfully" Feb 13 19:01:46.504648 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa-rootfs.mount: Deactivated successfully. Feb 13 19:01:46.509131 kubelet[3189]: I0213 19:01:46.509078 3189 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Feb 13 19:01:46.607233 systemd[1]: Created slice kubepods-burstable-pod45a3691c_b6a2_4296_b1c5_29de3f4e8fc3.slice - libcontainer container kubepods-burstable-pod45a3691c_b6a2_4296_b1c5_29de3f4e8fc3.slice. Feb 13 19:01:46.634778 systemd[1]: Created slice kubepods-burstable-pod214084d6_a69d_482f_83af_bd39751848d4.slice - libcontainer container kubepods-burstable-pod214084d6_a69d_482f_83af_bd39751848d4.slice. Feb 13 19:01:46.688856 containerd[1960]: time="2025-02-13T19:01:46.688474464Z" level=info msg="shim disconnected" id=33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa namespace=k8s.io Feb 13 19:01:46.688856 containerd[1960]: time="2025-02-13T19:01:46.688561020Z" level=warning msg="cleaning up after shim disconnected" id=33988a507adb70c7ef8c0f57d84f40f993c9dc6cbc7b6df6fbfbb47d28870bfa namespace=k8s.io Feb 13 19:01:46.688856 containerd[1960]: time="2025-02-13T19:01:46.688580748Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:01:46.689167 kubelet[3189]: I0213 19:01:46.688658 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45a3691c-b6a2-4296-b1c5-29de3f4e8fc3-config-volume\") pod \"coredns-668d6bf9bc-7sbgp\" (UID: \"45a3691c-b6a2-4296-b1c5-29de3f4e8fc3\") " pod="kube-system/coredns-668d6bf9bc-7sbgp" Feb 13 19:01:46.689167 kubelet[3189]: I0213 19:01:46.689102 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fdcp\" (UniqueName: \"kubernetes.io/projected/45a3691c-b6a2-4296-b1c5-29de3f4e8fc3-kube-api-access-5fdcp\") pod \"coredns-668d6bf9bc-7sbgp\" (UID: \"45a3691c-b6a2-4296-b1c5-29de3f4e8fc3\") " pod="kube-system/coredns-668d6bf9bc-7sbgp" Feb 13 19:01:46.689327 kubelet[3189]: I0213 19:01:46.689155 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx5rs\" (UniqueName: \"kubernetes.io/projected/214084d6-a69d-482f-83af-bd39751848d4-kube-api-access-tx5rs\") pod \"coredns-668d6bf9bc-qnvnt\" (UID: \"214084d6-a69d-482f-83af-bd39751848d4\") " pod="kube-system/coredns-668d6bf9bc-qnvnt" Feb 13 19:01:46.689327 kubelet[3189]: I0213 19:01:46.689229 3189 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/214084d6-a69d-482f-83af-bd39751848d4-config-volume\") pod \"coredns-668d6bf9bc-qnvnt\" (UID: \"214084d6-a69d-482f-83af-bd39751848d4\") " pod="kube-system/coredns-668d6bf9bc-qnvnt" Feb 13 19:01:46.924373 containerd[1960]: time="2025-02-13T19:01:46.923613361Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sbgp,Uid:45a3691c-b6a2-4296-b1c5-29de3f4e8fc3,Namespace:kube-system,Attempt:0,}" Feb 13 19:01:46.950512 containerd[1960]: time="2025-02-13T19:01:46.950444881Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qnvnt,Uid:214084d6-a69d-482f-83af-bd39751848d4,Namespace:kube-system,Attempt:0,}" Feb 13 19:01:46.983104 containerd[1960]: time="2025-02-13T19:01:46.982274341Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sbgp,Uid:45a3691c-b6a2-4296-b1c5-29de3f4e8fc3,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1293ad04a1f395a2ef0cd276af4592d89b89a58a723e0dbb4cca4350087bfe85\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:01:46.983831 kubelet[3189]: E0213 19:01:46.982585 3189 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1293ad04a1f395a2ef0cd276af4592d89b89a58a723e0dbb4cca4350087bfe85\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:01:46.983831 kubelet[3189]: E0213 19:01:46.982670 3189 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1293ad04a1f395a2ef0cd276af4592d89b89a58a723e0dbb4cca4350087bfe85\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-7sbgp" Feb 13 19:01:46.983831 kubelet[3189]: E0213 19:01:46.983483 3189 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1293ad04a1f395a2ef0cd276af4592d89b89a58a723e0dbb4cca4350087bfe85\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-7sbgp" Feb 13 19:01:46.984451 kubelet[3189]: E0213 19:01:46.983604 3189 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-7sbgp_kube-system(45a3691c-b6a2-4296-b1c5-29de3f4e8fc3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-7sbgp_kube-system(45a3691c-b6a2-4296-b1c5-29de3f4e8fc3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1293ad04a1f395a2ef0cd276af4592d89b89a58a723e0dbb4cca4350087bfe85\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-668d6bf9bc-7sbgp" podUID="45a3691c-b6a2-4296-b1c5-29de3f4e8fc3" Feb 13 19:01:47.007176 containerd[1960]: time="2025-02-13T19:01:47.007090018Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qnvnt,Uid:214084d6-a69d-482f-83af-bd39751848d4,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"bae06b36857065b5a1dcd0fc64190414f191f76971aa7e9aea7b264a7c9e788e\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:01:47.007561 kubelet[3189]: E0213 19:01:47.007485 3189 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bae06b36857065b5a1dcd0fc64190414f191f76971aa7e9aea7b264a7c9e788e\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:01:47.007648 kubelet[3189]: E0213 19:01:47.007569 3189 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bae06b36857065b5a1dcd0fc64190414f191f76971aa7e9aea7b264a7c9e788e\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-qnvnt" Feb 13 19:01:47.007648 kubelet[3189]: E0213 19:01:47.007604 3189 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bae06b36857065b5a1dcd0fc64190414f191f76971aa7e9aea7b264a7c9e788e\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-qnvnt" Feb 13 19:01:47.008301 kubelet[3189]: E0213 19:01:47.008128 3189 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-qnvnt_kube-system(214084d6-a69d-482f-83af-bd39751848d4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-qnvnt_kube-system(214084d6-a69d-482f-83af-bd39751848d4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bae06b36857065b5a1dcd0fc64190414f191f76971aa7e9aea7b264a7c9e788e\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-668d6bf9bc-qnvnt" podUID="214084d6-a69d-482f-83af-bd39751848d4" Feb 13 19:01:47.065948 containerd[1960]: time="2025-02-13T19:01:47.065852926Z" level=info msg="CreateContainer within sandbox \"1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Feb 13 19:01:47.089797 containerd[1960]: time="2025-02-13T19:01:47.089735458Z" level=info msg="CreateContainer within sandbox \"1798dc73034af8d431725cc6b99df15b33a0791ac2bae07d243510e96e0aaa47\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"9f03dc0e181c532c2b8302edd6e182316f8a4fb081e32c2dc04c6aadbb379af6\"" Feb 13 19:01:47.091143 containerd[1960]: time="2025-02-13T19:01:47.091080478Z" level=info msg="StartContainer for \"9f03dc0e181c532c2b8302edd6e182316f8a4fb081e32c2dc04c6aadbb379af6\"" Feb 13 19:01:47.142005 systemd[1]: Started cri-containerd-9f03dc0e181c532c2b8302edd6e182316f8a4fb081e32c2dc04c6aadbb379af6.scope - libcontainer container 9f03dc0e181c532c2b8302edd6e182316f8a4fb081e32c2dc04c6aadbb379af6. Feb 13 19:01:47.189622 containerd[1960]: time="2025-02-13T19:01:47.189359039Z" level=info msg="StartContainer for \"9f03dc0e181c532c2b8302edd6e182316f8a4fb081e32c2dc04c6aadbb379af6\" returns successfully" Feb 13 19:01:48.268024 (udev-worker)[4009]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:01:48.291931 systemd-networkd[1865]: flannel.1: Link UP Feb 13 19:01:48.291948 systemd-networkd[1865]: flannel.1: Gained carrier Feb 13 19:01:50.264925 systemd-networkd[1865]: flannel.1: Gained IPv6LL Feb 13 19:01:53.262921 ntpd[1926]: Listen normally on 8 flannel.1 192.168.0.0:123 Feb 13 19:01:53.263056 ntpd[1926]: Listen normally on 9 flannel.1 [fe80::2c74:7dff:fe37:ad28%4]:123 Feb 13 19:01:53.263570 ntpd[1926]: 13 Feb 19:01:53 ntpd[1926]: Listen normally on 8 flannel.1 192.168.0.0:123 Feb 13 19:01:53.263570 ntpd[1926]: 13 Feb 19:01:53 ntpd[1926]: Listen normally on 9 flannel.1 [fe80::2c74:7dff:fe37:ad28%4]:123 Feb 13 19:01:59.916016 containerd[1960]: time="2025-02-13T19:01:59.915941138Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sbgp,Uid:45a3691c-b6a2-4296-b1c5-29de3f4e8fc3,Namespace:kube-system,Attempt:0,}" Feb 13 19:01:59.953596 systemd-networkd[1865]: cni0: Link UP Feb 13 19:01:59.953611 systemd-networkd[1865]: cni0: Gained carrier Feb 13 19:01:59.964643 systemd-networkd[1865]: cni0: Lost carrier Feb 13 19:01:59.965097 systemd-networkd[1865]: veth9b2ae88c: Link UP Feb 13 19:01:59.969057 kernel: cni0: port 1(veth9b2ae88c) entered blocking state Feb 13 19:01:59.969168 kernel: cni0: port 1(veth9b2ae88c) entered disabled state Feb 13 19:01:59.969433 (udev-worker)[4148]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:01:59.971176 kernel: veth9b2ae88c: entered allmulticast mode Feb 13 19:01:59.972765 kernel: veth9b2ae88c: entered promiscuous mode Feb 13 19:01:59.974183 (udev-worker)[4152]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:01:59.984718 kernel: cni0: port 1(veth9b2ae88c) entered blocking state Feb 13 19:01:59.985590 kernel: cni0: port 1(veth9b2ae88c) entered forwarding state Feb 13 19:01:59.985153 systemd-networkd[1865]: veth9b2ae88c: Gained carrier Feb 13 19:01:59.986998 systemd-networkd[1865]: cni0: Gained carrier Feb 13 19:01:59.990978 containerd[1960]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x40000928e8), "name":"cbr0", "type":"bridge"} Feb 13 19:01:59.990978 containerd[1960]: delegateAdd: netconf sent to delegate plugin: Feb 13 19:02:00.041373 containerd[1960]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":8951,"name":"cbr0","type":"bridge"}time="2025-02-13T19:02:00.040902478Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:02:00.041373 containerd[1960]: time="2025-02-13T19:02:00.041011498Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:02:00.041373 containerd[1960]: time="2025-02-13T19:02:00.041040382Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:02:00.041373 containerd[1960]: time="2025-02-13T19:02:00.041191594Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:02:00.089009 systemd[1]: Started cri-containerd-f2b719771e85b5b31a36c8edc19fa7b502a07e0692eb40e46b840ef423cdeaa1.scope - libcontainer container f2b719771e85b5b31a36c8edc19fa7b502a07e0692eb40e46b840ef423cdeaa1. Feb 13 19:02:00.168377 containerd[1960]: time="2025-02-13T19:02:00.168110111Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-7sbgp,Uid:45a3691c-b6a2-4296-b1c5-29de3f4e8fc3,Namespace:kube-system,Attempt:0,} returns sandbox id \"f2b719771e85b5b31a36c8edc19fa7b502a07e0692eb40e46b840ef423cdeaa1\"" Feb 13 19:02:00.175081 containerd[1960]: time="2025-02-13T19:02:00.175009415Z" level=info msg="CreateContainer within sandbox \"f2b719771e85b5b31a36c8edc19fa7b502a07e0692eb40e46b840ef423cdeaa1\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:02:00.193280 containerd[1960]: time="2025-02-13T19:02:00.193202951Z" level=info msg="CreateContainer within sandbox \"f2b719771e85b5b31a36c8edc19fa7b502a07e0692eb40e46b840ef423cdeaa1\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"1109483031ff8b35d914139c75ee42068bb23e8bed2b49dc0e5943e3f24cbbbd\"" Feb 13 19:02:00.194329 containerd[1960]: time="2025-02-13T19:02:00.194031527Z" level=info msg="StartContainer for \"1109483031ff8b35d914139c75ee42068bb23e8bed2b49dc0e5943e3f24cbbbd\"" Feb 13 19:02:00.238985 systemd[1]: Started cri-containerd-1109483031ff8b35d914139c75ee42068bb23e8bed2b49dc0e5943e3f24cbbbd.scope - libcontainer container 1109483031ff8b35d914139c75ee42068bb23e8bed2b49dc0e5943e3f24cbbbd. Feb 13 19:02:00.284441 containerd[1960]: time="2025-02-13T19:02:00.284273904Z" level=info msg="StartContainer for \"1109483031ff8b35d914139c75ee42068bb23e8bed2b49dc0e5943e3f24cbbbd\" returns successfully" Feb 13 19:02:00.916462 containerd[1960]: time="2025-02-13T19:02:00.915936999Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qnvnt,Uid:214084d6-a69d-482f-83af-bd39751848d4,Namespace:kube-system,Attempt:0,}" Feb 13 19:02:00.957749 systemd-networkd[1865]: veth4e5e52af: Link UP Feb 13 19:02:00.960189 kernel: cni0: port 2(veth4e5e52af) entered blocking state Feb 13 19:02:00.961168 kernel: cni0: port 2(veth4e5e52af) entered disabled state Feb 13 19:02:00.961237 kernel: veth4e5e52af: entered allmulticast mode Feb 13 19:02:00.961279 kernel: veth4e5e52af: entered promiscuous mode Feb 13 19:02:00.962324 kernel: cni0: port 2(veth4e5e52af) entered blocking state Feb 13 19:02:00.964019 kernel: cni0: port 2(veth4e5e52af) entered forwarding state Feb 13 19:02:00.974932 systemd-networkd[1865]: veth4e5e52af: Gained carrier Feb 13 19:02:00.980285 containerd[1960]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x4000012938), "name":"cbr0", "type":"bridge"} Feb 13 19:02:00.980285 containerd[1960]: delegateAdd: netconf sent to delegate plugin: Feb 13 19:02:01.017492 systemd-networkd[1865]: cni0: Gained IPv6LL Feb 13 19:02:01.022249 containerd[1960]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":8951,"name":"cbr0","type":"bridge"}time="2025-02-13T19:02:01.022065515Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:02:01.022249 containerd[1960]: time="2025-02-13T19:02:01.022176587Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:02:01.022478 containerd[1960]: time="2025-02-13T19:02:01.022213823Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:02:01.022478 containerd[1960]: time="2025-02-13T19:02:01.022366403Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:02:01.060001 systemd[1]: run-containerd-runc-k8s.io-c75b269eca9dd082e962416aada9ccc3444e96eb9cf9a40f1572338f7ad0a457-runc.oOqdrv.mount: Deactivated successfully. Feb 13 19:02:01.072006 systemd[1]: Started cri-containerd-c75b269eca9dd082e962416aada9ccc3444e96eb9cf9a40f1572338f7ad0a457.scope - libcontainer container c75b269eca9dd082e962416aada9ccc3444e96eb9cf9a40f1572338f7ad0a457. Feb 13 19:02:01.128481 kubelet[3189]: I0213 19:02:01.128179 3189 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-p68sf" podStartSLOduration=16.090397782 podStartE2EDuration="22.128156256s" podCreationTimestamp="2025-02-13 19:01:39 +0000 UTC" firstStartedPulling="2025-02-13 19:01:40.29557504 +0000 UTC m=+4.060572933" lastFinishedPulling="2025-02-13 19:01:46.333333526 +0000 UTC m=+10.098331407" observedRunningTime="2025-02-13 19:01:48.087653363 +0000 UTC m=+11.852651256" watchObservedRunningTime="2025-02-13 19:02:01.128156256 +0000 UTC m=+24.893154149" Feb 13 19:02:01.145275 systemd-networkd[1865]: veth9b2ae88c: Gained IPv6LL Feb 13 19:02:01.152599 containerd[1960]: time="2025-02-13T19:02:01.152521464Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-qnvnt,Uid:214084d6-a69d-482f-83af-bd39751848d4,Namespace:kube-system,Attempt:0,} returns sandbox id \"c75b269eca9dd082e962416aada9ccc3444e96eb9cf9a40f1572338f7ad0a457\"" Feb 13 19:02:01.159403 containerd[1960]: time="2025-02-13T19:02:01.159348984Z" level=info msg="CreateContainer within sandbox \"c75b269eca9dd082e962416aada9ccc3444e96eb9cf9a40f1572338f7ad0a457\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:02:01.179147 containerd[1960]: time="2025-02-13T19:02:01.177094896Z" level=info msg="CreateContainer within sandbox \"c75b269eca9dd082e962416aada9ccc3444e96eb9cf9a40f1572338f7ad0a457\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4f488851588065e82ba438408231673611dee3afef4b757aa8acac4510519ae3\"" Feb 13 19:02:01.181598 containerd[1960]: time="2025-02-13T19:02:01.179589636Z" level=info msg="StartContainer for \"4f488851588065e82ba438408231673611dee3afef4b757aa8acac4510519ae3\"" Feb 13 19:02:01.224332 systemd[1]: Started cri-containerd-4f488851588065e82ba438408231673611dee3afef4b757aa8acac4510519ae3.scope - libcontainer container 4f488851588065e82ba438408231673611dee3afef4b757aa8acac4510519ae3. Feb 13 19:02:01.267820 containerd[1960]: time="2025-02-13T19:02:01.267561744Z" level=info msg="StartContainer for \"4f488851588065e82ba438408231673611dee3afef4b757aa8acac4510519ae3\" returns successfully" Feb 13 19:02:02.128382 kubelet[3189]: I0213 19:02:02.127382 3189 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-7sbgp" podStartSLOduration=22.127362373 podStartE2EDuration="22.127362373s" podCreationTimestamp="2025-02-13 19:01:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:02:01.128864616 +0000 UTC m=+24.893862533" watchObservedRunningTime="2025-02-13 19:02:02.127362373 +0000 UTC m=+25.892360266" Feb 13 19:02:02.128382 kubelet[3189]: I0213 19:02:02.127539 3189 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-qnvnt" podStartSLOduration=22.127530361 podStartE2EDuration="22.127530361s" podCreationTimestamp="2025-02-13 19:01:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:02:02.125591197 +0000 UTC m=+25.890589186" watchObservedRunningTime="2025-02-13 19:02:02.127530361 +0000 UTC m=+25.892528278" Feb 13 19:02:02.233458 systemd-networkd[1865]: veth4e5e52af: Gained IPv6LL Feb 13 19:02:04.262880 ntpd[1926]: Listen normally on 10 cni0 192.168.0.1:123 Feb 13 19:02:04.263027 ntpd[1926]: Listen normally on 11 cni0 [fe80::6c2c:1ff:fe75:2ed3%5]:123 Feb 13 19:02:04.263419 ntpd[1926]: 13 Feb 19:02:04 ntpd[1926]: Listen normally on 10 cni0 192.168.0.1:123 Feb 13 19:02:04.263419 ntpd[1926]: 13 Feb 19:02:04 ntpd[1926]: Listen normally on 11 cni0 [fe80::6c2c:1ff:fe75:2ed3%5]:123 Feb 13 19:02:04.263419 ntpd[1926]: 13 Feb 19:02:04 ntpd[1926]: Listen normally on 12 veth9b2ae88c [fe80::349c:cdff:fee2:81ec%6]:123 Feb 13 19:02:04.263419 ntpd[1926]: 13 Feb 19:02:04 ntpd[1926]: Listen normally on 13 veth4e5e52af [fe80::a86b:4dff:fec3:b70e%7]:123 Feb 13 19:02:04.263109 ntpd[1926]: Listen normally on 12 veth9b2ae88c [fe80::349c:cdff:fee2:81ec%6]:123 Feb 13 19:02:04.263176 ntpd[1926]: Listen normally on 13 veth4e5e52af [fe80::a86b:4dff:fec3:b70e%7]:123 Feb 13 19:02:12.921283 systemd[1]: Started sshd@5-172.31.18.134:22-139.178.89.65:42524.service - OpenSSH per-connection server daemon (139.178.89.65:42524). Feb 13 19:02:13.115617 sshd[4414]: Accepted publickey for core from 139.178.89.65 port 42524 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:13.118347 sshd-session[4414]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:13.130939 systemd-logind[1932]: New session 6 of user core. Feb 13 19:02:13.143047 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 19:02:13.403220 sshd[4416]: Connection closed by 139.178.89.65 port 42524 Feb 13 19:02:13.404133 sshd-session[4414]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:13.409441 systemd[1]: sshd@5-172.31.18.134:22-139.178.89.65:42524.service: Deactivated successfully. Feb 13 19:02:13.413903 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 19:02:13.417950 systemd-logind[1932]: Session 6 logged out. Waiting for processes to exit. Feb 13 19:02:13.420311 systemd-logind[1932]: Removed session 6. Feb 13 19:02:18.444210 systemd[1]: Started sshd@6-172.31.18.134:22-139.178.89.65:57448.service - OpenSSH per-connection server daemon (139.178.89.65:57448). Feb 13 19:02:18.641343 sshd[4450]: Accepted publickey for core from 139.178.89.65 port 57448 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:18.643990 sshd-session[4450]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:18.654905 systemd-logind[1932]: New session 7 of user core. Feb 13 19:02:18.663141 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 19:02:18.899618 sshd[4473]: Connection closed by 139.178.89.65 port 57448 Feb 13 19:02:18.900470 sshd-session[4450]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:18.907271 systemd[1]: sshd@6-172.31.18.134:22-139.178.89.65:57448.service: Deactivated successfully. Feb 13 19:02:18.911150 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 19:02:18.913467 systemd-logind[1932]: Session 7 logged out. Waiting for processes to exit. Feb 13 19:02:18.916829 systemd-logind[1932]: Removed session 7. Feb 13 19:02:23.942213 systemd[1]: Started sshd@7-172.31.18.134:22-139.178.89.65:57462.service - OpenSSH per-connection server daemon (139.178.89.65:57462). Feb 13 19:02:24.139069 sshd[4507]: Accepted publickey for core from 139.178.89.65 port 57462 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:24.141544 sshd-session[4507]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:24.150388 systemd-logind[1932]: New session 8 of user core. Feb 13 19:02:24.156971 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 19:02:24.405430 sshd[4509]: Connection closed by 139.178.89.65 port 57462 Feb 13 19:02:24.406049 sshd-session[4507]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:24.413325 systemd[1]: sshd@7-172.31.18.134:22-139.178.89.65:57462.service: Deactivated successfully. Feb 13 19:02:24.419366 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 19:02:24.422431 systemd-logind[1932]: Session 8 logged out. Waiting for processes to exit. Feb 13 19:02:24.424166 systemd-logind[1932]: Removed session 8. Feb 13 19:02:24.444317 systemd[1]: Started sshd@8-172.31.18.134:22-139.178.89.65:57468.service - OpenSSH per-connection server daemon (139.178.89.65:57468). Feb 13 19:02:24.636782 sshd[4522]: Accepted publickey for core from 139.178.89.65 port 57468 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:24.639249 sshd-session[4522]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:24.647817 systemd-logind[1932]: New session 9 of user core. Feb 13 19:02:24.655007 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 19:02:24.965355 sshd[4524]: Connection closed by 139.178.89.65 port 57468 Feb 13 19:02:24.966277 sshd-session[4522]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:24.974433 systemd-logind[1932]: Session 9 logged out. Waiting for processes to exit. Feb 13 19:02:24.978379 systemd[1]: sshd@8-172.31.18.134:22-139.178.89.65:57468.service: Deactivated successfully. Feb 13 19:02:24.989345 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 19:02:25.006289 systemd-logind[1932]: Removed session 9. Feb 13 19:02:25.014576 systemd[1]: Started sshd@9-172.31.18.134:22-139.178.89.65:53120.service - OpenSSH per-connection server daemon (139.178.89.65:53120). Feb 13 19:02:25.218118 sshd[4533]: Accepted publickey for core from 139.178.89.65 port 53120 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:25.220841 sshd-session[4533]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:25.228900 systemd-logind[1932]: New session 10 of user core. Feb 13 19:02:25.244115 systemd[1]: Started session-10.scope - Session 10 of User core. Feb 13 19:02:25.491617 sshd[4536]: Connection closed by 139.178.89.65 port 53120 Feb 13 19:02:25.492735 sshd-session[4533]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:25.499270 systemd[1]: sshd@9-172.31.18.134:22-139.178.89.65:53120.service: Deactivated successfully. Feb 13 19:02:25.506060 systemd[1]: session-10.scope: Deactivated successfully. Feb 13 19:02:25.512850 systemd-logind[1932]: Session 10 logged out. Waiting for processes to exit. Feb 13 19:02:25.516097 systemd-logind[1932]: Removed session 10. Feb 13 19:02:30.538166 systemd[1]: Started sshd@10-172.31.18.134:22-139.178.89.65:53132.service - OpenSSH per-connection server daemon (139.178.89.65:53132). Feb 13 19:02:30.734503 sshd[4571]: Accepted publickey for core from 139.178.89.65 port 53132 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:30.737021 sshd-session[4571]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:30.745467 systemd-logind[1932]: New session 11 of user core. Feb 13 19:02:30.757000 systemd[1]: Started session-11.scope - Session 11 of User core. Feb 13 19:02:31.003569 sshd[4573]: Connection closed by 139.178.89.65 port 53132 Feb 13 19:02:31.003446 sshd-session[4571]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:31.010247 systemd[1]: sshd@10-172.31.18.134:22-139.178.89.65:53132.service: Deactivated successfully. Feb 13 19:02:31.015861 systemd[1]: session-11.scope: Deactivated successfully. Feb 13 19:02:31.017420 systemd-logind[1932]: Session 11 logged out. Waiting for processes to exit. Feb 13 19:02:31.019178 systemd-logind[1932]: Removed session 11. Feb 13 19:02:36.049199 systemd[1]: Started sshd@11-172.31.18.134:22-139.178.89.65:55384.service - OpenSSH per-connection server daemon (139.178.89.65:55384). Feb 13 19:02:36.237153 sshd[4606]: Accepted publickey for core from 139.178.89.65 port 55384 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:36.240200 sshd-session[4606]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:36.249032 systemd-logind[1932]: New session 12 of user core. Feb 13 19:02:36.257951 systemd[1]: Started session-12.scope - Session 12 of User core. Feb 13 19:02:36.505762 sshd[4608]: Connection closed by 139.178.89.65 port 55384 Feb 13 19:02:36.506709 sshd-session[4606]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:36.513319 systemd[1]: sshd@11-172.31.18.134:22-139.178.89.65:55384.service: Deactivated successfully. Feb 13 19:02:36.518364 systemd[1]: session-12.scope: Deactivated successfully. Feb 13 19:02:36.522302 systemd-logind[1932]: Session 12 logged out. Waiting for processes to exit. Feb 13 19:02:36.526144 systemd-logind[1932]: Removed session 12. Feb 13 19:02:36.546213 systemd[1]: Started sshd@12-172.31.18.134:22-139.178.89.65:55398.service - OpenSSH per-connection server daemon (139.178.89.65:55398). Feb 13 19:02:36.745739 sshd[4620]: Accepted publickey for core from 139.178.89.65 port 55398 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:36.748177 sshd-session[4620]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:36.757340 systemd-logind[1932]: New session 13 of user core. Feb 13 19:02:36.764946 systemd[1]: Started session-13.scope - Session 13 of User core. Feb 13 19:02:37.120333 sshd[4622]: Connection closed by 139.178.89.65 port 55398 Feb 13 19:02:37.121370 sshd-session[4620]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:37.128068 systemd[1]: sshd@12-172.31.18.134:22-139.178.89.65:55398.service: Deactivated successfully. Feb 13 19:02:37.132234 systemd[1]: session-13.scope: Deactivated successfully. Feb 13 19:02:37.137946 systemd-logind[1932]: Session 13 logged out. Waiting for processes to exit. Feb 13 19:02:37.141968 systemd-logind[1932]: Removed session 13. Feb 13 19:02:37.170218 systemd[1]: Started sshd@13-172.31.18.134:22-139.178.89.65:55404.service - OpenSSH per-connection server daemon (139.178.89.65:55404). Feb 13 19:02:37.365380 sshd[4634]: Accepted publickey for core from 139.178.89.65 port 55404 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:37.367900 sshd-session[4634]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:37.377051 systemd-logind[1932]: New session 14 of user core. Feb 13 19:02:37.383979 systemd[1]: Started session-14.scope - Session 14 of User core. Feb 13 19:02:38.674195 sshd[4636]: Connection closed by 139.178.89.65 port 55404 Feb 13 19:02:38.675072 sshd-session[4634]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:38.687058 systemd-logind[1932]: Session 14 logged out. Waiting for processes to exit. Feb 13 19:02:38.688522 systemd[1]: sshd@13-172.31.18.134:22-139.178.89.65:55404.service: Deactivated successfully. Feb 13 19:02:38.702545 systemd[1]: session-14.scope: Deactivated successfully. Feb 13 19:02:38.726255 systemd[1]: Started sshd@14-172.31.18.134:22-139.178.89.65:55406.service - OpenSSH per-connection server daemon (139.178.89.65:55406). Feb 13 19:02:38.729309 systemd-logind[1932]: Removed session 14. Feb 13 19:02:38.918315 sshd[4659]: Accepted publickey for core from 139.178.89.65 port 55406 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:38.923531 sshd-session[4659]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:38.939048 systemd-logind[1932]: New session 15 of user core. Feb 13 19:02:38.943928 systemd[1]: Started session-15.scope - Session 15 of User core. Feb 13 19:02:39.417471 sshd[4677]: Connection closed by 139.178.89.65 port 55406 Feb 13 19:02:39.419795 sshd-session[4659]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:39.426334 systemd[1]: sshd@14-172.31.18.134:22-139.178.89.65:55406.service: Deactivated successfully. Feb 13 19:02:39.430334 systemd[1]: session-15.scope: Deactivated successfully. Feb 13 19:02:39.432885 systemd-logind[1932]: Session 15 logged out. Waiting for processes to exit. Feb 13 19:02:39.434641 systemd-logind[1932]: Removed session 15. Feb 13 19:02:39.470159 systemd[1]: Started sshd@15-172.31.18.134:22-139.178.89.65:55416.service - OpenSSH per-connection server daemon (139.178.89.65:55416). Feb 13 19:02:39.645384 sshd[4687]: Accepted publickey for core from 139.178.89.65 port 55416 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:39.648093 sshd-session[4687]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:39.658413 systemd-logind[1932]: New session 16 of user core. Feb 13 19:02:39.668967 systemd[1]: Started session-16.scope - Session 16 of User core. Feb 13 19:02:39.901744 sshd[4689]: Connection closed by 139.178.89.65 port 55416 Feb 13 19:02:39.902562 sshd-session[4687]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:39.909512 systemd[1]: sshd@15-172.31.18.134:22-139.178.89.65:55416.service: Deactivated successfully. Feb 13 19:02:39.913152 systemd[1]: session-16.scope: Deactivated successfully. Feb 13 19:02:39.914746 systemd-logind[1932]: Session 16 logged out. Waiting for processes to exit. Feb 13 19:02:39.917593 systemd-logind[1932]: Removed session 16. Feb 13 19:02:44.945466 systemd[1]: Started sshd@16-172.31.18.134:22-139.178.89.65:53050.service - OpenSSH per-connection server daemon (139.178.89.65:53050). Feb 13 19:02:45.141374 sshd[4727]: Accepted publickey for core from 139.178.89.65 port 53050 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:45.143830 sshd-session[4727]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:45.153033 systemd-logind[1932]: New session 17 of user core. Feb 13 19:02:45.159981 systemd[1]: Started session-17.scope - Session 17 of User core. Feb 13 19:02:45.408778 sshd[4729]: Connection closed by 139.178.89.65 port 53050 Feb 13 19:02:45.409583 sshd-session[4727]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:45.414531 systemd[1]: sshd@16-172.31.18.134:22-139.178.89.65:53050.service: Deactivated successfully. Feb 13 19:02:45.418457 systemd[1]: session-17.scope: Deactivated successfully. Feb 13 19:02:45.423202 systemd-logind[1932]: Session 17 logged out. Waiting for processes to exit. Feb 13 19:02:45.425280 systemd-logind[1932]: Removed session 17. Feb 13 19:02:50.450225 systemd[1]: Started sshd@17-172.31.18.134:22-139.178.89.65:53062.service - OpenSSH per-connection server daemon (139.178.89.65:53062). Feb 13 19:02:50.640767 sshd[4764]: Accepted publickey for core from 139.178.89.65 port 53062 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:50.643275 sshd-session[4764]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:50.651304 systemd-logind[1932]: New session 18 of user core. Feb 13 19:02:50.660978 systemd[1]: Started session-18.scope - Session 18 of User core. Feb 13 19:02:50.902731 sshd[4766]: Connection closed by 139.178.89.65 port 53062 Feb 13 19:02:50.901891 sshd-session[4764]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:50.909045 systemd[1]: sshd@17-172.31.18.134:22-139.178.89.65:53062.service: Deactivated successfully. Feb 13 19:02:50.913195 systemd[1]: session-18.scope: Deactivated successfully. Feb 13 19:02:50.915362 systemd-logind[1932]: Session 18 logged out. Waiting for processes to exit. Feb 13 19:02:50.919884 systemd-logind[1932]: Removed session 18. Feb 13 19:02:55.943217 systemd[1]: Started sshd@18-172.31.18.134:22-139.178.89.65:45372.service - OpenSSH per-connection server daemon (139.178.89.65:45372). Feb 13 19:02:56.130182 sshd[4798]: Accepted publickey for core from 139.178.89.65 port 45372 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:02:56.132752 sshd-session[4798]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:02:56.142793 systemd-logind[1932]: New session 19 of user core. Feb 13 19:02:56.151964 systemd[1]: Started session-19.scope - Session 19 of User core. Feb 13 19:02:56.383471 sshd[4800]: Connection closed by 139.178.89.65 port 45372 Feb 13 19:02:56.384395 sshd-session[4798]: pam_unix(sshd:session): session closed for user core Feb 13 19:02:56.390515 systemd[1]: sshd@18-172.31.18.134:22-139.178.89.65:45372.service: Deactivated successfully. Feb 13 19:02:56.394850 systemd[1]: session-19.scope: Deactivated successfully. Feb 13 19:02:56.397011 systemd-logind[1932]: Session 19 logged out. Waiting for processes to exit. Feb 13 19:02:56.398953 systemd-logind[1932]: Removed session 19. Feb 13 19:03:01.429220 systemd[1]: Started sshd@19-172.31.18.134:22-139.178.89.65:45384.service - OpenSSH per-connection server daemon (139.178.89.65:45384). Feb 13 19:03:01.615433 sshd[4833]: Accepted publickey for core from 139.178.89.65 port 45384 ssh2: RSA SHA256:N5jzFAPw/VkUdyH7hxgwbv5n548nUQy18zKQaYF7hgg Feb 13 19:03:01.618073 sshd-session[4833]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:03:01.626841 systemd-logind[1932]: New session 20 of user core. Feb 13 19:03:01.632981 systemd[1]: Started session-20.scope - Session 20 of User core. Feb 13 19:03:01.876780 sshd[4835]: Connection closed by 139.178.89.65 port 45384 Feb 13 19:03:01.877732 sshd-session[4833]: pam_unix(sshd:session): session closed for user core Feb 13 19:03:01.883967 systemd[1]: sshd@19-172.31.18.134:22-139.178.89.65:45384.service: Deactivated successfully. Feb 13 19:03:01.889331 systemd[1]: session-20.scope: Deactivated successfully. Feb 13 19:03:01.890967 systemd-logind[1932]: Session 20 logged out. Waiting for processes to exit. Feb 13 19:03:01.892623 systemd-logind[1932]: Removed session 20. Feb 13 19:03:16.629450 systemd[1]: cri-containerd-5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3.scope: Deactivated successfully. Feb 13 19:03:16.633170 systemd[1]: cri-containerd-5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3.scope: Consumed 3.797s CPU time, 51.3M memory peak. Feb 13 19:03:16.671964 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3-rootfs.mount: Deactivated successfully. Feb 13 19:03:16.679922 containerd[1960]: time="2025-02-13T19:03:16.679805403Z" level=info msg="shim disconnected" id=5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3 namespace=k8s.io Feb 13 19:03:16.680587 containerd[1960]: time="2025-02-13T19:03:16.679917759Z" level=warning msg="cleaning up after shim disconnected" id=5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3 namespace=k8s.io Feb 13 19:03:16.680587 containerd[1960]: time="2025-02-13T19:03:16.679962099Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:03:17.287849 kubelet[3189]: I0213 19:03:17.286716 3189 scope.go:117] "RemoveContainer" containerID="5fcc1175d0903678aa9f295a9bc21c5fa3af52c1ac84b8e6841a0a43814be9e3" Feb 13 19:03:17.290949 containerd[1960]: time="2025-02-13T19:03:17.290900870Z" level=info msg="CreateContainer within sandbox \"9dcdf7d091fe56e230053e1dd51d581c97b3b3efd01ec43e0dc46f784e0fffc5\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Feb 13 19:03:17.315014 containerd[1960]: time="2025-02-13T19:03:17.314951294Z" level=info msg="CreateContainer within sandbox \"9dcdf7d091fe56e230053e1dd51d581c97b3b3efd01ec43e0dc46f784e0fffc5\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"f9819d0f50409e450ec37f8732a744eb5db82ea99a1fe41f850e8d9f612bb863\"" Feb 13 19:03:17.315874 containerd[1960]: time="2025-02-13T19:03:17.315828566Z" level=info msg="StartContainer for \"f9819d0f50409e450ec37f8732a744eb5db82ea99a1fe41f850e8d9f612bb863\"" Feb 13 19:03:17.366008 systemd[1]: Started cri-containerd-f9819d0f50409e450ec37f8732a744eb5db82ea99a1fe41f850e8d9f612bb863.scope - libcontainer container f9819d0f50409e450ec37f8732a744eb5db82ea99a1fe41f850e8d9f612bb863. Feb 13 19:03:17.430063 containerd[1960]: time="2025-02-13T19:03:17.429884067Z" level=info msg="StartContainer for \"f9819d0f50409e450ec37f8732a744eb5db82ea99a1fe41f850e8d9f612bb863\" returns successfully" Feb 13 19:03:19.210879 kubelet[3189]: E0213 19:03:19.210798 3189 controller.go:195] "Failed to update lease" err="Put \"https://172.31.18.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-18-134?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 13 19:03:20.741331 systemd[1]: cri-containerd-528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c.scope: Deactivated successfully. Feb 13 19:03:20.741950 systemd[1]: cri-containerd-528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c.scope: Consumed 5.396s CPU time, 20.3M memory peak. Feb 13 19:03:20.787958 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c-rootfs.mount: Deactivated successfully. Feb 13 19:03:20.790786 containerd[1960]: time="2025-02-13T19:03:20.790659559Z" level=info msg="shim disconnected" id=528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c namespace=k8s.io Feb 13 19:03:20.790786 containerd[1960]: time="2025-02-13T19:03:20.790757407Z" level=warning msg="cleaning up after shim disconnected" id=528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c namespace=k8s.io Feb 13 19:03:20.790786 containerd[1960]: time="2025-02-13T19:03:20.790778743Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:03:21.313191 kubelet[3189]: I0213 19:03:21.313080 3189 scope.go:117] "RemoveContainer" containerID="528cb5a53e87c5bc152febe2e9cb4abef1448f2e1fc8eda7b042f2dcfd22033c" Feb 13 19:03:21.316665 containerd[1960]: time="2025-02-13T19:03:21.316577154Z" level=info msg="CreateContainer within sandbox \"9467c42bb09652a0a16192b7309ea87d656aa85b1dec0350ac09727e53b85134\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Feb 13 19:03:21.342383 containerd[1960]: time="2025-02-13T19:03:21.342304650Z" level=info msg="CreateContainer within sandbox \"9467c42bb09652a0a16192b7309ea87d656aa85b1dec0350ac09727e53b85134\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"539fac9effab603efed9d7c1fe52ac2ccfaea2380d3f50a34a8ccfba2f8a9ed7\"" Feb 13 19:03:21.343189 containerd[1960]: time="2025-02-13T19:03:21.343137666Z" level=info msg="StartContainer for \"539fac9effab603efed9d7c1fe52ac2ccfaea2380d3f50a34a8ccfba2f8a9ed7\"" Feb 13 19:03:21.391996 systemd[1]: Started cri-containerd-539fac9effab603efed9d7c1fe52ac2ccfaea2380d3f50a34a8ccfba2f8a9ed7.scope - libcontainer container 539fac9effab603efed9d7c1fe52ac2ccfaea2380d3f50a34a8ccfba2f8a9ed7. Feb 13 19:03:21.454749 containerd[1960]: time="2025-02-13T19:03:21.454661479Z" level=info msg="StartContainer for \"539fac9effab603efed9d7c1fe52ac2ccfaea2380d3f50a34a8ccfba2f8a9ed7\" returns successfully" Feb 13 19:03:29.212534 kubelet[3189]: E0213 19:03:29.212480 3189 request.go:1332] Unexpected error when reading response body: net/http: request canceled (Client.Timeout or context cancellation while reading body) Feb 13 19:03:29.213735 kubelet[3189]: E0213 19:03:29.212576 3189 controller.go:195] "Failed to update lease" err="unexpected error when reading response body. Please retry. Original error: net/http: request canceled (Client.Timeout or context cancellation while reading body)"