Jan 13 20:07:30.188514 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd083] Jan 13 20:07:30.188560 kernel: Linux version 6.6.71-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241116 p3) 14.2.1 20241116, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT Mon Jan 13 18:56:28 -00 2025 Jan 13 20:07:30.188585 kernel: KASLR disabled due to lack of seed Jan 13 20:07:30.188602 kernel: efi: EFI v2.7 by EDK II Jan 13 20:07:30.188618 kernel: efi: SMBIOS=0x7bed0000 SMBIOS 3.0=0x7beb0000 ACPI=0x786e0000 ACPI 2.0=0x786e0014 MEMATTR=0x7a736a98 MEMRESERVE=0x78557598 Jan 13 20:07:30.188634 kernel: secureboot: Secure boot disabled Jan 13 20:07:30.188652 kernel: ACPI: Early table checksum verification disabled Jan 13 20:07:30.188667 kernel: ACPI: RSDP 0x00000000786E0014 000024 (v02 AMAZON) Jan 13 20:07:30.188683 kernel: ACPI: XSDT 0x00000000786D00E8 000064 (v01 AMAZON AMZNFACP 00000001 01000013) Jan 13 20:07:30.188699 kernel: ACPI: FACP 0x00000000786B0000 000114 (v06 AMAZON AMZNFACP 00000001 AMZN 00000001) Jan 13 20:07:30.188719 kernel: ACPI: DSDT 0x0000000078640000 00159D (v02 AMAZON AMZNDSDT 00000001 INTL 20160527) Jan 13 20:07:30.188736 kernel: ACPI: APIC 0x00000000786C0000 000108 (v04 AMAZON AMZNAPIC 00000001 AMZN 00000001) Jan 13 20:07:30.188751 kernel: ACPI: SPCR 0x00000000786A0000 000050 (v02 AMAZON AMZNSPCR 00000001 AMZN 00000001) Jan 13 20:07:30.188767 kernel: ACPI: GTDT 0x0000000078690000 000060 (v02 AMAZON AMZNGTDT 00000001 AMZN 00000001) Jan 13 20:07:30.188786 kernel: ACPI: MCFG 0x0000000078680000 00003C (v02 AMAZON AMZNMCFG 00000001 AMZN 00000001) Jan 13 20:07:30.188807 kernel: ACPI: SLIT 0x0000000078670000 00002D (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Jan 13 20:07:30.188824 kernel: ACPI: IORT 0x0000000078660000 000078 (v01 AMAZON AMZNIORT 00000001 AMZN 00000001) Jan 13 20:07:30.188840 kernel: ACPI: PPTT 0x0000000078650000 0000EC (v01 AMAZON AMZNPPTT 00000001 AMZN 00000001) Jan 13 20:07:30.188856 kernel: ACPI: SPCR: console: uart,mmio,0x90a0000,115200 Jan 13 20:07:30.188873 kernel: earlycon: uart0 at MMIO 0x00000000090a0000 (options '115200') Jan 13 20:07:30.188889 kernel: printk: bootconsole [uart0] enabled Jan 13 20:07:30.188906 kernel: NUMA: Failed to initialise from firmware Jan 13 20:07:30.188922 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000004b5ffffff] Jan 13 20:07:30.188939 kernel: NUMA: NODE_DATA [mem 0x4b583f800-0x4b5844fff] Jan 13 20:07:30.188955 kernel: Zone ranges: Jan 13 20:07:30.188992 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Jan 13 20:07:30.189018 kernel: DMA32 empty Jan 13 20:07:30.189036 kernel: Normal [mem 0x0000000100000000-0x00000004b5ffffff] Jan 13 20:07:30.189053 kernel: Movable zone start for each node Jan 13 20:07:30.189069 kernel: Early memory node ranges Jan 13 20:07:30.189085 kernel: node 0: [mem 0x0000000040000000-0x000000007862ffff] Jan 13 20:07:30.189102 kernel: node 0: [mem 0x0000000078630000-0x000000007863ffff] Jan 13 20:07:30.189142 kernel: node 0: [mem 0x0000000078640000-0x00000000786effff] Jan 13 20:07:30.189161 kernel: node 0: [mem 0x00000000786f0000-0x000000007872ffff] Jan 13 20:07:30.189177 kernel: node 0: [mem 0x0000000078730000-0x000000007bbfffff] Jan 13 20:07:30.189194 kernel: node 0: [mem 0x000000007bc00000-0x000000007bfdffff] Jan 13 20:07:30.189212 kernel: node 0: [mem 0x000000007bfe0000-0x000000007fffffff] Jan 13 20:07:30.189304 kernel: node 0: [mem 0x0000000400000000-0x00000004b5ffffff] Jan 13 20:07:30.189501 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000004b5ffffff] Jan 13 20:07:30.189523 kernel: On node 0, zone Normal: 8192 pages in unavailable ranges Jan 13 20:07:30.189549 kernel: psci: probing for conduit method from ACPI. Jan 13 20:07:30.189567 kernel: psci: PSCIv1.0 detected in firmware. Jan 13 20:07:30.189584 kernel: psci: Using standard PSCI v0.2 function IDs Jan 13 20:07:30.189606 kernel: psci: Trusted OS migration not required Jan 13 20:07:30.189623 kernel: psci: SMC Calling Convention v1.1 Jan 13 20:07:30.189640 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Jan 13 20:07:30.189657 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Jan 13 20:07:30.189675 kernel: pcpu-alloc: [0] 0 [0] 1 Jan 13 20:07:30.189692 kernel: Detected PIPT I-cache on CPU0 Jan 13 20:07:30.189709 kernel: CPU features: detected: GIC system register CPU interface Jan 13 20:07:30.189727 kernel: CPU features: detected: Spectre-v2 Jan 13 20:07:30.189743 kernel: CPU features: detected: Spectre-v3a Jan 13 20:07:30.189761 kernel: CPU features: detected: Spectre-BHB Jan 13 20:07:30.189778 kernel: CPU features: detected: ARM erratum 1742098 Jan 13 20:07:30.189795 kernel: CPU features: detected: ARM errata 1165522, 1319367, or 1530923 Jan 13 20:07:30.189817 kernel: alternatives: applying boot alternatives Jan 13 20:07:30.189836 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=9798117b3b15ef802e3d618077f87253cc08e0d5280b8fe28b307e7558b7ebcc Jan 13 20:07:30.189856 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 13 20:07:30.189873 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 13 20:07:30.189891 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 13 20:07:30.189908 kernel: Fallback order for Node 0: 0 Jan 13 20:07:30.189925 kernel: Built 1 zonelists, mobility grouping on. Total pages: 991872 Jan 13 20:07:30.189942 kernel: Policy zone: Normal Jan 13 20:07:30.189959 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 13 20:07:30.189976 kernel: software IO TLB: area num 2. Jan 13 20:07:30.189998 kernel: software IO TLB: mapped [mem 0x000000007c000000-0x0000000080000000] (64MB) Jan 13 20:07:30.190016 kernel: Memory: 3819640K/4030464K available (10304K kernel code, 2184K rwdata, 8092K rodata, 39936K init, 897K bss, 210824K reserved, 0K cma-reserved) Jan 13 20:07:30.190034 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jan 13 20:07:30.190051 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 13 20:07:30.190069 kernel: rcu: RCU event tracing is enabled. Jan 13 20:07:30.190087 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jan 13 20:07:30.190105 kernel: Trampoline variant of Tasks RCU enabled. Jan 13 20:07:30.193161 kernel: Tracing variant of Tasks RCU enabled. Jan 13 20:07:30.193181 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 13 20:07:30.193199 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jan 13 20:07:30.193217 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Jan 13 20:07:30.193243 kernel: GICv3: 96 SPIs implemented Jan 13 20:07:30.193262 kernel: GICv3: 0 Extended SPIs implemented Jan 13 20:07:30.193279 kernel: Root IRQ handler: gic_handle_irq Jan 13 20:07:30.193296 kernel: GICv3: GICv3 features: 16 PPIs Jan 13 20:07:30.193314 kernel: GICv3: CPU0: found redistributor 0 region 0:0x0000000010200000 Jan 13 20:07:30.193331 kernel: ITS [mem 0x10080000-0x1009ffff] Jan 13 20:07:30.193348 kernel: ITS@0x0000000010080000: allocated 8192 Devices @4000b0000 (indirect, esz 8, psz 64K, shr 1) Jan 13 20:07:30.193366 kernel: ITS@0x0000000010080000: allocated 8192 Interrupt Collections @4000c0000 (flat, esz 8, psz 64K, shr 1) Jan 13 20:07:30.193383 kernel: GICv3: using LPI property table @0x00000004000d0000 Jan 13 20:07:30.193401 kernel: ITS: Using hypervisor restricted LPI range [128] Jan 13 20:07:30.193418 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000004000e0000 Jan 13 20:07:30.193435 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 13 20:07:30.193457 kernel: arch_timer: cp15 timer(s) running at 83.33MHz (virt). Jan 13 20:07:30.193475 kernel: clocksource: arch_sys_counter: mask: 0x1ffffffffffffff max_cycles: 0x13381ebeec, max_idle_ns: 440795203145 ns Jan 13 20:07:30.193492 kernel: sched_clock: 57 bits at 83MHz, resolution 12ns, wraps every 4398046511100ns Jan 13 20:07:30.193510 kernel: Console: colour dummy device 80x25 Jan 13 20:07:30.193528 kernel: printk: console [tty1] enabled Jan 13 20:07:30.193546 kernel: ACPI: Core revision 20230628 Jan 13 20:07:30.193564 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 166.66 BogoMIPS (lpj=83333) Jan 13 20:07:30.193582 kernel: pid_max: default: 32768 minimum: 301 Jan 13 20:07:30.193600 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 13 20:07:30.193617 kernel: landlock: Up and running. Jan 13 20:07:30.193639 kernel: SELinux: Initializing. Jan 13 20:07:30.193657 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 13 20:07:30.193675 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 13 20:07:30.193693 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 13 20:07:30.193711 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 13 20:07:30.193729 kernel: rcu: Hierarchical SRCU implementation. Jan 13 20:07:30.193747 kernel: rcu: Max phase no-delay instances is 400. Jan 13 20:07:30.193764 kernel: Platform MSI: ITS@0x10080000 domain created Jan 13 20:07:30.193786 kernel: PCI/MSI: ITS@0x10080000 domain created Jan 13 20:07:30.193804 kernel: Remapping and enabling EFI services. Jan 13 20:07:30.193821 kernel: smp: Bringing up secondary CPUs ... Jan 13 20:07:30.193839 kernel: Detected PIPT I-cache on CPU1 Jan 13 20:07:30.193857 kernel: GICv3: CPU1: found redistributor 1 region 0:0x0000000010220000 Jan 13 20:07:30.193874 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000004000f0000 Jan 13 20:07:30.193892 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd083] Jan 13 20:07:30.193909 kernel: smp: Brought up 1 node, 2 CPUs Jan 13 20:07:30.193927 kernel: SMP: Total of 2 processors activated. Jan 13 20:07:30.193944 kernel: CPU features: detected: 32-bit EL0 Support Jan 13 20:07:30.193967 kernel: CPU features: detected: 32-bit EL1 Support Jan 13 20:07:30.193985 kernel: CPU features: detected: CRC32 instructions Jan 13 20:07:30.194013 kernel: CPU: All CPU(s) started at EL1 Jan 13 20:07:30.194036 kernel: alternatives: applying system-wide alternatives Jan 13 20:07:30.194054 kernel: devtmpfs: initialized Jan 13 20:07:30.194073 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 13 20:07:30.194091 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jan 13 20:07:30.194122 kernel: pinctrl core: initialized pinctrl subsystem Jan 13 20:07:30.194148 kernel: SMBIOS 3.0.0 present. Jan 13 20:07:30.194173 kernel: DMI: Amazon EC2 a1.large/, BIOS 1.0 11/1/2018 Jan 13 20:07:30.194192 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 13 20:07:30.194211 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Jan 13 20:07:30.194230 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Jan 13 20:07:30.194248 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Jan 13 20:07:30.194266 kernel: audit: initializing netlink subsys (disabled) Jan 13 20:07:30.194285 kernel: audit: type=2000 audit(0.221:1): state=initialized audit_enabled=0 res=1 Jan 13 20:07:30.194308 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 13 20:07:30.194327 kernel: cpuidle: using governor menu Jan 13 20:07:30.194345 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Jan 13 20:07:30.194363 kernel: ASID allocator initialised with 65536 entries Jan 13 20:07:30.194381 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 13 20:07:30.194400 kernel: Serial: AMBA PL011 UART driver Jan 13 20:07:30.194418 kernel: Modules: 17360 pages in range for non-PLT usage Jan 13 20:07:30.194436 kernel: Modules: 508880 pages in range for PLT usage Jan 13 20:07:30.194455 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 13 20:07:30.194478 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Jan 13 20:07:30.194496 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Jan 13 20:07:30.194515 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Jan 13 20:07:30.194533 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 13 20:07:30.194551 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Jan 13 20:07:30.194569 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Jan 13 20:07:30.194588 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Jan 13 20:07:30.194606 kernel: ACPI: Added _OSI(Module Device) Jan 13 20:07:30.194624 kernel: ACPI: Added _OSI(Processor Device) Jan 13 20:07:30.194646 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 13 20:07:30.194665 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 13 20:07:30.194684 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 13 20:07:30.194702 kernel: ACPI: Interpreter enabled Jan 13 20:07:30.194720 kernel: ACPI: Using GIC for interrupt routing Jan 13 20:07:30.194738 kernel: ACPI: MCFG table detected, 1 entries Jan 13 20:07:30.194756 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-0f]) Jan 13 20:07:30.195062 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 13 20:07:30.205299 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Jan 13 20:07:30.205518 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Jan 13 20:07:30.205729 kernel: acpi PNP0A08:00: ECAM area [mem 0x20000000-0x20ffffff] reserved by PNP0C02:00 Jan 13 20:07:30.205934 kernel: acpi PNP0A08:00: ECAM at [mem 0x20000000-0x20ffffff] for [bus 00-0f] Jan 13 20:07:30.205961 kernel: ACPI: Remapped I/O 0x000000001fff0000 to [io 0x0000-0xffff window] Jan 13 20:07:30.205981 kernel: acpiphp: Slot [1] registered Jan 13 20:07:30.206000 kernel: acpiphp: Slot [2] registered Jan 13 20:07:30.206019 kernel: acpiphp: Slot [3] registered Jan 13 20:07:30.206048 kernel: acpiphp: Slot [4] registered Jan 13 20:07:30.206067 kernel: acpiphp: Slot [5] registered Jan 13 20:07:30.206086 kernel: acpiphp: Slot [6] registered Jan 13 20:07:30.206104 kernel: acpiphp: Slot [7] registered Jan 13 20:07:30.206155 kernel: acpiphp: Slot [8] registered Jan 13 20:07:30.206175 kernel: acpiphp: Slot [9] registered Jan 13 20:07:30.206194 kernel: acpiphp: Slot [10] registered Jan 13 20:07:30.206212 kernel: acpiphp: Slot [11] registered Jan 13 20:07:30.206231 kernel: acpiphp: Slot [12] registered Jan 13 20:07:30.206249 kernel: acpiphp: Slot [13] registered Jan 13 20:07:30.206275 kernel: acpiphp: Slot [14] registered Jan 13 20:07:30.206293 kernel: acpiphp: Slot [15] registered Jan 13 20:07:30.206311 kernel: acpiphp: Slot [16] registered Jan 13 20:07:30.206329 kernel: acpiphp: Slot [17] registered Jan 13 20:07:30.206347 kernel: acpiphp: Slot [18] registered Jan 13 20:07:30.206365 kernel: acpiphp: Slot [19] registered Jan 13 20:07:30.206383 kernel: acpiphp: Slot [20] registered Jan 13 20:07:30.206402 kernel: acpiphp: Slot [21] registered Jan 13 20:07:30.206420 kernel: acpiphp: Slot [22] registered Jan 13 20:07:30.206442 kernel: acpiphp: Slot [23] registered Jan 13 20:07:30.206461 kernel: acpiphp: Slot [24] registered Jan 13 20:07:30.206479 kernel: acpiphp: Slot [25] registered Jan 13 20:07:30.206497 kernel: acpiphp: Slot [26] registered Jan 13 20:07:30.206515 kernel: acpiphp: Slot [27] registered Jan 13 20:07:30.206534 kernel: acpiphp: Slot [28] registered Jan 13 20:07:30.206553 kernel: acpiphp: Slot [29] registered Jan 13 20:07:30.206572 kernel: acpiphp: Slot [30] registered Jan 13 20:07:30.206591 kernel: acpiphp: Slot [31] registered Jan 13 20:07:30.206609 kernel: PCI host bridge to bus 0000:00 Jan 13 20:07:30.206846 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xffffffff window] Jan 13 20:07:30.207033 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Jan 13 20:07:30.207251 kernel: pci_bus 0000:00: root bus resource [mem 0x400000000000-0x407fffffffff window] Jan 13 20:07:30.207438 kernel: pci_bus 0000:00: root bus resource [bus 00-0f] Jan 13 20:07:30.207679 kernel: pci 0000:00:00.0: [1d0f:0200] type 00 class 0x060000 Jan 13 20:07:30.207915 kernel: pci 0000:00:01.0: [1d0f:8250] type 00 class 0x070003 Jan 13 20:07:30.208191 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80118000-0x80118fff] Jan 13 20:07:30.208457 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Jan 13 20:07:30.208674 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80114000-0x80117fff] Jan 13 20:07:30.208884 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D2 D3hot D3cold Jan 13 20:07:30.209174 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Jan 13 20:07:30.209397 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80110000-0x80113fff] Jan 13 20:07:30.209609 kernel: pci 0000:00:05.0: reg 0x18: [mem 0x80000000-0x800fffff pref] Jan 13 20:07:30.209821 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x80100000-0x8010ffff] Jan 13 20:07:30.210026 kernel: pci 0000:00:05.0: PME# supported from D0 D1 D2 D3hot D3cold Jan 13 20:07:30.210281 kernel: pci 0000:00:05.0: BAR 2: assigned [mem 0x80000000-0x800fffff pref] Jan 13 20:07:30.210491 kernel: pci 0000:00:05.0: BAR 4: assigned [mem 0x80100000-0x8010ffff] Jan 13 20:07:30.210700 kernel: pci 0000:00:04.0: BAR 0: assigned [mem 0x80110000-0x80113fff] Jan 13 20:07:30.210909 kernel: pci 0000:00:05.0: BAR 0: assigned [mem 0x80114000-0x80117fff] Jan 13 20:07:30.211135 kernel: pci 0000:00:01.0: BAR 0: assigned [mem 0x80118000-0x80118fff] Jan 13 20:07:30.211342 kernel: pci_bus 0000:00: resource 4 [mem 0x80000000-0xffffffff window] Jan 13 20:07:30.211528 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Jan 13 20:07:30.211710 kernel: pci_bus 0000:00: resource 6 [mem 0x400000000000-0x407fffffffff window] Jan 13 20:07:30.211736 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Jan 13 20:07:30.211755 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Jan 13 20:07:30.211774 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Jan 13 20:07:30.211793 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Jan 13 20:07:30.211811 kernel: iommu: Default domain type: Translated Jan 13 20:07:30.211836 kernel: iommu: DMA domain TLB invalidation policy: strict mode Jan 13 20:07:30.211855 kernel: efivars: Registered efivars operations Jan 13 20:07:30.211873 kernel: vgaarb: loaded Jan 13 20:07:30.211892 kernel: clocksource: Switched to clocksource arch_sys_counter Jan 13 20:07:30.211910 kernel: VFS: Disk quotas dquot_6.6.0 Jan 13 20:07:30.211928 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 13 20:07:30.211946 kernel: pnp: PnP ACPI init Jan 13 20:07:30.212205 kernel: system 00:00: [mem 0x20000000-0x2fffffff] could not be reserved Jan 13 20:07:30.212238 kernel: pnp: PnP ACPI: found 1 devices Jan 13 20:07:30.212258 kernel: NET: Registered PF_INET protocol family Jan 13 20:07:30.212276 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 13 20:07:30.212295 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 13 20:07:30.212314 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 13 20:07:30.212333 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 13 20:07:30.212352 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 13 20:07:30.212370 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 13 20:07:30.212388 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 13 20:07:30.212412 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 13 20:07:30.212431 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 13 20:07:30.212449 kernel: PCI: CLS 0 bytes, default 64 Jan 13 20:07:30.212467 kernel: kvm [1]: HYP mode not available Jan 13 20:07:30.212486 kernel: Initialise system trusted keyrings Jan 13 20:07:30.212505 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 13 20:07:30.212524 kernel: Key type asymmetric registered Jan 13 20:07:30.212543 kernel: Asymmetric key parser 'x509' registered Jan 13 20:07:30.212561 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 13 20:07:30.212584 kernel: io scheduler mq-deadline registered Jan 13 20:07:30.212603 kernel: io scheduler kyber registered Jan 13 20:07:30.212621 kernel: io scheduler bfq registered Jan 13 20:07:30.212844 kernel: pl061_gpio ARMH0061:00: PL061 GPIO chip registered Jan 13 20:07:30.212873 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Jan 13 20:07:30.212893 kernel: ACPI: button: Power Button [PWRB] Jan 13 20:07:30.212912 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input1 Jan 13 20:07:30.212931 kernel: ACPI: button: Sleep Button [SLPB] Jan 13 20:07:30.212955 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 13 20:07:30.212995 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Jan 13 20:07:30.213326 kernel: serial 0000:00:01.0: enabling device (0010 -> 0012) Jan 13 20:07:30.213356 kernel: printk: console [ttyS0] disabled Jan 13 20:07:30.213376 kernel: 0000:00:01.0: ttyS0 at MMIO 0x80118000 (irq = 14, base_baud = 115200) is a 16550A Jan 13 20:07:30.213396 kernel: printk: console [ttyS0] enabled Jan 13 20:07:30.213415 kernel: printk: bootconsole [uart0] disabled Jan 13 20:07:30.213433 kernel: thunder_xcv, ver 1.0 Jan 13 20:07:30.213452 kernel: thunder_bgx, ver 1.0 Jan 13 20:07:30.213471 kernel: nicpf, ver 1.0 Jan 13 20:07:30.213499 kernel: nicvf, ver 1.0 Jan 13 20:07:30.213730 kernel: rtc-efi rtc-efi.0: registered as rtc0 Jan 13 20:07:30.213924 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-01-13T20:07:29 UTC (1736798849) Jan 13 20:07:30.213950 kernel: hid: raw HID events driver (C) Jiri Kosina Jan 13 20:07:30.213969 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 3 counters available Jan 13 20:07:30.213987 kernel: watchdog: Delayed init of the lockup detector failed: -19 Jan 13 20:07:30.214006 kernel: watchdog: Hard watchdog permanently disabled Jan 13 20:07:30.214034 kernel: NET: Registered PF_INET6 protocol family Jan 13 20:07:30.214091 kernel: Segment Routing with IPv6 Jan 13 20:07:30.214140 kernel: In-situ OAM (IOAM) with IPv6 Jan 13 20:07:30.214161 kernel: NET: Registered PF_PACKET protocol family Jan 13 20:07:30.214180 kernel: Key type dns_resolver registered Jan 13 20:07:30.214199 kernel: registered taskstats version 1 Jan 13 20:07:30.214218 kernel: Loading compiled-in X.509 certificates Jan 13 20:07:30.214237 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.71-flatcar: 46cb4d1b22f3a5974766fe7d7b651e2f296d4fe0' Jan 13 20:07:30.214256 kernel: Key type .fscrypt registered Jan 13 20:07:30.214274 kernel: Key type fscrypt-provisioning registered Jan 13 20:07:30.214299 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 13 20:07:30.214318 kernel: ima: Allocated hash algorithm: sha1 Jan 13 20:07:30.214336 kernel: ima: No architecture policies found Jan 13 20:07:30.214355 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Jan 13 20:07:30.214373 kernel: clk: Disabling unused clocks Jan 13 20:07:30.214391 kernel: Freeing unused kernel memory: 39936K Jan 13 20:07:30.214410 kernel: Run /init as init process Jan 13 20:07:30.214428 kernel: with arguments: Jan 13 20:07:30.214446 kernel: /init Jan 13 20:07:30.214468 kernel: with environment: Jan 13 20:07:30.214487 kernel: HOME=/ Jan 13 20:07:30.214505 kernel: TERM=linux Jan 13 20:07:30.214523 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 13 20:07:30.214546 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 13 20:07:30.214569 systemd[1]: Detected virtualization amazon. Jan 13 20:07:30.214590 systemd[1]: Detected architecture arm64. Jan 13 20:07:30.214615 systemd[1]: Running in initrd. Jan 13 20:07:30.214635 systemd[1]: No hostname configured, using default hostname. Jan 13 20:07:30.214655 systemd[1]: Hostname set to . Jan 13 20:07:30.214676 systemd[1]: Initializing machine ID from VM UUID. Jan 13 20:07:30.214696 systemd[1]: Queued start job for default target initrd.target. Jan 13 20:07:30.214716 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 13 20:07:30.214736 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 13 20:07:30.214757 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 13 20:07:30.214783 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 13 20:07:30.214804 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 13 20:07:30.214825 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 13 20:07:30.214849 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 13 20:07:30.214869 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 13 20:07:30.214890 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 13 20:07:30.214910 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 13 20:07:30.214934 systemd[1]: Reached target paths.target - Path Units. Jan 13 20:07:30.214955 systemd[1]: Reached target slices.target - Slice Units. Jan 13 20:07:30.214975 systemd[1]: Reached target swap.target - Swaps. Jan 13 20:07:30.214995 systemd[1]: Reached target timers.target - Timer Units. Jan 13 20:07:30.215016 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 13 20:07:30.215036 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 13 20:07:30.215056 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 13 20:07:30.215076 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 13 20:07:30.215096 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 13 20:07:30.217177 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 13 20:07:30.217210 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 13 20:07:30.217232 systemd[1]: Reached target sockets.target - Socket Units. Jan 13 20:07:30.217252 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 13 20:07:30.217273 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 13 20:07:30.217294 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 13 20:07:30.217314 systemd[1]: Starting systemd-fsck-usr.service... Jan 13 20:07:30.217334 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 13 20:07:30.217363 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 13 20:07:30.217384 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:07:30.217404 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 13 20:07:30.217425 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 13 20:07:30.217445 systemd[1]: Finished systemd-fsck-usr.service. Jan 13 20:07:30.217512 systemd-journald[251]: Collecting audit messages is disabled. Jan 13 20:07:30.217562 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 13 20:07:30.217584 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:07:30.217604 systemd-journald[251]: Journal started Jan 13 20:07:30.217653 systemd-journald[251]: Runtime Journal (/run/log/journal/ec203ce844de19d765bb1c1bc41889cd) is 8.0M, max 75.3M, 67.3M free. Jan 13 20:07:30.216806 systemd-modules-load[252]: Inserted module 'overlay' Jan 13 20:07:30.227187 systemd[1]: Started systemd-journald.service - Journal Service. Jan 13 20:07:30.232159 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 13 20:07:30.258303 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 13 20:07:30.260321 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 13 20:07:30.264781 kernel: Bridge firewalling registered Jan 13 20:07:30.263321 systemd-modules-load[252]: Inserted module 'br_netfilter' Jan 13 20:07:30.272388 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 13 20:07:30.295427 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 13 20:07:30.304448 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 13 20:07:30.310948 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:07:30.329829 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 13 20:07:30.336407 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 13 20:07:30.352533 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 13 20:07:30.358374 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 13 20:07:30.382343 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:07:30.396179 dracut-cmdline[282]: dracut-dracut-053 Jan 13 20:07:30.399848 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 13 20:07:30.403657 dracut-cmdline[282]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=9798117b3b15ef802e3d618077f87253cc08e0d5280b8fe28b307e7558b7ebcc Jan 13 20:07:30.477743 systemd-resolved[296]: Positive Trust Anchors: Jan 13 20:07:30.477781 systemd-resolved[296]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 13 20:07:30.477842 systemd-resolved[296]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 13 20:07:30.566152 kernel: SCSI subsystem initialized Jan 13 20:07:30.575136 kernel: Loading iSCSI transport class v2.0-870. Jan 13 20:07:30.586152 kernel: iscsi: registered transport (tcp) Jan 13 20:07:30.608157 kernel: iscsi: registered transport (qla4xxx) Jan 13 20:07:30.608234 kernel: QLogic iSCSI HBA Driver Jan 13 20:07:30.682141 kernel: random: crng init done Jan 13 20:07:30.682323 systemd-resolved[296]: Defaulting to hostname 'linux'. Jan 13 20:07:30.687570 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 13 20:07:30.690667 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 13 20:07:30.717205 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 13 20:07:30.726432 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 13 20:07:30.758393 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 13 20:07:30.758508 kernel: device-mapper: uevent: version 1.0.3 Jan 13 20:07:30.758537 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 13 20:07:30.824162 kernel: raid6: neonx8 gen() 6436 MB/s Jan 13 20:07:30.841145 kernel: raid6: neonx4 gen() 6467 MB/s Jan 13 20:07:30.858148 kernel: raid6: neonx2 gen() 5351 MB/s Jan 13 20:07:30.875144 kernel: raid6: neonx1 gen() 3931 MB/s Jan 13 20:07:30.892143 kernel: raid6: int64x8 gen() 3603 MB/s Jan 13 20:07:30.909158 kernel: raid6: int64x4 gen() 3663 MB/s Jan 13 20:07:30.926153 kernel: raid6: int64x2 gen() 3536 MB/s Jan 13 20:07:30.943905 kernel: raid6: int64x1 gen() 2758 MB/s Jan 13 20:07:30.943939 kernel: raid6: using algorithm neonx4 gen() 6467 MB/s Jan 13 20:07:30.961902 kernel: raid6: .... xor() 4984 MB/s, rmw enabled Jan 13 20:07:30.961939 kernel: raid6: using neon recovery algorithm Jan 13 20:07:30.969148 kernel: xor: measuring software checksum speed Jan 13 20:07:30.970144 kernel: 8regs : 11505 MB/sec Jan 13 20:07:30.971147 kernel: 32regs : 12017 MB/sec Jan 13 20:07:30.973146 kernel: arm64_neon : 8963 MB/sec Jan 13 20:07:30.973179 kernel: xor: using function: 32regs (12017 MB/sec) Jan 13 20:07:31.056157 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 13 20:07:31.076171 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 13 20:07:31.087456 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 13 20:07:31.130805 systemd-udevd[472]: Using default interface naming scheme 'v255'. Jan 13 20:07:31.138563 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 13 20:07:31.153443 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 13 20:07:31.184008 dracut-pre-trigger[474]: rd.md=0: removing MD RAID activation Jan 13 20:07:31.239021 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 13 20:07:31.250585 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 13 20:07:31.370215 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 13 20:07:31.386494 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 13 20:07:31.440926 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 13 20:07:31.451953 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 13 20:07:31.457092 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 13 20:07:31.464098 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 13 20:07:31.488549 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 13 20:07:31.515910 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 13 20:07:31.557395 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Jan 13 20:07:31.557458 kernel: ena 0000:00:05.0: enabling device (0010 -> 0012) Jan 13 20:07:31.578103 kernel: ena 0000:00:05.0: ENA device version: 0.10 Jan 13 20:07:31.578408 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Jan 13 20:07:31.578649 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80114000, mac addr 06:db:2b:e0:d9:ff Jan 13 20:07:31.582865 (udev-worker)[517]: Network interface NamePolicy= disabled on kernel command line. Jan 13 20:07:31.589753 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 13 20:07:31.589997 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:07:31.599105 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 13 20:07:31.603749 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 13 20:07:31.604064 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:07:31.608866 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:07:31.644802 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Jan 13 20:07:31.644845 kernel: nvme nvme0: pci function 0000:00:04.0 Jan 13 20:07:31.634545 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:07:31.652168 kernel: nvme nvme0: 2/0/0 default/read/poll queues Jan 13 20:07:31.659280 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 13 20:07:31.659362 kernel: GPT:9289727 != 16777215 Jan 13 20:07:31.660506 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 13 20:07:31.661234 kernel: GPT:9289727 != 16777215 Jan 13 20:07:31.662301 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 13 20:07:31.663198 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jan 13 20:07:31.671500 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:07:31.687470 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 13 20:07:31.714374 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:07:31.835281 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Jan 13 20:07:31.855164 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by (udev-worker) (528) Jan 13 20:07:31.889879 kernel: BTRFS: device fsid 2be7cc1c-29d4-4496-b29b-8561323213d2 devid 1 transid 38 /dev/nvme0n1p3 scanned by (udev-worker) (522) Jan 13 20:07:31.939725 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Jan 13 20:07:31.973984 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Jan 13 20:07:32.000879 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Jan 13 20:07:32.001038 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Jan 13 20:07:32.028501 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 13 20:07:32.044940 disk-uuid[663]: Primary Header is updated. Jan 13 20:07:32.044940 disk-uuid[663]: Secondary Entries is updated. Jan 13 20:07:32.044940 disk-uuid[663]: Secondary Header is updated. Jan 13 20:07:32.057159 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jan 13 20:07:33.071151 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Jan 13 20:07:33.074108 disk-uuid[664]: The operation has completed successfully. Jan 13 20:07:33.262790 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 13 20:07:33.264169 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 13 20:07:33.319411 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 13 20:07:33.328392 sh[925]: Success Jan 13 20:07:33.356171 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Jan 13 20:07:33.464696 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 13 20:07:33.472504 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 13 20:07:33.486411 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 13 20:07:33.508528 kernel: BTRFS info (device dm-0): first mount of filesystem 2be7cc1c-29d4-4496-b29b-8561323213d2 Jan 13 20:07:33.508591 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Jan 13 20:07:33.510350 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 13 20:07:33.511578 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 13 20:07:33.512634 kernel: BTRFS info (device dm-0): using free space tree Jan 13 20:07:33.638144 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jan 13 20:07:33.672238 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 13 20:07:33.678510 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 13 20:07:33.693357 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 13 20:07:33.702566 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 13 20:07:33.722031 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 9f8ecb6c-ace6-4d16-8781-f4e964dc0779 Jan 13 20:07:33.722100 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Jan 13 20:07:33.723262 kernel: BTRFS info (device nvme0n1p6): using free space tree Jan 13 20:07:33.733523 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Jan 13 20:07:33.748830 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 13 20:07:33.752908 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 9f8ecb6c-ace6-4d16-8781-f4e964dc0779 Jan 13 20:07:33.782211 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 13 20:07:33.798463 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 13 20:07:33.879691 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 13 20:07:33.906364 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 13 20:07:33.953260 systemd-networkd[1117]: lo: Link UP Jan 13 20:07:33.953281 systemd-networkd[1117]: lo: Gained carrier Jan 13 20:07:33.955864 systemd-networkd[1117]: Enumeration completed Jan 13 20:07:33.956006 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 13 20:07:33.957224 systemd-networkd[1117]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:07:33.957231 systemd-networkd[1117]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 13 20:07:33.961849 systemd[1]: Reached target network.target - Network. Jan 13 20:07:33.982065 systemd-networkd[1117]: eth0: Link UP Jan 13 20:07:33.982080 systemd-networkd[1117]: eth0: Gained carrier Jan 13 20:07:33.982097 systemd-networkd[1117]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:07:34.008204 systemd-networkd[1117]: eth0: DHCPv4 address 172.31.20.118/20, gateway 172.31.16.1 acquired from 172.31.16.1 Jan 13 20:07:34.323507 ignition[1042]: Ignition 2.20.0 Jan 13 20:07:34.324022 ignition[1042]: Stage: fetch-offline Jan 13 20:07:34.324505 ignition[1042]: no configs at "/usr/lib/ignition/base.d" Jan 13 20:07:34.329474 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 13 20:07:34.324529 ignition[1042]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jan 13 20:07:34.325844 ignition[1042]: Ignition finished successfully Jan 13 20:07:34.353451 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jan 13 20:07:34.378883 ignition[1127]: Ignition 2.20.0 Jan 13 20:07:34.378911 ignition[1127]: Stage: fetch Jan 13 20:07:34.379516 ignition[1127]: no configs at "/usr/lib/ignition/base.d" Jan 13 20:07:34.379541 ignition[1127]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jan 13 20:07:34.379706 ignition[1127]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jan 13 20:07:34.393205 ignition[1127]: PUT result: OK Jan 13 20:07:34.396583 ignition[1127]: parsed url from cmdline: "" Jan 13 20:07:34.396609 ignition[1127]: no config URL provided Jan 13 20:07:34.396624 ignition[1127]: reading system config file "/usr/lib/ignition/user.ign" Jan 13 20:07:34.396649 ignition[1127]: no config at "/usr/lib/ignition/user.ign" Jan 13 20:07:34.396684 ignition[1127]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jan 13 20:07:34.405637 ignition[1127]: PUT result: OK Jan 13 20:07:34.405726 ignition[1127]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Jan 13 20:07:34.410214 ignition[1127]: GET result: OK Jan 13 20:07:34.410359 ignition[1127]: parsing config with SHA512: 3fd60908aefb1313062ab272bad186728c898c4582d466b662b47a254048e93d87ef0769a37605e843730a784a8d7d6688b8a63a9e473f37afe1b30ebe29ee32 Jan 13 20:07:34.418351 unknown[1127]: fetched base config from "system" Jan 13 20:07:34.418961 ignition[1127]: fetch: fetch complete Jan 13 20:07:34.418368 unknown[1127]: fetched base config from "system" Jan 13 20:07:34.418971 ignition[1127]: fetch: fetch passed Jan 13 20:07:34.418382 unknown[1127]: fetched user config from "aws" Jan 13 20:07:34.419047 ignition[1127]: Ignition finished successfully Jan 13 20:07:34.433658 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jan 13 20:07:34.451505 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 13 20:07:34.473796 ignition[1134]: Ignition 2.20.0 Jan 13 20:07:34.473825 ignition[1134]: Stage: kargs Jan 13 20:07:34.474824 ignition[1134]: no configs at "/usr/lib/ignition/base.d" Jan 13 20:07:34.474849 ignition[1134]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jan 13 20:07:34.474998 ignition[1134]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jan 13 20:07:34.477543 ignition[1134]: PUT result: OK Jan 13 20:07:34.488957 ignition[1134]: kargs: kargs passed Jan 13 20:07:34.489312 ignition[1134]: Ignition finished successfully Jan 13 20:07:34.494688 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 13 20:07:34.510400 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 13 20:07:34.530921 ignition[1140]: Ignition 2.20.0 Jan 13 20:07:34.530955 ignition[1140]: Stage: disks Jan 13 20:07:34.531635 ignition[1140]: no configs at "/usr/lib/ignition/base.d" Jan 13 20:07:34.531661 ignition[1140]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Jan 13 20:07:34.531809 ignition[1140]: PUT http://169.254.169.254/latest/api/token: attempt #1 Jan 13 20:07:34.534838 ignition[1140]: PUT result: OK Jan 13 20:07:34.546164 ignition[1140]: disks: disks passed Jan 13 20:07:34.546568 ignition[1140]: Ignition finished successfully Jan 13 20:07:34.552192 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 13 20:07:34.557878 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 13 20:07:34.562839 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 13 20:07:34.563024 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 13 20:07:34.572934 systemd[1]: Reached target sysinit.target - System Initialization. Jan 13 20:07:34.575352 systemd[1]: Reached target basic.target - Basic System. Jan 13 20:07:34.592419 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 13 20:07:34.645626 systemd-fsck[1148]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jan 13 20:07:34.658462 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 13 20:07:34.669417 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 13 20:07:34.760179 kernel: EXT4-fs (nvme0n1p9): mounted filesystem f9a95e53-2d63-4443-b523-cb2108fb48f6 r/w with ordered data mode. Quota mode: none. Jan 13 20:07:34.761197 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 13 20:07:34.767269 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 13 20:07:34.797253 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 13 20:07:34.804544 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 13 20:07:34.812479 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 13 20:07:34.812747 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 13 20:07:34.840141 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/nvme0n1p6 scanned by mount (1167) Jan 13 20:07:34.840188 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 9f8ecb6c-ace6-4d16-8781-f4e964dc0779 Jan 13 20:07:34.840215 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Jan 13 20:07:34.840242 kernel: BTRFS info (device nvme0n1p6): using free space tree Jan 13 20:07:34.840268 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Jan 13 20:07:34.812795 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 13 20:07:34.849391 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 13 20:07:34.860596 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 13 20:07:34.876422 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 13 20:07:35.390147 initrd-setup-root[1191]: cut: /sysroot/etc/passwd: No such file or directory Jan 13 20:07:35.431017 initrd-setup-root[1198]: cut: /sysroot/etc/group: No such file or directory Jan 13 20:07:35.440356 initrd-setup-root[1205]: cut: /sysroot/etc/shadow: No such file or directory Jan 13 20:07:35.449378 initrd-setup-root[1212]: cut: /sysroot/etc/gshadow: No such file or directory Jan 13 20:07:35.818213 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 13 20:07:35.828481 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 13 20:07:35.829826 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 13 20:07:35.853798 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 13 20:07:35.858098 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 9f8ecb6c-ace6-4d16-8781-f4e964dc0779 Jan 13 20:07:35.892606 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 13 20:07:35.909153 ignition[1281]: INFO : Ignition 2.20.0 Jan 13 20:07:35.909153 ignition[1281]: INFO : Stage: mount Jan 13 20:07:35.914071 ignition[1281]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 13 20:07:35.914071 ignition[1281]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jan 13 20:07:35.914071 ignition[1281]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jan 13 20:07:35.923430 ignition[1281]: INFO : PUT result: OK Jan 13 20:07:35.927871 ignition[1281]: INFO : mount: mount passed Jan 13 20:07:35.929770 ignition[1281]: INFO : Ignition finished successfully Jan 13 20:07:35.934080 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 13 20:07:35.940481 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 13 20:07:35.961531 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 13 20:07:35.982173 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/nvme0n1p6 scanned by mount (1291) Jan 13 20:07:35.985643 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 9f8ecb6c-ace6-4d16-8781-f4e964dc0779 Jan 13 20:07:35.985681 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Jan 13 20:07:35.985799 systemd-networkd[1117]: eth0: Gained IPv6LL Jan 13 20:07:35.989885 kernel: BTRFS info (device nvme0n1p6): using free space tree Jan 13 20:07:35.994147 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Jan 13 20:07:35.997742 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 13 20:07:36.040156 ignition[1308]: INFO : Ignition 2.20.0 Jan 13 20:07:36.040156 ignition[1308]: INFO : Stage: files Jan 13 20:07:36.040156 ignition[1308]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 13 20:07:36.040156 ignition[1308]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jan 13 20:07:36.040156 ignition[1308]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jan 13 20:07:36.054244 ignition[1308]: INFO : PUT result: OK Jan 13 20:07:36.066314 ignition[1308]: DEBUG : files: compiled without relabeling support, skipping Jan 13 20:07:36.082894 ignition[1308]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 13 20:07:36.082894 ignition[1308]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 13 20:07:36.126045 ignition[1308]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 13 20:07:36.129650 ignition[1308]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 13 20:07:36.133549 unknown[1308]: wrote ssh authorized keys file for user: core Jan 13 20:07:36.135863 ignition[1308]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 13 20:07:36.152286 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jan 13 20:07:36.157338 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Jan 13 20:07:36.277202 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 13 20:07:36.432692 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jan 13 20:07:36.432692 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Jan 13 20:07:36.441067 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-arm64.raw: attempt #1 Jan 13 20:07:36.966200 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 13 20:07:37.329211 ignition[1308]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Jan 13 20:07:37.329211 ignition[1308]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 13 20:07:37.345270 ignition[1308]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 13 20:07:37.345270 ignition[1308]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 13 20:07:37.345270 ignition[1308]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 13 20:07:37.345270 ignition[1308]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Jan 13 20:07:37.345270 ignition[1308]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Jan 13 20:07:37.345270 ignition[1308]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 13 20:07:37.366936 ignition[1308]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 13 20:07:37.366936 ignition[1308]: INFO : files: files passed Jan 13 20:07:37.372727 ignition[1308]: INFO : Ignition finished successfully Jan 13 20:07:37.387732 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 13 20:07:37.409515 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 13 20:07:37.418459 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 13 20:07:37.424678 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 13 20:07:37.425033 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 13 20:07:37.453892 initrd-setup-root-after-ignition[1337]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 13 20:07:37.453892 initrd-setup-root-after-ignition[1337]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 13 20:07:37.463509 initrd-setup-root-after-ignition[1341]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 13 20:07:37.470095 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 13 20:07:37.475838 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 13 20:07:37.501497 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 13 20:07:37.555503 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 13 20:07:37.555779 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 13 20:07:37.566015 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 13 20:07:37.569578 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 13 20:07:37.572129 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 13 20:07:37.592376 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 13 20:07:37.617793 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 13 20:07:37.630544 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 13 20:07:37.659199 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 13 20:07:37.662819 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 13 20:07:37.666276 systemd[1]: Stopped target timers.target - Timer Units. Jan 13 20:07:37.673968 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 13 20:07:37.674228 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 13 20:07:37.682465 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 13 20:07:37.685047 systemd[1]: Stopped target basic.target - Basic System. Jan 13 20:07:37.687295 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 13 20:07:37.690074 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 13 20:07:37.701874 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 13 20:07:37.705042 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 13 20:07:37.711733 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 13 20:07:37.714948 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 13 20:07:37.722280 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 13 20:07:37.725007 systemd[1]: Stopped target swap.target - Swaps. Jan 13 20:07:37.730799 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 13 20:07:37.731028 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 13 20:07:37.733999 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 13 20:07:37.743411 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 13 20:07:37.746387 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 13 20:07:37.755244 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 13 20:07:37.755556 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 13 20:07:37.755779 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 13 20:07:37.764510 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 13 20:07:37.764734 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 13 20:07:37.767906 systemd[1]: ignition-files.service: Deactivated successfully. Jan 13 20:07:37.768108 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 13 20:07:37.799310 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 13 20:07:37.801314 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 13 20:07:37.801586 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 13 20:07:37.810676 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 13 20:07:37.815479 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 13 20:07:37.816353 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 13 20:07:37.825526 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 13 20:07:37.825755 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 13 20:07:37.842319 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 13 20:07:37.842513 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 13 20:07:37.862658 ignition[1361]: INFO : Ignition 2.20.0 Jan 13 20:07:37.862658 ignition[1361]: INFO : Stage: umount Jan 13 20:07:37.862658 ignition[1361]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 13 20:07:37.862658 ignition[1361]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Jan 13 20:07:37.862658 ignition[1361]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Jan 13 20:07:37.862658 ignition[1361]: INFO : PUT result: OK Jan 13 20:07:37.892502 ignition[1361]: INFO : umount: umount passed Jan 13 20:07:37.892502 ignition[1361]: INFO : Ignition finished successfully Jan 13 20:07:37.874149 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 13 20:07:37.874342 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 13 20:07:37.880431 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 13 20:07:37.892944 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 13 20:07:37.893316 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 13 20:07:37.903514 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 13 20:07:37.904241 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 13 20:07:37.910630 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 13 20:07:37.910724 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 13 20:07:37.913025 systemd[1]: ignition-fetch.service: Deactivated successfully. Jan 13 20:07:37.913102 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jan 13 20:07:37.915403 systemd[1]: Stopped target network.target - Network. Jan 13 20:07:37.917477 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 13 20:07:37.917569 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 13 20:07:37.920348 systemd[1]: Stopped target paths.target - Path Units. Jan 13 20:07:37.922488 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 13 20:07:37.951154 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 13 20:07:37.954040 systemd[1]: Stopped target slices.target - Slice Units. Jan 13 20:07:37.960023 systemd[1]: Stopped target sockets.target - Socket Units. Jan 13 20:07:37.962011 systemd[1]: iscsid.socket: Deactivated successfully. Jan 13 20:07:37.962096 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 13 20:07:37.964147 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 13 20:07:37.964218 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 13 20:07:37.966893 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 13 20:07:37.966979 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 13 20:07:37.969527 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 13 20:07:37.969677 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 13 20:07:37.973898 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 13 20:07:37.973978 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 13 20:07:37.977186 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 13 20:07:37.995960 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 13 20:07:38.011196 systemd-networkd[1117]: eth0: DHCPv6 lease lost Jan 13 20:07:38.013850 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 13 20:07:38.014155 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 13 20:07:38.017947 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 13 20:07:38.018018 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 13 20:07:38.043424 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 13 20:07:38.045771 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 13 20:07:38.045880 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 13 20:07:38.048904 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 13 20:07:38.051788 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 13 20:07:38.051979 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 13 20:07:38.055742 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 13 20:07:38.055892 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:07:38.060629 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 13 20:07:38.060742 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 13 20:07:38.089865 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 13 20:07:38.089978 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 13 20:07:38.110769 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 13 20:07:38.111323 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 13 20:07:38.121327 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 13 20:07:38.121784 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 13 20:07:38.126939 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 13 20:07:38.127022 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 13 20:07:38.129552 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 13 20:07:38.129744 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 13 20:07:38.141791 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 13 20:07:38.141882 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 13 20:07:38.144636 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 13 20:07:38.144717 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 13 20:07:38.158340 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 13 20:07:38.158432 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 13 20:07:38.175513 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 13 20:07:38.178351 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 13 20:07:38.178458 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 13 20:07:38.181430 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 13 20:07:38.181513 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:07:38.209145 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 13 20:07:38.209501 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 13 20:07:38.217327 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 13 20:07:38.223451 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 13 20:07:38.245013 systemd[1]: Switching root. Jan 13 20:07:38.304360 systemd-journald[251]: Journal stopped Jan 13 20:07:41.600736 systemd-journald[251]: Received SIGTERM from PID 1 (systemd). Jan 13 20:07:41.600882 kernel: SELinux: policy capability network_peer_controls=1 Jan 13 20:07:41.600945 kernel: SELinux: policy capability open_perms=1 Jan 13 20:07:41.600978 kernel: SELinux: policy capability extended_socket_class=1 Jan 13 20:07:41.601017 kernel: SELinux: policy capability always_check_network=0 Jan 13 20:07:41.601054 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 13 20:07:41.601095 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 13 20:07:41.606797 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 13 20:07:41.607097 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 13 20:07:41.608328 kernel: audit: type=1403 audit(1736798859.637:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 13 20:07:41.608367 systemd[1]: Successfully loaded SELinux policy in 94.170ms. Jan 13 20:07:41.608415 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 28.281ms. Jan 13 20:07:41.608449 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 13 20:07:41.608488 systemd[1]: Detected virtualization amazon. Jan 13 20:07:41.608520 systemd[1]: Detected architecture arm64. Jan 13 20:07:41.608550 systemd[1]: Detected first boot. Jan 13 20:07:41.608581 systemd[1]: Initializing machine ID from VM UUID. Jan 13 20:07:41.608614 zram_generator::config[1404]: No configuration found. Jan 13 20:07:41.608648 systemd[1]: Populated /etc with preset unit settings. Jan 13 20:07:41.608679 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 13 20:07:41.608712 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 13 20:07:41.608745 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 13 20:07:41.608781 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 13 20:07:41.608813 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 13 20:07:41.608843 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 13 20:07:41.608874 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 13 20:07:41.611630 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 13 20:07:41.611691 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 13 20:07:41.611811 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 13 20:07:41.611851 systemd[1]: Created slice user.slice - User and Session Slice. Jan 13 20:07:41.611889 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 13 20:07:41.611922 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 13 20:07:41.611953 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 13 20:07:41.611987 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 13 20:07:41.612016 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 13 20:07:41.612047 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 13 20:07:41.612079 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 13 20:07:41.612128 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 13 20:07:41.612190 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 13 20:07:41.612227 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 13 20:07:41.612257 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 13 20:07:41.612290 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 13 20:07:41.612318 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 13 20:07:41.612350 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 13 20:07:41.612381 systemd[1]: Reached target slices.target - Slice Units. Jan 13 20:07:41.612413 systemd[1]: Reached target swap.target - Swaps. Jan 13 20:07:41.612445 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 13 20:07:41.612481 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 13 20:07:41.612512 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 13 20:07:41.612541 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 13 20:07:41.612572 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 13 20:07:41.612600 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 13 20:07:41.612631 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 13 20:07:41.612663 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 13 20:07:41.612692 systemd[1]: Mounting media.mount - External Media Directory... Jan 13 20:07:41.612731 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 13 20:07:41.612764 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 13 20:07:41.612794 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 13 20:07:41.612825 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 13 20:07:41.612854 systemd[1]: Reached target machines.target - Containers. Jan 13 20:07:41.612883 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 13 20:07:41.612935 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 13 20:07:41.612966 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 13 20:07:41.612994 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 13 20:07:41.613028 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 13 20:07:41.613059 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 13 20:07:41.613091 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 13 20:07:41.614177 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 13 20:07:41.614220 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 13 20:07:41.614253 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 13 20:07:41.614283 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 13 20:07:41.614311 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 13 20:07:41.614345 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 13 20:07:41.614381 systemd[1]: Stopped systemd-fsck-usr.service. Jan 13 20:07:41.614409 kernel: fuse: init (API version 7.39) Jan 13 20:07:41.614437 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 13 20:07:41.614468 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 13 20:07:41.614495 kernel: loop: module loaded Jan 13 20:07:41.614522 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 13 20:07:41.614551 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 13 20:07:41.614582 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 13 20:07:41.614612 systemd[1]: verity-setup.service: Deactivated successfully. Jan 13 20:07:41.614645 systemd[1]: Stopped verity-setup.service. Jan 13 20:07:41.614674 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 13 20:07:41.614702 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 13 20:07:41.614731 systemd[1]: Mounted media.mount - External Media Directory. Jan 13 20:07:41.614760 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 13 20:07:41.614789 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 13 20:07:41.614818 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 13 20:07:41.614852 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 13 20:07:41.614881 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 13 20:07:41.614909 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 13 20:07:41.614937 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 13 20:07:41.614966 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 13 20:07:41.614994 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 13 20:07:41.615028 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 13 20:07:41.615057 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 13 20:07:41.615085 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 13 20:07:41.619402 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 13 20:07:41.623752 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 13 20:07:41.623796 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 13 20:07:41.623836 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 13 20:07:41.623867 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 13 20:07:41.623898 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 13 20:07:41.623926 kernel: ACPI: bus type drm_connector registered Jan 13 20:07:41.623954 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 13 20:07:41.624024 systemd-journald[1486]: Collecting audit messages is disabled. Jan 13 20:07:41.624072 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 13 20:07:41.624107 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 13 20:07:41.625211 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 13 20:07:41.625248 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 13 20:07:41.625278 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 13 20:07:41.625307 systemd-journald[1486]: Journal started Jan 13 20:07:41.625364 systemd-journald[1486]: Runtime Journal (/run/log/journal/ec203ce844de19d765bb1c1bc41889cd) is 8.0M, max 75.3M, 67.3M free. Jan 13 20:07:40.892102 systemd[1]: Queued start job for default target multi-user.target. Jan 13 20:07:40.964803 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Jan 13 20:07:40.965596 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 13 20:07:41.634890 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 13 20:07:41.642231 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 13 20:07:41.657882 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 13 20:07:41.657966 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 13 20:07:41.678787 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 13 20:07:41.678940 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 13 20:07:41.694179 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 13 20:07:41.706147 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 13 20:07:41.718137 systemd[1]: Started systemd-journald.service - Journal Service. Jan 13 20:07:41.719819 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 13 20:07:41.731035 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 13 20:07:41.732426 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 13 20:07:41.738473 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 13 20:07:41.743499 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 13 20:07:41.748879 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 13 20:07:41.756505 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 13 20:07:41.799186 kernel: loop0: detected capacity change from 0 to 113552 Jan 13 20:07:41.815904 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 13 20:07:41.831527 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 13 20:07:41.848577 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 13 20:07:41.861444 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 13 20:07:41.873723 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 13 20:07:41.911004 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 13 20:07:41.914498 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 13 20:07:41.926434 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 13 20:07:41.939330 systemd-journald[1486]: Time spent on flushing to /var/log/journal/ec203ce844de19d765bb1c1bc41889cd is 35.759ms for 913 entries. Jan 13 20:07:41.939330 systemd-journald[1486]: System Journal (/var/log/journal/ec203ce844de19d765bb1c1bc41889cd) is 8.0M, max 195.6M, 187.6M free. Jan 13 20:07:41.997262 systemd-journald[1486]: Received client request to flush runtime journal. Jan 13 20:07:41.997338 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 13 20:07:41.946464 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 13 20:07:42.002652 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 13 20:07:42.016758 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 13 20:07:42.028143 kernel: loop1: detected capacity change from 0 to 189592 Jan 13 20:07:42.037911 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 13 20:07:42.045705 udevadm[1546]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jan 13 20:07:42.100166 kernel: loop2: detected capacity change from 0 to 116784 Jan 13 20:07:42.124305 systemd-tmpfiles[1552]: ACLs are not supported, ignoring. Jan 13 20:07:42.124837 systemd-tmpfiles[1552]: ACLs are not supported, ignoring. Jan 13 20:07:42.139226 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 13 20:07:42.266314 kernel: loop3: detected capacity change from 0 to 53784 Jan 13 20:07:42.425160 kernel: loop4: detected capacity change from 0 to 113552 Jan 13 20:07:42.452728 kernel: loop5: detected capacity change from 0 to 189592 Jan 13 20:07:42.478158 kernel: loop6: detected capacity change from 0 to 116784 Jan 13 20:07:42.497467 kernel: loop7: detected capacity change from 0 to 53784 Jan 13 20:07:42.514634 (sd-merge)[1559]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Jan 13 20:07:42.515636 (sd-merge)[1559]: Merged extensions into '/usr'. Jan 13 20:07:42.523766 systemd[1]: Reloading requested from client PID 1515 ('systemd-sysext') (unit systemd-sysext.service)... Jan 13 20:07:42.523922 systemd[1]: Reloading... Jan 13 20:07:42.679165 zram_generator::config[1581]: No configuration found. Jan 13 20:07:43.016719 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 13 20:07:43.130057 systemd[1]: Reloading finished in 605 ms. Jan 13 20:07:43.172210 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 13 20:07:43.176055 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 13 20:07:43.195447 systemd[1]: Starting ensure-sysext.service... Jan 13 20:07:43.202470 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 13 20:07:43.220416 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 13 20:07:43.231025 systemd[1]: Reloading requested from client PID 1637 ('systemctl') (unit ensure-sysext.service)... Jan 13 20:07:43.231051 systemd[1]: Reloading... Jan 13 20:07:43.289465 systemd-tmpfiles[1638]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 13 20:07:43.289999 systemd-tmpfiles[1638]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 13 20:07:43.292999 systemd-tmpfiles[1638]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 13 20:07:43.294758 systemd-tmpfiles[1638]: ACLs are not supported, ignoring. Jan 13 20:07:43.294925 systemd-tmpfiles[1638]: ACLs are not supported, ignoring. Jan 13 20:07:43.305750 systemd-tmpfiles[1638]: Detected autofs mount point /boot during canonicalization of boot. Jan 13 20:07:43.305778 systemd-tmpfiles[1638]: Skipping /boot Jan 13 20:07:43.309177 systemd-udevd[1639]: Using default interface naming scheme 'v255'. Jan 13 20:07:43.340834 systemd-tmpfiles[1638]: Detected autofs mount point /boot during canonicalization of boot. Jan 13 20:07:43.340866 systemd-tmpfiles[1638]: Skipping /boot Jan 13 20:07:43.461470 zram_generator::config[1674]: No configuration found. Jan 13 20:07:43.594219 (udev-worker)[1702]: Network interface NamePolicy= disabled on kernel command line. Jan 13 20:07:43.715541 ldconfig[1511]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 13 20:07:43.829826 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 13 20:07:43.932152 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (1702) Jan 13 20:07:43.993939 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 13 20:07:43.994627 systemd[1]: Reloading finished in 762 ms. Jan 13 20:07:44.023262 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 13 20:07:44.029734 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 13 20:07:44.043293 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 13 20:07:44.101218 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 13 20:07:44.121882 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 13 20:07:44.133007 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 13 20:07:44.164757 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 13 20:07:44.187622 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 13 20:07:44.195874 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 13 20:07:44.213664 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 13 20:07:44.274494 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 13 20:07:44.283360 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 13 20:07:44.300801 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 13 20:07:44.311808 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 13 20:07:44.316444 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 13 20:07:44.330074 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 13 20:07:44.340403 augenrules[1865]: No rules Jan 13 20:07:44.338265 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 13 20:07:44.346679 systemd[1]: audit-rules.service: Deactivated successfully. Jan 13 20:07:44.347260 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 13 20:07:44.354331 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 13 20:07:44.360809 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 13 20:07:44.363600 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 13 20:07:44.365219 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 13 20:07:44.367464 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 13 20:07:44.367861 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 13 20:07:44.369071 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 13 20:07:44.370538 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 13 20:07:44.389998 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Jan 13 20:07:44.419433 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 13 20:07:44.432107 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 13 20:07:44.432799 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 13 20:07:44.433165 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 13 20:07:44.445410 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 13 20:07:44.467259 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 13 20:07:44.476829 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 13 20:07:44.487627 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 13 20:07:44.497668 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 13 20:07:44.514719 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 13 20:07:44.523230 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 13 20:07:44.526442 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 13 20:07:44.526830 systemd[1]: Reached target time-set.target - System Time Set. Jan 13 20:07:44.531503 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 13 20:07:44.537352 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 13 20:07:44.537711 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 13 20:07:44.544012 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 13 20:07:44.549870 systemd[1]: Finished ensure-sysext.service. Jan 13 20:07:44.562158 lvm[1877]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 13 20:07:44.565346 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 13 20:07:44.565706 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 13 20:07:44.568668 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 13 20:07:44.576622 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 13 20:07:44.577088 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 13 20:07:44.606398 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 13 20:07:44.611989 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 13 20:07:44.624759 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 13 20:07:44.630463 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 13 20:07:44.633315 augenrules[1880]: /sbin/augenrules: No change Jan 13 20:07:44.657430 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 13 20:07:44.657828 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 13 20:07:44.663396 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 13 20:07:44.670358 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 13 20:07:44.673040 augenrules[1918]: No rules Jan 13 20:07:44.684549 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 13 20:07:44.689508 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 13 20:07:44.692447 systemd[1]: audit-rules.service: Deactivated successfully. Jan 13 20:07:44.692860 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 13 20:07:44.719147 lvm[1925]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 13 20:07:44.778236 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 13 20:07:44.785659 systemd-resolved[1843]: Positive Trust Anchors: Jan 13 20:07:44.785704 systemd-resolved[1843]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 13 20:07:44.785769 systemd-resolved[1843]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 13 20:07:44.794320 systemd-resolved[1843]: Defaulting to hostname 'linux'. Jan 13 20:07:44.797033 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 13 20:07:44.800169 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 13 20:07:44.802957 systemd[1]: Reached target sysinit.target - System Initialization. Jan 13 20:07:44.805669 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 13 20:07:44.808640 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 13 20:07:44.811889 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 13 20:07:44.814755 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 13 20:07:44.817713 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 13 20:07:44.820718 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 13 20:07:44.820795 systemd[1]: Reached target paths.target - Path Units. Jan 13 20:07:44.825334 systemd-networkd[1836]: lo: Link UP Jan 13 20:07:44.825351 systemd-networkd[1836]: lo: Gained carrier Jan 13 20:07:44.828431 systemd-networkd[1836]: Enumeration completed Jan 13 20:07:44.829875 systemd-networkd[1836]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:07:44.830002 systemd-networkd[1836]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 13 20:07:44.832371 systemd[1]: Reached target timers.target - Timer Units. Jan 13 20:07:44.836466 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 13 20:07:44.836909 systemd-networkd[1836]: eth0: Link UP Jan 13 20:07:44.840345 systemd-networkd[1836]: eth0: Gained carrier Jan 13 20:07:44.840390 systemd-networkd[1836]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 13 20:07:44.842217 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 13 20:07:44.854904 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 13 20:07:44.859313 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 13 20:07:44.862411 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 13 20:07:44.865799 systemd[1]: Reached target network.target - Network. Jan 13 20:07:44.868026 systemd[1]: Reached target sockets.target - Socket Units. Jan 13 20:07:44.870484 systemd[1]: Reached target basic.target - Basic System. Jan 13 20:07:44.872728 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 13 20:07:44.872779 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 13 20:07:44.875239 systemd-networkd[1836]: eth0: DHCPv4 address 172.31.20.118/20, gateway 172.31.16.1 acquired from 172.31.16.1 Jan 13 20:07:44.882402 systemd[1]: Starting containerd.service - containerd container runtime... Jan 13 20:07:44.892145 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jan 13 20:07:44.898462 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 13 20:07:44.913522 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 13 20:07:44.922446 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 13 20:07:44.926872 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 13 20:07:44.931480 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 13 20:07:44.941442 systemd[1]: Started ntpd.service - Network Time Service. Jan 13 20:07:44.944941 jq[1934]: false Jan 13 20:07:44.951365 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 13 20:07:44.963489 systemd[1]: Starting setup-oem.service - Setup OEM... Jan 13 20:07:44.971421 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 13 20:07:44.976736 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 13 20:07:44.988561 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 13 20:07:44.997414 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 13 20:07:45.001095 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 13 20:07:45.001982 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 13 20:07:45.006138 systemd[1]: Starting update-engine.service - Update Engine... Jan 13 20:07:45.015353 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 13 20:07:45.026102 dbus-daemon[1933]: [system] SELinux support is enabled Jan 13 20:07:45.030700 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 13 20:07:45.039614 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 13 20:07:45.040231 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 13 20:07:45.042932 dbus-daemon[1933]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.1' (uid=244 pid=1836 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Jan 13 20:07:45.066522 dbus-daemon[1933]: [system] Successfully activated service 'org.freedesktop.systemd1' Jan 13 20:07:45.082373 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 13 20:07:45.084245 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 13 20:07:45.100437 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Jan 13 20:07:45.106244 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 13 20:07:45.106297 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 13 20:07:45.115929 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 13 20:07:45.117239 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 13 20:07:45.200440 systemd[1]: motdgen.service: Deactivated successfully. Jan 13 20:07:45.207889 jq[1948]: true Jan 13 20:07:45.203090 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 13 20:07:45.203810 (ntainerd)[1972]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 13 20:07:45.236213 extend-filesystems[1935]: Found loop4 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found loop5 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found loop6 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found loop7 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1p1 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1p2 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1p3 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found usr Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1p4 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1p6 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1p7 Jan 13 20:07:45.236213 extend-filesystems[1935]: Found nvme0n1p9 Jan 13 20:07:45.223837 systemd[1]: Finished setup-oem.service - Setup OEM. Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: ntpd 4.2.8p17@1.4004-o Mon Jan 13 18:25:48 UTC 2025 (1): Starting Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: ---------------------------------------------------- Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: ntp-4 is maintained by Network Time Foundation, Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: corporation. Support and training for ntp-4 are Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: available at https://www.nwtime.org/support Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: ---------------------------------------------------- Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: proto: precision = 0.096 usec (-23) Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: basedate set to 2025-01-01 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: gps base set to 2025-01-05 (week 2348) Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Listen and drop on 0 v6wildcard [::]:123 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Listen normally on 2 lo 127.0.0.1:123 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Listen normally on 3 eth0 172.31.20.118:123 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Listen normally on 4 lo [::1]:123 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: bind(21) AF_INET6 fe80::4db:2bff:fee0:d9ff%2#123 flags 0x11 failed: Cannot assign requested address Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: unable to create socket on eth0 (5) for fe80::4db:2bff:fee0:d9ff%2#123 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: failed to init interface for address fe80::4db:2bff:fee0:d9ff%2 Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: Listening on routing socket on fd #21 for interface updates Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jan 13 20:07:45.304061 ntpd[1937]: 13 Jan 20:07:45 ntpd[1937]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jan 13 20:07:45.323677 extend-filesystems[1935]: Checking size of /dev/nvme0n1p9 Jan 13 20:07:45.247568 ntpd[1937]: ntpd 4.2.8p17@1.4004-o Mon Jan 13 18:25:48 UTC 2025 (1): Starting Jan 13 20:07:45.247616 ntpd[1937]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Jan 13 20:07:45.247635 ntpd[1937]: ---------------------------------------------------- Jan 13 20:07:45.247654 ntpd[1937]: ntp-4 is maintained by Network Time Foundation, Jan 13 20:07:45.247672 ntpd[1937]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Jan 13 20:07:45.247690 ntpd[1937]: corporation. Support and training for ntp-4 are Jan 13 20:07:45.247707 ntpd[1937]: available at https://www.nwtime.org/support Jan 13 20:07:45.247725 ntpd[1937]: ---------------------------------------------------- Jan 13 20:07:45.251956 ntpd[1937]: proto: precision = 0.096 usec (-23) Jan 13 20:07:45.254084 ntpd[1937]: basedate set to 2025-01-01 Jan 13 20:07:45.254164 ntpd[1937]: gps base set to 2025-01-05 (week 2348) Jan 13 20:07:45.258724 ntpd[1937]: Listen and drop on 0 v6wildcard [::]:123 Jan 13 20:07:45.258808 ntpd[1937]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Jan 13 20:07:45.259078 ntpd[1937]: Listen normally on 2 lo 127.0.0.1:123 Jan 13 20:07:45.260174 ntpd[1937]: Listen normally on 3 eth0 172.31.20.118:123 Jan 13 20:07:45.260271 ntpd[1937]: Listen normally on 4 lo [::1]:123 Jan 13 20:07:45.260365 ntpd[1937]: bind(21) AF_INET6 fe80::4db:2bff:fee0:d9ff%2#123 flags 0x11 failed: Cannot assign requested address Jan 13 20:07:45.260407 ntpd[1937]: unable to create socket on eth0 (5) for fe80::4db:2bff:fee0:d9ff%2#123 Jan 13 20:07:45.260434 ntpd[1937]: failed to init interface for address fe80::4db:2bff:fee0:d9ff%2 Jan 13 20:07:45.260501 ntpd[1937]: Listening on routing socket on fd #21 for interface updates Jan 13 20:07:45.266713 ntpd[1937]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jan 13 20:07:45.266768 ntpd[1937]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Jan 13 20:07:45.333179 systemd-logind[1942]: Watching system buttons on /dev/input/event0 (Power Button) Jan 13 20:07:45.333259 systemd-logind[1942]: Watching system buttons on /dev/input/event1 (Sleep Button) Jan 13 20:07:45.366498 tar[1968]: linux-arm64/helm Jan 13 20:07:45.366834 jq[1976]: true Jan 13 20:07:45.367171 update_engine[1944]: I20250113 20:07:45.353299 1944 main.cc:92] Flatcar Update Engine starting Jan 13 20:07:45.334064 systemd-logind[1942]: New seat seat0. Jan 13 20:07:45.384254 update_engine[1944]: I20250113 20:07:45.381533 1944 update_check_scheduler.cc:74] Next update check in 9m14s Jan 13 20:07:45.387577 extend-filesystems[1935]: Resized partition /dev/nvme0n1p9 Jan 13 20:07:45.429549 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Jan 13 20:07:45.429649 extend-filesystems[1987]: resize2fs 1.47.1 (20-May-2024) Jan 13 20:07:45.388715 systemd[1]: Started update-engine.service - Update Engine. Jan 13 20:07:45.437962 systemd[1]: Started systemd-logind.service - User Login Management. Jan 13 20:07:45.460735 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch failed with 404: resource not found Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Jan 13 20:07:45.471213 coreos-metadata[1932]: Jan 13 20:07:45.470 INFO Fetch successful Jan 13 20:07:45.482178 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Jan 13 20:07:45.535931 extend-filesystems[1987]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Jan 13 20:07:45.535931 extend-filesystems[1987]: old_desc_blocks = 1, new_desc_blocks = 1 Jan 13 20:07:45.535931 extend-filesystems[1987]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Jan 13 20:07:45.554635 extend-filesystems[1935]: Resized filesystem in /dev/nvme0n1p9 Jan 13 20:07:45.565409 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 13 20:07:45.565780 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 13 20:07:45.622473 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jan 13 20:07:45.628049 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 13 20:07:45.650739 dbus-daemon[1933]: [system] Successfully activated service 'org.freedesktop.hostname1' Jan 13 20:07:45.651527 dbus-daemon[1933]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=1963 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Jan 13 20:07:45.683797 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Jan 13 20:07:45.692618 bash[2020]: Updated "/home/core/.ssh/authorized_keys" Jan 13 20:07:45.695497 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (1704) Jan 13 20:07:45.708241 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 13 20:07:45.723591 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 13 20:07:45.781880 systemd[1]: Starting polkit.service - Authorization Manager... Jan 13 20:07:45.793450 systemd[1]: Starting sshkeys.service... Jan 13 20:07:45.821739 locksmithd[1989]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 13 20:07:45.862587 polkitd[2037]: Started polkitd version 121 Jan 13 20:07:45.906244 systemd-networkd[1836]: eth0: Gained IPv6LL Jan 13 20:07:45.915293 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jan 13 20:07:45.929932 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jan 13 20:07:45.946993 polkitd[2037]: Loading rules from directory /etc/polkit-1/rules.d Jan 13 20:07:45.949574 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 13 20:07:45.947144 polkitd[2037]: Loading rules from directory /usr/share/polkit-1/rules.d Jan 13 20:07:45.957659 systemd[1]: Reached target network-online.target - Network is Online. Jan 13 20:07:45.960377 polkitd[2037]: Finished loading, compiling and executing 2 rules Jan 13 20:07:45.968780 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Jan 13 20:07:45.970211 dbus-daemon[1933]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Jan 13 20:07:45.971058 polkitd[2037]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Jan 13 20:07:45.978652 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:07:45.982802 containerd[1972]: time="2025-01-13T20:07:45.982675285Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Jan 13 20:07:45.987720 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 13 20:07:45.999634 systemd[1]: Started polkit.service - Authorization Manager. Jan 13 20:07:46.070887 systemd-hostnamed[1963]: Hostname set to (transient) Jan 13 20:07:46.079229 systemd-resolved[1843]: System hostname changed to 'ip-172-31-20-118'. Jan 13 20:07:46.202266 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 13 20:07:46.211321 containerd[1972]: time="2025-01-13T20:07:46.207850570Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:07:46.217187 containerd[1972]: time="2025-01-13T20:07:46.216885011Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.71-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:07:46.217187 containerd[1972]: time="2025-01-13T20:07:46.216959339Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 13 20:07:46.217187 containerd[1972]: time="2025-01-13T20:07:46.216998771Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 13 20:07:46.217424 containerd[1972]: time="2025-01-13T20:07:46.217397879Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 13 20:07:46.217474 containerd[1972]: time="2025-01-13T20:07:46.217440659Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 13 20:07:46.219712 containerd[1972]: time="2025-01-13T20:07:46.217599503Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:07:46.219712 containerd[1972]: time="2025-01-13T20:07:46.217640255Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:07:46.219712 containerd[1972]: time="2025-01-13T20:07:46.217968347Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:07:46.219712 containerd[1972]: time="2025-01-13T20:07:46.218003267Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 13 20:07:46.219712 containerd[1972]: time="2025-01-13T20:07:46.218036507Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:07:46.219712 containerd[1972]: time="2025-01-13T20:07:46.218060771Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 13 20:07:46.242167 containerd[1972]: time="2025-01-13T20:07:46.240327575Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:07:46.242167 containerd[1972]: time="2025-01-13T20:07:46.240768035Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 13 20:07:46.242167 containerd[1972]: time="2025-01-13T20:07:46.241023815Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 13 20:07:46.242167 containerd[1972]: time="2025-01-13T20:07:46.241056167Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 13 20:07:46.248994 containerd[1972]: time="2025-01-13T20:07:46.248666243Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 13 20:07:46.248994 containerd[1972]: time="2025-01-13T20:07:46.248807447Z" level=info msg="metadata content store policy set" policy=shared Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.280942991Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.281041835Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.281077199Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.281133095Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.281174027Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.281438099Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.281838863Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.282042119Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.282076703Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.282133115Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.282171299Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.282202643Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.282232319Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.283451 containerd[1972]: time="2025-01-13T20:07:46.282265235Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282298283Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282329171Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282359399Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282386219Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282427811Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282459419Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282489047Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282520019Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282548471Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282596471Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282625259Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282654971Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282685007Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285050 containerd[1972]: time="2025-01-13T20:07:46.282719807Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285691 containerd[1972]: time="2025-01-13T20:07:46.282750407Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285691 containerd[1972]: time="2025-01-13T20:07:46.282780047Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285691 containerd[1972]: time="2025-01-13T20:07:46.282809099Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285691 containerd[1972]: time="2025-01-13T20:07:46.282840515Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 13 20:07:46.285691 containerd[1972]: time="2025-01-13T20:07:46.282884819Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285691 containerd[1972]: time="2025-01-13T20:07:46.282920039Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.285691 containerd[1972]: time="2025-01-13T20:07:46.282947351Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 13 20:07:46.301135 amazon-ssm-agent[2100]: Initializing new seelog logger Jan 13 20:07:46.301135 amazon-ssm-agent[2100]: New Seelog Logger Creation Complete Jan 13 20:07:46.301135 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.301135 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.301135 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 processing appconfig overrides Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295282019Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295342919Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295372691Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295402271Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295432295Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295465391Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295489775Z" level=info msg="NRI interface is disabled by configuration." Jan 13 20:07:46.301757 containerd[1972]: time="2025-01-13T20:07:46.295515755Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 13 20:07:46.305147 containerd[1972]: time="2025-01-13T20:07:46.296038931Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 13 20:07:46.305147 containerd[1972]: time="2025-01-13T20:07:46.303628751Z" level=info msg="Connect containerd service" Jan 13 20:07:46.305147 containerd[1972]: time="2025-01-13T20:07:46.303754211Z" level=info msg="using legacy CRI server" Jan 13 20:07:46.305147 containerd[1972]: time="2025-01-13T20:07:46.303774323Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 13 20:07:46.305147 containerd[1972]: time="2025-01-13T20:07:46.304008647Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 13 20:07:46.305147 containerd[1972]: time="2025-01-13T20:07:46.305069015Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 13 20:07:46.314033 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.314033 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.314033 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 processing appconfig overrides Jan 13 20:07:46.315702 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO Proxy environment variables: Jan 13 20:07:46.319882 coreos-metadata[2090]: Jan 13 20:07:46.316 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Jan 13 20:07:46.320420 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.320420 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.320420 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 processing appconfig overrides Jan 13 20:07:46.323315 coreos-metadata[2090]: Jan 13 20:07:46.320 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Jan 13 20:07:46.323425 containerd[1972]: time="2025-01-13T20:07:46.322535495Z" level=info msg="Start subscribing containerd event" Jan 13 20:07:46.323425 containerd[1972]: time="2025-01-13T20:07:46.322624247Z" level=info msg="Start recovering state" Jan 13 20:07:46.323425 containerd[1972]: time="2025-01-13T20:07:46.322743443Z" level=info msg="Start event monitor" Jan 13 20:07:46.323425 containerd[1972]: time="2025-01-13T20:07:46.322769243Z" level=info msg="Start snapshots syncer" Jan 13 20:07:46.323425 containerd[1972]: time="2025-01-13T20:07:46.322800239Z" level=info msg="Start cni network conf syncer for default" Jan 13 20:07:46.323425 containerd[1972]: time="2025-01-13T20:07:46.322821887Z" level=info msg="Start streaming server" Jan 13 20:07:46.329287 coreos-metadata[2090]: Jan 13 20:07:46.326 INFO Fetch successful Jan 13 20:07:46.329287 coreos-metadata[2090]: Jan 13 20:07:46.326 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Jan 13 20:07:46.329287 coreos-metadata[2090]: Jan 13 20:07:46.329 INFO Fetch successful Jan 13 20:07:46.333156 containerd[1972]: time="2025-01-13T20:07:46.332356607Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 13 20:07:46.333156 containerd[1972]: time="2025-01-13T20:07:46.332511815Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 13 20:07:46.333156 containerd[1972]: time="2025-01-13T20:07:46.332623091Z" level=info msg="containerd successfully booted in 0.364005s" Jan 13 20:07:46.332756 systemd[1]: Started containerd.service - containerd container runtime. Jan 13 20:07:46.340896 unknown[2090]: wrote ssh authorized keys file for user: core Jan 13 20:07:46.353808 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.353808 amazon-ssm-agent[2100]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Jan 13 20:07:46.353808 amazon-ssm-agent[2100]: 2025/01/13 20:07:46 processing appconfig overrides Jan 13 20:07:46.401152 update-ssh-keys[2147]: Updated "/home/core/.ssh/authorized_keys" Jan 13 20:07:46.403438 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jan 13 20:07:46.416491 systemd[1]: Finished sshkeys.service. Jan 13 20:07:46.423220 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO https_proxy: Jan 13 20:07:46.522864 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO http_proxy: Jan 13 20:07:46.625294 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO no_proxy: Jan 13 20:07:46.724587 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO Checking if agent identity type OnPrem can be assumed Jan 13 20:07:46.825221 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO Checking if agent identity type EC2 can be assumed Jan 13 20:07:46.925553 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO Agent will take identity from EC2 Jan 13 20:07:47.027465 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [amazon-ssm-agent] using named pipe channel for IPC Jan 13 20:07:47.127263 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [amazon-ssm-agent] using named pipe channel for IPC Jan 13 20:07:47.229643 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [amazon-ssm-agent] using named pipe channel for IPC Jan 13 20:07:47.328576 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Jan 13 20:07:47.429400 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [amazon-ssm-agent] OS: linux, Arch: arm64 Jan 13 20:07:47.514499 tar[1968]: linux-arm64/LICENSE Jan 13 20:07:47.514499 tar[1968]: linux-arm64/README.md Jan 13 20:07:47.530252 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [amazon-ssm-agent] Starting Core Agent Jan 13 20:07:47.553699 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 13 20:07:47.631133 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [amazon-ssm-agent] registrar detected. Attempting registration Jan 13 20:07:47.730484 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [Registrar] Starting registrar module Jan 13 20:07:47.800389 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:07:47.820808 (kubelet)[2165]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 13 20:07:47.831299 amazon-ssm-agent[2100]: 2025-01-13 20:07:46 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Jan 13 20:07:48.245285 amazon-ssm-agent[2100]: 2025-01-13 20:07:48 INFO [EC2Identity] EC2 registration was successful. Jan 13 20:07:48.248304 ntpd[1937]: Listen normally on 6 eth0 [fe80::4db:2bff:fee0:d9ff%2]:123 Jan 13 20:07:48.249005 ntpd[1937]: 13 Jan 20:07:48 ntpd[1937]: Listen normally on 6 eth0 [fe80::4db:2bff:fee0:d9ff%2]:123 Jan 13 20:07:48.282801 amazon-ssm-agent[2100]: 2025-01-13 20:07:48 INFO [CredentialRefresher] credentialRefresher has started Jan 13 20:07:48.283004 amazon-ssm-agent[2100]: 2025-01-13 20:07:48 INFO [CredentialRefresher] Starting credentials refresher loop Jan 13 20:07:48.283761 amazon-ssm-agent[2100]: 2025-01-13 20:07:48 INFO EC2RoleProvider Successfully connected with instance profile role credentials Jan 13 20:07:48.346079 amazon-ssm-agent[2100]: 2025-01-13 20:07:48 INFO [CredentialRefresher] Next credential rotation will be in 30.033274302633334 minutes Jan 13 20:07:48.667225 kubelet[2165]: E0113 20:07:48.665751 2165 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 13 20:07:48.670378 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 13 20:07:48.670736 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 13 20:07:48.672290 systemd[1]: kubelet.service: Consumed 1.239s CPU time. Jan 13 20:07:49.310398 amazon-ssm-agent[2100]: 2025-01-13 20:07:49 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Jan 13 20:07:49.412941 amazon-ssm-agent[2100]: 2025-01-13 20:07:49 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2175) started Jan 13 20:07:49.513439 amazon-ssm-agent[2100]: 2025-01-13 20:07:49 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Jan 13 20:07:50.206085 sshd_keygen[1967]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 13 20:07:50.245200 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 13 20:07:50.257622 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 13 20:07:50.262648 systemd[1]: Started sshd@0-172.31.20.118:22-147.75.109.163:56828.service - OpenSSH per-connection server daemon (147.75.109.163:56828). Jan 13 20:07:50.282603 systemd[1]: issuegen.service: Deactivated successfully. Jan 13 20:07:50.283013 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 13 20:07:50.289697 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 13 20:07:50.321432 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 13 20:07:50.332936 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 13 20:07:50.346812 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 13 20:07:50.351731 systemd[1]: Reached target getty.target - Login Prompts. Jan 13 20:07:50.355907 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 13 20:07:50.363684 systemd[1]: Startup finished in 1.075s (kernel) + 9.810s (initrd) + 10.819s (userspace) = 21.705s. Jan 13 20:07:50.384624 agetty[2201]: failed to open credentials directory Jan 13 20:07:50.390891 agetty[2200]: failed to open credentials directory Jan 13 20:07:50.580670 sshd[2194]: Accepted publickey for core from 147.75.109.163 port 56828 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:07:50.584087 sshd-session[2194]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:07:50.602387 systemd-logind[1942]: New session 1 of user core. Jan 13 20:07:50.605508 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 13 20:07:50.614622 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 13 20:07:50.638995 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 13 20:07:50.648695 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 13 20:07:50.664368 (systemd)[2209]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 13 20:07:50.889768 systemd[2209]: Queued start job for default target default.target. Jan 13 20:07:50.901225 systemd[2209]: Created slice app.slice - User Application Slice. Jan 13 20:07:50.901291 systemd[2209]: Reached target paths.target - Paths. Jan 13 20:07:50.901324 systemd[2209]: Reached target timers.target - Timers. Jan 13 20:07:50.903842 systemd[2209]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 13 20:07:50.936087 systemd[2209]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 13 20:07:50.936352 systemd[2209]: Reached target sockets.target - Sockets. Jan 13 20:07:50.936386 systemd[2209]: Reached target basic.target - Basic System. Jan 13 20:07:50.936469 systemd[2209]: Reached target default.target - Main User Target. Jan 13 20:07:50.936532 systemd[2209]: Startup finished in 260ms. Jan 13 20:07:50.936671 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 13 20:07:50.943417 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 13 20:07:51.100685 systemd[1]: Started sshd@1-172.31.20.118:22-147.75.109.163:42974.service - OpenSSH per-connection server daemon (147.75.109.163:42974). Jan 13 20:07:51.286375 sshd[2220]: Accepted publickey for core from 147.75.109.163 port 42974 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:07:51.288927 sshd-session[2220]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:07:51.297729 systemd-logind[1942]: New session 2 of user core. Jan 13 20:07:51.309386 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 13 20:07:51.436487 sshd[2222]: Connection closed by 147.75.109.163 port 42974 Jan 13 20:07:51.437356 sshd-session[2220]: pam_unix(sshd:session): session closed for user core Jan 13 20:07:51.443353 systemd[1]: sshd@1-172.31.20.118:22-147.75.109.163:42974.service: Deactivated successfully. Jan 13 20:07:51.446817 systemd[1]: session-2.scope: Deactivated successfully. Jan 13 20:07:51.449164 systemd-logind[1942]: Session 2 logged out. Waiting for processes to exit. Jan 13 20:07:51.450976 systemd-logind[1942]: Removed session 2. Jan 13 20:07:51.477642 systemd[1]: Started sshd@2-172.31.20.118:22-147.75.109.163:42986.service - OpenSSH per-connection server daemon (147.75.109.163:42986). Jan 13 20:07:51.672477 sshd[2227]: Accepted publickey for core from 147.75.109.163 port 42986 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:07:51.674822 sshd-session[2227]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:07:51.681991 systemd-logind[1942]: New session 3 of user core. Jan 13 20:07:51.690373 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 13 20:07:51.808851 sshd[2229]: Connection closed by 147.75.109.163 port 42986 Jan 13 20:07:51.809654 sshd-session[2227]: pam_unix(sshd:session): session closed for user core Jan 13 20:07:51.815409 systemd[1]: sshd@2-172.31.20.118:22-147.75.109.163:42986.service: Deactivated successfully. Jan 13 20:07:51.818714 systemd[1]: session-3.scope: Deactivated successfully. Jan 13 20:07:51.820085 systemd-logind[1942]: Session 3 logged out. Waiting for processes to exit. Jan 13 20:07:51.822334 systemd-logind[1942]: Removed session 3. Jan 13 20:07:51.846651 systemd[1]: Started sshd@3-172.31.20.118:22-147.75.109.163:42994.service - OpenSSH per-connection server daemon (147.75.109.163:42994). Jan 13 20:07:52.028982 sshd[2234]: Accepted publickey for core from 147.75.109.163 port 42994 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:07:52.031401 sshd-session[2234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:07:52.040477 systemd-logind[1942]: New session 4 of user core. Jan 13 20:07:52.057392 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 13 20:07:52.182667 sshd[2236]: Connection closed by 147.75.109.163 port 42994 Jan 13 20:07:52.183200 sshd-session[2234]: pam_unix(sshd:session): session closed for user core Jan 13 20:07:52.189657 systemd-logind[1942]: Session 4 logged out. Waiting for processes to exit. Jan 13 20:07:52.191470 systemd[1]: sshd@3-172.31.20.118:22-147.75.109.163:42994.service: Deactivated successfully. Jan 13 20:07:52.194663 systemd[1]: session-4.scope: Deactivated successfully. Jan 13 20:07:52.196731 systemd-logind[1942]: Removed session 4. Jan 13 20:07:52.228847 systemd[1]: Started sshd@4-172.31.20.118:22-147.75.109.163:43004.service - OpenSSH per-connection server daemon (147.75.109.163:43004). Jan 13 20:07:52.394330 systemd-resolved[1843]: Clock change detected. Flushing caches. Jan 13 20:07:52.557479 sshd[2241]: Accepted publickey for core from 147.75.109.163 port 43004 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:07:52.559228 sshd-session[2241]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:07:52.566060 systemd-logind[1942]: New session 5 of user core. Jan 13 20:07:52.576678 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 13 20:07:52.739025 sudo[2244]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 13 20:07:52.739701 sudo[2244]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 13 20:07:53.489906 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 13 20:07:53.492257 (dockerd)[2261]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 13 20:07:53.964530 dockerd[2261]: time="2025-01-13T20:07:53.964402983Z" level=info msg="Starting up" Jan 13 20:07:54.782759 dockerd[2261]: time="2025-01-13T20:07:54.782669067Z" level=info msg="Loading containers: start." Jan 13 20:07:55.073488 kernel: Initializing XFRM netlink socket Jan 13 20:07:55.105086 (udev-worker)[2284]: Network interface NamePolicy= disabled on kernel command line. Jan 13 20:07:55.196980 systemd-networkd[1836]: docker0: Link UP Jan 13 20:07:55.240859 dockerd[2261]: time="2025-01-13T20:07:55.240765842Z" level=info msg="Loading containers: done." Jan 13 20:07:55.266615 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2770100790-merged.mount: Deactivated successfully. Jan 13 20:07:55.270627 dockerd[2261]: time="2025-01-13T20:07:55.270479678Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 13 20:07:55.270799 dockerd[2261]: time="2025-01-13T20:07:55.270670442Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Jan 13 20:07:55.270962 dockerd[2261]: time="2025-01-13T20:07:55.270899366Z" level=info msg="Daemon has completed initialization" Jan 13 20:07:55.328655 dockerd[2261]: time="2025-01-13T20:07:55.327877694Z" level=info msg="API listen on /run/docker.sock" Jan 13 20:07:55.328230 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 13 20:07:56.446469 containerd[1972]: time="2025-01-13T20:07:56.446333380Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.4\"" Jan 13 20:07:57.189475 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3584691634.mount: Deactivated successfully. Jan 13 20:07:58.767378 containerd[1972]: time="2025-01-13T20:07:58.767307535Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:07:58.770521 containerd[1972]: time="2025-01-13T20:07:58.770412811Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.4: active requests=0, bytes read=25615585" Jan 13 20:07:58.772735 containerd[1972]: time="2025-01-13T20:07:58.772656883Z" level=info msg="ImageCreate event name:\"sha256:3e1123d6ebadbafa6eb77a9047f23f20befbbe2f177eb473a81b27a5de8c2ec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:07:58.778113 containerd[1972]: time="2025-01-13T20:07:58.778012651Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:ace6a943b058439bd6daeb74f152e7c36e6fc0b5e481cdff9364cd6ca0473e5e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:07:58.780556 containerd[1972]: time="2025-01-13T20:07:58.780506899Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.4\" with image id \"sha256:3e1123d6ebadbafa6eb77a9047f23f20befbbe2f177eb473a81b27a5de8c2ec5\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:ace6a943b058439bd6daeb74f152e7c36e6fc0b5e481cdff9364cd6ca0473e5e\", size \"25612385\" in 2.334112247s" Jan 13 20:07:58.781279 containerd[1972]: time="2025-01-13T20:07:58.780704659Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.4\" returns image reference \"sha256:3e1123d6ebadbafa6eb77a9047f23f20befbbe2f177eb473a81b27a5de8c2ec5\"" Jan 13 20:07:58.782148 containerd[1972]: time="2025-01-13T20:07:58.781704295Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.4\"" Jan 13 20:07:58.820397 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 13 20:07:58.831760 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:07:59.324621 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:07:59.339108 (kubelet)[2512]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 13 20:07:59.411636 kubelet[2512]: E0113 20:07:59.411504 2512 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 13 20:07:59.418382 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 13 20:07:59.418776 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 13 20:08:00.348534 containerd[1972]: time="2025-01-13T20:08:00.348453859Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:00.350646 containerd[1972]: time="2025-01-13T20:08:00.350573599Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.4: active requests=0, bytes read=22470096" Jan 13 20:08:00.352710 containerd[1972]: time="2025-01-13T20:08:00.352632811Z" level=info msg="ImageCreate event name:\"sha256:d5369864a42bf2c01d3ad462832526b7d3e40620c0e75fecefbffc203562ad55\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:00.364939 containerd[1972]: time="2025-01-13T20:08:00.364137127Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:4bd1d4a449e7a1a4f375bd7c71abf48a95f8949b38f725ded255077329f21f7b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:00.366610 containerd[1972]: time="2025-01-13T20:08:00.365957287Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.4\" with image id \"sha256:d5369864a42bf2c01d3ad462832526b7d3e40620c0e75fecefbffc203562ad55\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:4bd1d4a449e7a1a4f375bd7c71abf48a95f8949b38f725ded255077329f21f7b\", size \"23872417\" in 1.584196136s" Jan 13 20:08:00.366610 containerd[1972]: time="2025-01-13T20:08:00.366057859Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.4\" returns image reference \"sha256:d5369864a42bf2c01d3ad462832526b7d3e40620c0e75fecefbffc203562ad55\"" Jan 13 20:08:00.367011 containerd[1972]: time="2025-01-13T20:08:00.366955531Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.4\"" Jan 13 20:08:01.745885 containerd[1972]: time="2025-01-13T20:08:01.745783546Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:01.759680 containerd[1972]: time="2025-01-13T20:08:01.759583474Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.4: active requests=0, bytes read=17024202" Jan 13 20:08:01.778332 containerd[1972]: time="2025-01-13T20:08:01.778228174Z" level=info msg="ImageCreate event name:\"sha256:d99fc9a32f6b42ab5537eec09d599efae0f61c109406dae1ba255cec288fcb95\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:01.799745 containerd[1972]: time="2025-01-13T20:08:01.799661614Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:1a3081cb7d21763d22eb2c0781cc462d89f501ed523ad558dea1226f128fbfdd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:01.801687 containerd[1972]: time="2025-01-13T20:08:01.801506974Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.4\" with image id \"sha256:d99fc9a32f6b42ab5537eec09d599efae0f61c109406dae1ba255cec288fcb95\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:1a3081cb7d21763d22eb2c0781cc462d89f501ed523ad558dea1226f128fbfdd\", size \"18426541\" in 1.434487495s" Jan 13 20:08:01.801687 containerd[1972]: time="2025-01-13T20:08:01.801560674Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.4\" returns image reference \"sha256:d99fc9a32f6b42ab5537eec09d599efae0f61c109406dae1ba255cec288fcb95\"" Jan 13 20:08:01.802760 containerd[1972]: time="2025-01-13T20:08:01.802696066Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.4\"" Jan 13 20:08:03.244338 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1420514091.mount: Deactivated successfully. Jan 13 20:08:03.815333 containerd[1972]: time="2025-01-13T20:08:03.815254020Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:03.816988 containerd[1972]: time="2025-01-13T20:08:03.816907536Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.4: active requests=0, bytes read=26771426" Jan 13 20:08:03.818847 containerd[1972]: time="2025-01-13T20:08:03.818772744Z" level=info msg="ImageCreate event name:\"sha256:34e142197cb996099cc1e98902c112642b3fb3dc559140c0a95279aa8d254d3a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:03.823349 containerd[1972]: time="2025-01-13T20:08:03.823253112Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:1739b3febca392035bf6edfe31efdfa55226be7b57389b2001ae357f7dcb99cf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:03.824792 containerd[1972]: time="2025-01-13T20:08:03.824549928Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.4\" with image id \"sha256:34e142197cb996099cc1e98902c112642b3fb3dc559140c0a95279aa8d254d3a\", repo tag \"registry.k8s.io/kube-proxy:v1.31.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:1739b3febca392035bf6edfe31efdfa55226be7b57389b2001ae357f7dcb99cf\", size \"26770445\" in 2.021616538s" Jan 13 20:08:03.824792 containerd[1972]: time="2025-01-13T20:08:03.824602944Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.4\" returns image reference \"sha256:34e142197cb996099cc1e98902c112642b3fb3dc559140c0a95279aa8d254d3a\"" Jan 13 20:08:03.825390 containerd[1972]: time="2025-01-13T20:08:03.825301164Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Jan 13 20:08:04.391310 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1391919513.mount: Deactivated successfully. Jan 13 20:08:05.558601 containerd[1972]: time="2025-01-13T20:08:05.558521173Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:05.560747 containerd[1972]: time="2025-01-13T20:08:05.560663317Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485381" Jan 13 20:08:05.562237 containerd[1972]: time="2025-01-13T20:08:05.562183177Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:05.568534 containerd[1972]: time="2025-01-13T20:08:05.568410541Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:05.571177 containerd[1972]: time="2025-01-13T20:08:05.570906841Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.745556225s" Jan 13 20:08:05.571177 containerd[1972]: time="2025-01-13T20:08:05.570971833Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Jan 13 20:08:05.571829 containerd[1972]: time="2025-01-13T20:08:05.571777741Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jan 13 20:08:06.101388 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2906117096.mount: Deactivated successfully. Jan 13 20:08:06.113526 containerd[1972]: time="2025-01-13T20:08:06.112800432Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:06.114758 containerd[1972]: time="2025-01-13T20:08:06.114689076Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Jan 13 20:08:06.116981 containerd[1972]: time="2025-01-13T20:08:06.116911812Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:06.123842 containerd[1972]: time="2025-01-13T20:08:06.123743712Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:06.125528 containerd[1972]: time="2025-01-13T20:08:06.125287620Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 553.454079ms" Jan 13 20:08:06.125528 containerd[1972]: time="2025-01-13T20:08:06.125343912Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Jan 13 20:08:06.127075 containerd[1972]: time="2025-01-13T20:08:06.127011852Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Jan 13 20:08:06.736701 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2778153577.mount: Deactivated successfully. Jan 13 20:08:09.571556 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 13 20:08:09.580027 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:08:10.292826 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:08:10.298854 (kubelet)[2648]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 13 20:08:10.312475 containerd[1972]: time="2025-01-13T20:08:10.311060740Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:10.314873 containerd[1972]: time="2025-01-13T20:08:10.314761685Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66406425" Jan 13 20:08:10.318478 containerd[1972]: time="2025-01-13T20:08:10.317704325Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:10.325316 containerd[1972]: time="2025-01-13T20:08:10.325257401Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:10.331041 containerd[1972]: time="2025-01-13T20:08:10.330934121Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 4.203859173s" Jan 13 20:08:10.331250 containerd[1972]: time="2025-01-13T20:08:10.331216193Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Jan 13 20:08:10.393422 kubelet[2648]: E0113 20:08:10.393332 2648 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 13 20:08:10.401615 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 13 20:08:10.401940 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 13 20:08:16.226571 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Jan 13 20:08:17.092411 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:08:17.103937 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:08:17.162201 systemd[1]: Reloading requested from client PID 2684 ('systemctl') (unit session-5.scope)... Jan 13 20:08:17.162397 systemd[1]: Reloading... Jan 13 20:08:17.375491 zram_generator::config[2727]: No configuration found. Jan 13 20:08:17.617127 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 13 20:08:17.793389 systemd[1]: Reloading finished in 630 ms. Jan 13 20:08:17.897636 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 13 20:08:17.897880 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 13 20:08:17.898524 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:08:17.909568 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:08:18.268754 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:08:18.286003 (kubelet)[2788]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 13 20:08:18.354468 kubelet[2788]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 13 20:08:18.354468 kubelet[2788]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 13 20:08:18.354468 kubelet[2788]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 13 20:08:18.355060 kubelet[2788]: I0113 20:08:18.354643 2788 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 13 20:08:19.725319 kubelet[2788]: I0113 20:08:19.725260 2788 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Jan 13 20:08:19.725319 kubelet[2788]: I0113 20:08:19.725314 2788 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 13 20:08:19.726253 kubelet[2788]: I0113 20:08:19.725797 2788 server.go:929] "Client rotation is on, will bootstrap in background" Jan 13 20:08:19.765242 kubelet[2788]: E0113 20:08:19.765177 2788 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.20.118:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:19.767264 kubelet[2788]: I0113 20:08:19.767024 2788 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 13 20:08:19.778250 kubelet[2788]: E0113 20:08:19.778183 2788 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 13 20:08:19.778250 kubelet[2788]: I0113 20:08:19.778238 2788 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 13 20:08:19.787569 kubelet[2788]: I0113 20:08:19.787165 2788 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 13 20:08:19.787569 kubelet[2788]: I0113 20:08:19.787584 2788 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 13 20:08:19.787894 kubelet[2788]: I0113 20:08:19.787839 2788 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 13 20:08:19.788687 kubelet[2788]: I0113 20:08:19.787892 2788 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-20-118","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 13 20:08:19.788894 kubelet[2788]: I0113 20:08:19.788743 2788 topology_manager.go:138] "Creating topology manager with none policy" Jan 13 20:08:19.788894 kubelet[2788]: I0113 20:08:19.788775 2788 container_manager_linux.go:300] "Creating device plugin manager" Jan 13 20:08:19.789000 kubelet[2788]: I0113 20:08:19.788967 2788 state_mem.go:36] "Initialized new in-memory state store" Jan 13 20:08:19.795777 kubelet[2788]: I0113 20:08:19.795717 2788 kubelet.go:408] "Attempting to sync node with API server" Jan 13 20:08:19.795777 kubelet[2788]: I0113 20:08:19.795774 2788 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 13 20:08:19.796567 kubelet[2788]: I0113 20:08:19.796524 2788 kubelet.go:314] "Adding apiserver pod source" Jan 13 20:08:19.796567 kubelet[2788]: I0113 20:08:19.796564 2788 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 13 20:08:19.804144 kubelet[2788]: I0113 20:08:19.803874 2788 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Jan 13 20:08:19.806798 kubelet[2788]: I0113 20:08:19.806749 2788 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 13 20:08:19.808182 kubelet[2788]: W0113 20:08:19.808133 2788 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 13 20:08:19.809259 kubelet[2788]: I0113 20:08:19.809204 2788 server.go:1269] "Started kubelet" Jan 13 20:08:19.809609 kubelet[2788]: W0113 20:08:19.809531 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.20.118:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-118&limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:19.809684 kubelet[2788]: E0113 20:08:19.809636 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.20.118:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-118&limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:19.812468 kubelet[2788]: W0113 20:08:19.812342 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.20.118:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:19.812468 kubelet[2788]: E0113 20:08:19.812456 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.20.118:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:19.812716 kubelet[2788]: I0113 20:08:19.812548 2788 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 13 20:08:19.817675 kubelet[2788]: I0113 20:08:19.816242 2788 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 13 20:08:19.817675 kubelet[2788]: I0113 20:08:19.816841 2788 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 13 20:08:19.817675 kubelet[2788]: I0113 20:08:19.817228 2788 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 13 20:08:19.824557 kubelet[2788]: E0113 20:08:19.820718 2788 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.20.118:6443/api/v1/namespaces/default/events\": dial tcp 172.31.20.118:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-20-118.181a5965bd6a2d94 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-20-118,UID:ip-172-31-20-118,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-20-118,},FirstTimestamp:2025-01-13 20:08:19.809168788 +0000 UTC m=+1.517360169,LastTimestamp:2025-01-13 20:08:19.809168788 +0000 UTC m=+1.517360169,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-20-118,}" Jan 13 20:08:19.826880 kubelet[2788]: I0113 20:08:19.826846 2788 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 13 20:08:19.827468 kubelet[2788]: I0113 20:08:19.827103 2788 server.go:460] "Adding debug handlers to kubelet server" Jan 13 20:08:19.827630 kubelet[2788]: E0113 20:08:19.827601 2788 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-20-118\" not found" Jan 13 20:08:19.828791 kubelet[2788]: I0113 20:08:19.828717 2788 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 13 20:08:19.833803 kubelet[2788]: I0113 20:08:19.833371 2788 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 13 20:08:19.833803 kubelet[2788]: I0113 20:08:19.833548 2788 reconciler.go:26] "Reconciler: start to sync state" Jan 13 20:08:19.834316 kubelet[2788]: E0113 20:08:19.834244 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.118:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-118?timeout=10s\": dial tcp 172.31.20.118:6443: connect: connection refused" interval="200ms" Jan 13 20:08:19.834651 kubelet[2788]: I0113 20:08:19.834608 2788 factory.go:221] Registration of the systemd container factory successfully Jan 13 20:08:19.835527 kubelet[2788]: I0113 20:08:19.834768 2788 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 13 20:08:19.838333 kubelet[2788]: I0113 20:08:19.838281 2788 factory.go:221] Registration of the containerd container factory successfully Jan 13 20:08:19.853563 kubelet[2788]: I0113 20:08:19.853506 2788 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 13 20:08:19.855851 kubelet[2788]: I0113 20:08:19.855810 2788 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 13 20:08:19.856043 kubelet[2788]: I0113 20:08:19.856022 2788 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 13 20:08:19.856176 kubelet[2788]: I0113 20:08:19.856158 2788 kubelet.go:2321] "Starting kubelet main sync loop" Jan 13 20:08:19.856334 kubelet[2788]: E0113 20:08:19.856305 2788 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 13 20:08:19.869015 kubelet[2788]: W0113 20:08:19.868938 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.20.118:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:19.869215 kubelet[2788]: E0113 20:08:19.869184 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.20.118:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:19.870037 kubelet[2788]: W0113 20:08:19.869719 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.20.118:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:19.870037 kubelet[2788]: E0113 20:08:19.869790 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.20.118:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:19.878393 kubelet[2788]: E0113 20:08:19.877861 2788 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 13 20:08:19.889017 kubelet[2788]: I0113 20:08:19.888976 2788 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 13 20:08:19.889017 kubelet[2788]: I0113 20:08:19.889007 2788 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 13 20:08:19.889212 kubelet[2788]: I0113 20:08:19.889043 2788 state_mem.go:36] "Initialized new in-memory state store" Jan 13 20:08:19.928689 kubelet[2788]: E0113 20:08:19.928647 2788 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-20-118\" not found" Jan 13 20:08:19.957273 kubelet[2788]: E0113 20:08:19.957216 2788 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 13 20:08:20.019500 kubelet[2788]: I0113 20:08:20.019171 2788 policy_none.go:49] "None policy: Start" Jan 13 20:08:20.022200 kubelet[2788]: I0113 20:08:20.022150 2788 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 13 20:08:20.022327 kubelet[2788]: I0113 20:08:20.022223 2788 state_mem.go:35] "Initializing new in-memory state store" Jan 13 20:08:20.029114 kubelet[2788]: E0113 20:08:20.029060 2788 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ip-172-31-20-118\" not found" Jan 13 20:08:20.035854 kubelet[2788]: E0113 20:08:20.035787 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.118:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-118?timeout=10s\": dial tcp 172.31.20.118:6443: connect: connection refused" interval="400ms" Jan 13 20:08:20.077320 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 13 20:08:20.090783 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 13 20:08:20.098308 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 13 20:08:20.109822 kubelet[2788]: I0113 20:08:20.109377 2788 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 13 20:08:20.109822 kubelet[2788]: I0113 20:08:20.109695 2788 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 13 20:08:20.109822 kubelet[2788]: I0113 20:08:20.109715 2788 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 13 20:08:20.121934 kubelet[2788]: I0113 20:08:20.121131 2788 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 13 20:08:20.124810 kubelet[2788]: E0113 20:08:20.124764 2788 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-20-118\" not found" Jan 13 20:08:20.182863 systemd[1]: Created slice kubepods-burstable-pod0d3164495219bb25c05182931c5420ab.slice - libcontainer container kubepods-burstable-pod0d3164495219bb25c05182931c5420ab.slice. Jan 13 20:08:20.195515 systemd[1]: Created slice kubepods-burstable-poded98ea7da991063c7fb4b23f131aa385.slice - libcontainer container kubepods-burstable-poded98ea7da991063c7fb4b23f131aa385.slice. Jan 13 20:08:20.212188 kubelet[2788]: I0113 20:08:20.212107 2788 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-20-118" Jan 13 20:08:20.212709 systemd[1]: Created slice kubepods-burstable-pod0521c3369b30f001a16ecc0cb2c8c54d.slice - libcontainer container kubepods-burstable-pod0521c3369b30f001a16ecc0cb2c8c54d.slice. Jan 13 20:08:20.214736 kubelet[2788]: E0113 20:08:20.213685 2788 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.20.118:6443/api/v1/nodes\": dial tcp 172.31.20.118:6443: connect: connection refused" node="ip-172-31-20-118" Jan 13 20:08:20.236003 kubelet[2788]: I0113 20:08:20.235950 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0d3164495219bb25c05182931c5420ab-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-118\" (UID: \"0d3164495219bb25c05182931c5420ab\") " pod="kube-system/kube-apiserver-ip-172-31-20-118" Jan 13 20:08:20.236136 kubelet[2788]: I0113 20:08:20.236015 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:20.236136 kubelet[2788]: I0113 20:08:20.236057 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ed98ea7da991063c7fb4b23f131aa385-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-118\" (UID: \"ed98ea7da991063c7fb4b23f131aa385\") " pod="kube-system/kube-scheduler-ip-172-31-20-118" Jan 13 20:08:20.236136 kubelet[2788]: I0113 20:08:20.236093 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0d3164495219bb25c05182931c5420ab-ca-certs\") pod \"kube-apiserver-ip-172-31-20-118\" (UID: \"0d3164495219bb25c05182931c5420ab\") " pod="kube-system/kube-apiserver-ip-172-31-20-118" Jan 13 20:08:20.236136 kubelet[2788]: I0113 20:08:20.236129 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0d3164495219bb25c05182931c5420ab-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-118\" (UID: \"0d3164495219bb25c05182931c5420ab\") " pod="kube-system/kube-apiserver-ip-172-31-20-118" Jan 13 20:08:20.236580 kubelet[2788]: I0113 20:08:20.236169 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:20.236580 kubelet[2788]: I0113 20:08:20.236203 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:20.236580 kubelet[2788]: I0113 20:08:20.236241 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:20.236580 kubelet[2788]: I0113 20:08:20.236278 2788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:20.417260 kubelet[2788]: I0113 20:08:20.417023 2788 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-20-118" Jan 13 20:08:20.417625 kubelet[2788]: E0113 20:08:20.417564 2788 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.20.118:6443/api/v1/nodes\": dial tcp 172.31.20.118:6443: connect: connection refused" node="ip-172-31-20-118" Jan 13 20:08:20.437604 kubelet[2788]: E0113 20:08:20.437511 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.118:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-118?timeout=10s\": dial tcp 172.31.20.118:6443: connect: connection refused" interval="800ms" Jan 13 20:08:20.491914 containerd[1972]: time="2025-01-13T20:08:20.491811027Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-118,Uid:0d3164495219bb25c05182931c5420ab,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:20.507872 containerd[1972]: time="2025-01-13T20:08:20.507403167Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-118,Uid:ed98ea7da991063c7fb4b23f131aa385,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:20.518276 containerd[1972]: time="2025-01-13T20:08:20.518222019Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-118,Uid:0521c3369b30f001a16ecc0cb2c8c54d,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:20.805711 kubelet[2788]: W0113 20:08:20.805415 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.20.118:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:20.805711 kubelet[2788]: E0113 20:08:20.805534 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.20.118:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:20.819924 kubelet[2788]: I0113 20:08:20.819865 2788 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-20-118" Jan 13 20:08:20.820415 kubelet[2788]: E0113 20:08:20.820351 2788 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.20.118:6443/api/v1/nodes\": dial tcp 172.31.20.118:6443: connect: connection refused" node="ip-172-31-20-118" Jan 13 20:08:20.886221 kubelet[2788]: W0113 20:08:20.886059 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.20.118:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:20.886221 kubelet[2788]: E0113 20:08:20.886155 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.20.118:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:21.064963 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2421465359.mount: Deactivated successfully. Jan 13 20:08:21.078269 containerd[1972]: time="2025-01-13T20:08:21.078153686Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:08:21.081538 containerd[1972]: time="2025-01-13T20:08:21.081480734Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:08:21.085745 containerd[1972]: time="2025-01-13T20:08:21.085664138Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Jan 13 20:08:21.087139 containerd[1972]: time="2025-01-13T20:08:21.087067970Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 13 20:08:21.090636 containerd[1972]: time="2025-01-13T20:08:21.090572798Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:08:21.092878 containerd[1972]: time="2025-01-13T20:08:21.092781290Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:08:21.093582 containerd[1972]: time="2025-01-13T20:08:21.093388934Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 13 20:08:21.098497 containerd[1972]: time="2025-01-13T20:08:21.098400902Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 13 20:08:21.102703 containerd[1972]: time="2025-01-13T20:08:21.102652310Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 584.323023ms" Jan 13 20:08:21.107336 containerd[1972]: time="2025-01-13T20:08:21.107094374Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 615.169911ms" Jan 13 20:08:21.119216 containerd[1972]: time="2025-01-13T20:08:21.119148266Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 611.588271ms" Jan 13 20:08:21.239145 kubelet[2788]: E0113 20:08:21.239079 2788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.118:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-118?timeout=10s\": dial tcp 172.31.20.118:6443: connect: connection refused" interval="1.6s" Jan 13 20:08:21.288475 kubelet[2788]: W0113 20:08:21.288361 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.20.118:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:21.288637 kubelet[2788]: E0113 20:08:21.288502 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.20.118:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:21.402365 kubelet[2788]: W0113 20:08:21.402270 2788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.20.118:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-118&limit=500&resourceVersion=0": dial tcp 172.31.20.118:6443: connect: connection refused Jan 13 20:08:21.402541 kubelet[2788]: E0113 20:08:21.402374 2788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.20.118:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-118&limit=500&resourceVersion=0\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:21.459992 containerd[1972]: time="2025-01-13T20:08:21.459544408Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:08:21.459992 containerd[1972]: time="2025-01-13T20:08:21.459668704Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:08:21.459992 containerd[1972]: time="2025-01-13T20:08:21.459734620Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:21.459992 containerd[1972]: time="2025-01-13T20:08:21.459897808Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:21.462463 containerd[1972]: time="2025-01-13T20:08:21.462098416Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:08:21.462463 containerd[1972]: time="2025-01-13T20:08:21.462215716Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:08:21.462463 containerd[1972]: time="2025-01-13T20:08:21.462256348Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:21.464630 containerd[1972]: time="2025-01-13T20:08:21.462971920Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:21.469395 containerd[1972]: time="2025-01-13T20:08:21.469212868Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:08:21.470158 containerd[1972]: time="2025-01-13T20:08:21.469989412Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:08:21.470158 containerd[1972]: time="2025-01-13T20:08:21.470111392Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:21.470407 containerd[1972]: time="2025-01-13T20:08:21.470316868Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:21.511694 systemd[1]: Started cri-containerd-2dc0640630805fff70f7416dbcd18199d313f521d58b929e5844dfd5cafef48e.scope - libcontainer container 2dc0640630805fff70f7416dbcd18199d313f521d58b929e5844dfd5cafef48e. Jan 13 20:08:21.529791 systemd[1]: Started cri-containerd-3f4e6113ccf37dd6f982bacab2a1b9f4174599f5ea18a0b581129b29bcb2e59c.scope - libcontainer container 3f4e6113ccf37dd6f982bacab2a1b9f4174599f5ea18a0b581129b29bcb2e59c. Jan 13 20:08:21.534044 systemd[1]: Started cri-containerd-cf5ce157ce50049b9cc6feef6f92074be3eb00eea3888cd80fbbe130c3e77a28.scope - libcontainer container cf5ce157ce50049b9cc6feef6f92074be3eb00eea3888cd80fbbe130c3e77a28. Jan 13 20:08:21.624700 kubelet[2788]: I0113 20:08:21.624626 2788 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-20-118" Jan 13 20:08:21.626660 kubelet[2788]: E0113 20:08:21.626582 2788 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://172.31.20.118:6443/api/v1/nodes\": dial tcp 172.31.20.118:6443: connect: connection refused" node="ip-172-31-20-118" Jan 13 20:08:21.636880 containerd[1972]: time="2025-01-13T20:08:21.636673829Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-118,Uid:0521c3369b30f001a16ecc0cb2c8c54d,Namespace:kube-system,Attempt:0,} returns sandbox id \"2dc0640630805fff70f7416dbcd18199d313f521d58b929e5844dfd5cafef48e\"" Jan 13 20:08:21.642116 containerd[1972]: time="2025-01-13T20:08:21.641720909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-118,Uid:0d3164495219bb25c05182931c5420ab,Namespace:kube-system,Attempt:0,} returns sandbox id \"3f4e6113ccf37dd6f982bacab2a1b9f4174599f5ea18a0b581129b29bcb2e59c\"" Jan 13 20:08:21.656656 containerd[1972]: time="2025-01-13T20:08:21.656221025Z" level=info msg="CreateContainer within sandbox \"2dc0640630805fff70f7416dbcd18199d313f521d58b929e5844dfd5cafef48e\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 13 20:08:21.658101 containerd[1972]: time="2025-01-13T20:08:21.657761765Z" level=info msg="CreateContainer within sandbox \"3f4e6113ccf37dd6f982bacab2a1b9f4174599f5ea18a0b581129b29bcb2e59c\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 13 20:08:21.676953 containerd[1972]: time="2025-01-13T20:08:21.676873901Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-118,Uid:ed98ea7da991063c7fb4b23f131aa385,Namespace:kube-system,Attempt:0,} returns sandbox id \"cf5ce157ce50049b9cc6feef6f92074be3eb00eea3888cd80fbbe130c3e77a28\"" Jan 13 20:08:21.686858 containerd[1972]: time="2025-01-13T20:08:21.686403449Z" level=info msg="CreateContainer within sandbox \"cf5ce157ce50049b9cc6feef6f92074be3eb00eea3888cd80fbbe130c3e77a28\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 13 20:08:21.702686 containerd[1972]: time="2025-01-13T20:08:21.702633221Z" level=info msg="CreateContainer within sandbox \"2dc0640630805fff70f7416dbcd18199d313f521d58b929e5844dfd5cafef48e\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc\"" Jan 13 20:08:21.703659 containerd[1972]: time="2025-01-13T20:08:21.703609025Z" level=info msg="StartContainer for \"5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc\"" Jan 13 20:08:21.706923 containerd[1972]: time="2025-01-13T20:08:21.706659965Z" level=info msg="CreateContainer within sandbox \"3f4e6113ccf37dd6f982bacab2a1b9f4174599f5ea18a0b581129b29bcb2e59c\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3addb30a9dd7591b896258e072724224dbe2cf6736cbac74d6a2c3d5662ecd03\"" Jan 13 20:08:21.709473 containerd[1972]: time="2025-01-13T20:08:21.707984657Z" level=info msg="StartContainer for \"3addb30a9dd7591b896258e072724224dbe2cf6736cbac74d6a2c3d5662ecd03\"" Jan 13 20:08:21.722912 containerd[1972]: time="2025-01-13T20:08:21.722838953Z" level=info msg="CreateContainer within sandbox \"cf5ce157ce50049b9cc6feef6f92074be3eb00eea3888cd80fbbe130c3e77a28\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5\"" Jan 13 20:08:21.723678 containerd[1972]: time="2025-01-13T20:08:21.723625109Z" level=info msg="StartContainer for \"79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5\"" Jan 13 20:08:21.766793 systemd[1]: Started cri-containerd-3addb30a9dd7591b896258e072724224dbe2cf6736cbac74d6a2c3d5662ecd03.scope - libcontainer container 3addb30a9dd7591b896258e072724224dbe2cf6736cbac74d6a2c3d5662ecd03. Jan 13 20:08:21.791766 systemd[1]: Started cri-containerd-5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc.scope - libcontainer container 5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc. Jan 13 20:08:21.807041 kubelet[2788]: E0113 20:08:21.806933 2788 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.20.118:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.20.118:6443: connect: connection refused" logger="UnhandledError" Jan 13 20:08:21.834723 systemd[1]: Started cri-containerd-79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5.scope - libcontainer container 79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5. Jan 13 20:08:21.911796 containerd[1972]: time="2025-01-13T20:08:21.910156374Z" level=info msg="StartContainer for \"3addb30a9dd7591b896258e072724224dbe2cf6736cbac74d6a2c3d5662ecd03\" returns successfully" Jan 13 20:08:21.943307 containerd[1972]: time="2025-01-13T20:08:21.942794274Z" level=info msg="StartContainer for \"5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc\" returns successfully" Jan 13 20:08:22.028208 containerd[1972]: time="2025-01-13T20:08:22.026851035Z" level=info msg="StartContainer for \"79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5\" returns successfully" Jan 13 20:08:23.231166 kubelet[2788]: I0113 20:08:23.230228 2788 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-20-118" Jan 13 20:08:25.642757 kubelet[2788]: E0113 20:08:25.642662 2788 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-20-118\" not found" node="ip-172-31-20-118" Jan 13 20:08:25.732270 kubelet[2788]: I0113 20:08:25.731821 2788 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-20-118" Jan 13 20:08:25.814589 kubelet[2788]: I0113 20:08:25.814490 2788 apiserver.go:52] "Watching apiserver" Jan 13 20:08:25.834395 kubelet[2788]: I0113 20:08:25.834305 2788 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 13 20:08:27.767233 systemd[1]: Reloading requested from client PID 3064 ('systemctl') (unit session-5.scope)... Jan 13 20:08:27.767265 systemd[1]: Reloading... Jan 13 20:08:28.010477 zram_generator::config[3113]: No configuration found. Jan 13 20:08:28.251817 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 13 20:08:28.455967 systemd[1]: Reloading finished in 688 ms. Jan 13 20:08:28.544251 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:08:28.563979 systemd[1]: kubelet.service: Deactivated successfully. Jan 13 20:08:28.564334 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:08:28.564398 systemd[1]: kubelet.service: Consumed 2.166s CPU time, 114.9M memory peak, 0B memory swap peak. Jan 13 20:08:28.574059 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 13 20:08:28.941712 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 13 20:08:28.951980 (kubelet)[3164]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 13 20:08:29.059480 kubelet[3164]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 13 20:08:29.059480 kubelet[3164]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 13 20:08:29.059480 kubelet[3164]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 13 20:08:29.060032 kubelet[3164]: I0113 20:08:29.059632 3164 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 13 20:08:29.077521 kubelet[3164]: I0113 20:08:29.076736 3164 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Jan 13 20:08:29.077521 kubelet[3164]: I0113 20:08:29.076778 3164 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 13 20:08:29.078032 kubelet[3164]: I0113 20:08:29.078000 3164 server.go:929] "Client rotation is on, will bootstrap in background" Jan 13 20:08:29.082135 kubelet[3164]: I0113 20:08:29.082095 3164 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 13 20:08:29.086321 kubelet[3164]: I0113 20:08:29.086260 3164 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 13 20:08:29.094098 kubelet[3164]: E0113 20:08:29.093917 3164 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 13 20:08:29.094606 kubelet[3164]: I0113 20:08:29.094570 3164 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 13 20:08:29.104145 kubelet[3164]: I0113 20:08:29.104079 3164 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 13 20:08:29.105732 kubelet[3164]: I0113 20:08:29.105682 3164 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 13 20:08:29.106006 kubelet[3164]: I0113 20:08:29.105941 3164 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 13 20:08:29.106326 kubelet[3164]: I0113 20:08:29.106000 3164 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-20-118","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 13 20:08:29.106554 kubelet[3164]: I0113 20:08:29.106336 3164 topology_manager.go:138] "Creating topology manager with none policy" Jan 13 20:08:29.106554 kubelet[3164]: I0113 20:08:29.106357 3164 container_manager_linux.go:300] "Creating device plugin manager" Jan 13 20:08:29.106554 kubelet[3164]: I0113 20:08:29.106412 3164 state_mem.go:36] "Initialized new in-memory state store" Jan 13 20:08:29.107847 kubelet[3164]: I0113 20:08:29.107800 3164 kubelet.go:408] "Attempting to sync node with API server" Jan 13 20:08:29.110065 kubelet[3164]: I0113 20:08:29.109498 3164 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 13 20:08:29.110065 kubelet[3164]: I0113 20:08:29.109565 3164 kubelet.go:314] "Adding apiserver pod source" Jan 13 20:08:29.110065 kubelet[3164]: I0113 20:08:29.109587 3164 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 13 20:08:29.113448 kubelet[3164]: I0113 20:08:29.112332 3164 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Jan 13 20:08:29.113448 kubelet[3164]: I0113 20:08:29.113242 3164 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 13 20:08:29.147248 kubelet[3164]: I0113 20:08:29.146169 3164 server.go:1269] "Started kubelet" Jan 13 20:08:29.151446 kubelet[3164]: I0113 20:08:29.150069 3164 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 13 20:08:29.153628 kubelet[3164]: I0113 20:08:29.153192 3164 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 13 20:08:29.155057 kubelet[3164]: I0113 20:08:29.154384 3164 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 13 20:08:29.163328 kubelet[3164]: I0113 20:08:29.162742 3164 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 13 20:08:29.168469 kubelet[3164]: I0113 20:08:29.166928 3164 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 13 20:08:29.168469 kubelet[3164]: I0113 20:08:29.167375 3164 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 13 20:08:29.170179 kubelet[3164]: I0113 20:08:29.170122 3164 server.go:460] "Adding debug handlers to kubelet server" Jan 13 20:08:29.172971 kubelet[3164]: I0113 20:08:29.172910 3164 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 13 20:08:29.190194 kubelet[3164]: I0113 20:08:29.190147 3164 reconciler.go:26] "Reconciler: start to sync state" Jan 13 20:08:29.194943 kubelet[3164]: E0113 20:08:29.194787 3164 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 13 20:08:29.200250 kubelet[3164]: I0113 20:08:29.198935 3164 factory.go:221] Registration of the containerd container factory successfully Jan 13 20:08:29.200250 kubelet[3164]: I0113 20:08:29.198976 3164 factory.go:221] Registration of the systemd container factory successfully Jan 13 20:08:29.200250 kubelet[3164]: I0113 20:08:29.199132 3164 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 13 20:08:29.227478 kubelet[3164]: I0113 20:08:29.226808 3164 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 13 20:08:29.231037 kubelet[3164]: I0113 20:08:29.230979 3164 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 13 20:08:29.231037 kubelet[3164]: I0113 20:08:29.231025 3164 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 13 20:08:29.231251 kubelet[3164]: I0113 20:08:29.231057 3164 kubelet.go:2321] "Starting kubelet main sync loop" Jan 13 20:08:29.231251 kubelet[3164]: E0113 20:08:29.231134 3164 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 13 20:08:29.330563 kubelet[3164]: I0113 20:08:29.330192 3164 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 13 20:08:29.330563 kubelet[3164]: I0113 20:08:29.330226 3164 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 13 20:08:29.330563 kubelet[3164]: I0113 20:08:29.330262 3164 state_mem.go:36] "Initialized new in-memory state store" Jan 13 20:08:29.330563 kubelet[3164]: I0113 20:08:29.330553 3164 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 13 20:08:29.330975 kubelet[3164]: I0113 20:08:29.330575 3164 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 13 20:08:29.330975 kubelet[3164]: I0113 20:08:29.330609 3164 policy_none.go:49] "None policy: Start" Jan 13 20:08:29.331627 kubelet[3164]: E0113 20:08:29.331314 3164 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 13 20:08:29.333028 kubelet[3164]: I0113 20:08:29.332478 3164 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 13 20:08:29.333028 kubelet[3164]: I0113 20:08:29.332519 3164 state_mem.go:35] "Initializing new in-memory state store" Jan 13 20:08:29.333028 kubelet[3164]: I0113 20:08:29.332821 3164 state_mem.go:75] "Updated machine memory state" Jan 13 20:08:29.344932 kubelet[3164]: I0113 20:08:29.343373 3164 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 13 20:08:29.345515 kubelet[3164]: I0113 20:08:29.345488 3164 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 13 20:08:29.346914 kubelet[3164]: I0113 20:08:29.345985 3164 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 13 20:08:29.346914 kubelet[3164]: I0113 20:08:29.346754 3164 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 13 20:08:29.467805 kubelet[3164]: I0113 20:08:29.467638 3164 kubelet_node_status.go:72] "Attempting to register node" node="ip-172-31-20-118" Jan 13 20:08:29.484920 kubelet[3164]: I0113 20:08:29.483602 3164 kubelet_node_status.go:111] "Node was previously registered" node="ip-172-31-20-118" Jan 13 20:08:29.484920 kubelet[3164]: I0113 20:08:29.483728 3164 kubelet_node_status.go:75] "Successfully registered node" node="ip-172-31-20-118" Jan 13 20:08:29.542653 kubelet[3164]: E0113 20:08:29.542535 3164 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ip-172-31-20-118\" already exists" pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:29.603703 kubelet[3164]: I0113 20:08:29.603633 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0d3164495219bb25c05182931c5420ab-ca-certs\") pod \"kube-apiserver-ip-172-31-20-118\" (UID: \"0d3164495219bb25c05182931c5420ab\") " pod="kube-system/kube-apiserver-ip-172-31-20-118" Jan 13 20:08:29.603837 kubelet[3164]: I0113 20:08:29.603709 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:29.603837 kubelet[3164]: I0113 20:08:29.603755 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:29.603837 kubelet[3164]: I0113 20:08:29.603798 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ed98ea7da991063c7fb4b23f131aa385-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-118\" (UID: \"ed98ea7da991063c7fb4b23f131aa385\") " pod="kube-system/kube-scheduler-ip-172-31-20-118" Jan 13 20:08:29.604003 kubelet[3164]: I0113 20:08:29.603840 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0d3164495219bb25c05182931c5420ab-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-118\" (UID: \"0d3164495219bb25c05182931c5420ab\") " pod="kube-system/kube-apiserver-ip-172-31-20-118" Jan 13 20:08:29.604003 kubelet[3164]: I0113 20:08:29.603875 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:29.604003 kubelet[3164]: I0113 20:08:29.603928 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:29.604003 kubelet[3164]: I0113 20:08:29.603964 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0521c3369b30f001a16ecc0cb2c8c54d-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-118\" (UID: \"0521c3369b30f001a16ecc0cb2c8c54d\") " pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:29.604003 kubelet[3164]: I0113 20:08:29.604003 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0d3164495219bb25c05182931c5420ab-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-118\" (UID: \"0d3164495219bb25c05182931c5420ab\") " pod="kube-system/kube-apiserver-ip-172-31-20-118" Jan 13 20:08:30.121112 kubelet[3164]: I0113 20:08:30.121051 3164 apiserver.go:52] "Watching apiserver" Jan 13 20:08:30.168382 kubelet[3164]: I0113 20:08:30.168331 3164 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 13 20:08:30.290506 kubelet[3164]: E0113 20:08:30.290459 3164 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ip-172-31-20-118\" already exists" pod="kube-system/kube-controller-manager-ip-172-31-20-118" Jan 13 20:08:30.294041 kubelet[3164]: E0113 20:08:30.293738 3164 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ip-172-31-20-118\" already exists" pod="kube-system/kube-apiserver-ip-172-31-20-118" Jan 13 20:08:30.440561 kubelet[3164]: I0113 20:08:30.440065 3164 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-20-118" podStartSLOduration=1.4400410080000001 podStartE2EDuration="1.440041008s" podCreationTimestamp="2025-01-13 20:08:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-13 20:08:30.362994384 +0000 UTC m=+1.403385404" watchObservedRunningTime="2025-01-13 20:08:30.440041008 +0000 UTC m=+1.480432016" Jan 13 20:08:30.494786 kubelet[3164]: I0113 20:08:30.494665 3164 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-20-118" podStartSLOduration=2.494642185 podStartE2EDuration="2.494642185s" podCreationTimestamp="2025-01-13 20:08:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-13 20:08:30.444779341 +0000 UTC m=+1.485170373" watchObservedRunningTime="2025-01-13 20:08:30.494642185 +0000 UTC m=+1.535033181" Jan 13 20:08:30.496743 kubelet[3164]: I0113 20:08:30.496620 3164 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-20-118" podStartSLOduration=1.496595413 podStartE2EDuration="1.496595413s" podCreationTimestamp="2025-01-13 20:08:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-13 20:08:30.493510033 +0000 UTC m=+1.533901101" watchObservedRunningTime="2025-01-13 20:08:30.496595413 +0000 UTC m=+1.536986661" Jan 13 20:08:30.805671 update_engine[1944]: I20250113 20:08:30.805490 1944 update_attempter.cc:509] Updating boot flags... Jan 13 20:08:30.829687 sudo[2244]: pam_unix(sudo:session): session closed for user root Jan 13 20:08:30.855491 sshd[2243]: Connection closed by 147.75.109.163 port 43004 Jan 13 20:08:30.855726 sshd-session[2241]: pam_unix(sshd:session): session closed for user core Jan 13 20:08:30.862337 systemd[1]: sshd@4-172.31.20.118:22-147.75.109.163:43004.service: Deactivated successfully. Jan 13 20:08:30.867907 systemd[1]: session-5.scope: Deactivated successfully. Jan 13 20:08:30.868377 systemd[1]: session-5.scope: Consumed 8.800s CPU time, 156.9M memory peak, 0B memory swap peak. Jan 13 20:08:30.870840 systemd-logind[1942]: Session 5 logged out. Waiting for processes to exit. Jan 13 20:08:30.875084 systemd-logind[1942]: Removed session 5. Jan 13 20:08:30.908473 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (3234) Jan 13 20:08:31.218534 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (3238) Jan 13 20:08:32.366273 kubelet[3164]: I0113 20:08:32.366222 3164 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 13 20:08:32.367294 containerd[1972]: time="2025-01-13T20:08:32.367208966Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 13 20:08:32.368557 kubelet[3164]: I0113 20:08:32.368231 3164 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 13 20:08:33.175145 systemd[1]: Created slice kubepods-besteffort-pod358d6950_e250_47ed_ae1c_f709aab3d186.slice - libcontainer container kubepods-besteffort-pod358d6950_e250_47ed_ae1c_f709aab3d186.slice. Jan 13 20:08:33.222942 systemd[1]: Created slice kubepods-burstable-podce79dad4_602d_45d1_8d51_cc327d785446.slice - libcontainer container kubepods-burstable-podce79dad4_602d_45d1_8d51_cc327d785446.slice. Jan 13 20:08:33.225583 kubelet[3164]: I0113 20:08:33.225399 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/358d6950-e250-47ed-ae1c-f709aab3d186-xtables-lock\") pod \"kube-proxy-zbs4d\" (UID: \"358d6950-e250-47ed-ae1c-f709aab3d186\") " pod="kube-system/kube-proxy-zbs4d" Jan 13 20:08:33.225892 kubelet[3164]: I0113 20:08:33.225735 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/358d6950-e250-47ed-ae1c-f709aab3d186-lib-modules\") pod \"kube-proxy-zbs4d\" (UID: \"358d6950-e250-47ed-ae1c-f709aab3d186\") " pod="kube-system/kube-proxy-zbs4d" Jan 13 20:08:33.225892 kubelet[3164]: I0113 20:08:33.225809 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcqfw\" (UniqueName: \"kubernetes.io/projected/358d6950-e250-47ed-ae1c-f709aab3d186-kube-api-access-fcqfw\") pod \"kube-proxy-zbs4d\" (UID: \"358d6950-e250-47ed-ae1c-f709aab3d186\") " pod="kube-system/kube-proxy-zbs4d" Jan 13 20:08:33.226524 kubelet[3164]: I0113 20:08:33.226272 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/358d6950-e250-47ed-ae1c-f709aab3d186-kube-proxy\") pod \"kube-proxy-zbs4d\" (UID: \"358d6950-e250-47ed-ae1c-f709aab3d186\") " pod="kube-system/kube-proxy-zbs4d" Jan 13 20:08:33.327842 kubelet[3164]: I0113 20:08:33.327271 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/ce79dad4-602d-45d1-8d51-cc327d785446-cni-plugin\") pod \"kube-flannel-ds-z8xpq\" (UID: \"ce79dad4-602d-45d1-8d51-cc327d785446\") " pod="kube-flannel/kube-flannel-ds-z8xpq" Jan 13 20:08:33.327842 kubelet[3164]: I0113 20:08:33.327360 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ce79dad4-602d-45d1-8d51-cc327d785446-run\") pod \"kube-flannel-ds-z8xpq\" (UID: \"ce79dad4-602d-45d1-8d51-cc327d785446\") " pod="kube-flannel/kube-flannel-ds-z8xpq" Jan 13 20:08:33.327842 kubelet[3164]: I0113 20:08:33.327495 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ce79dad4-602d-45d1-8d51-cc327d785446-xtables-lock\") pod \"kube-flannel-ds-z8xpq\" (UID: \"ce79dad4-602d-45d1-8d51-cc327d785446\") " pod="kube-flannel/kube-flannel-ds-z8xpq" Jan 13 20:08:33.327842 kubelet[3164]: I0113 20:08:33.327556 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvz6m\" (UniqueName: \"kubernetes.io/projected/ce79dad4-602d-45d1-8d51-cc327d785446-kube-api-access-kvz6m\") pod \"kube-flannel-ds-z8xpq\" (UID: \"ce79dad4-602d-45d1-8d51-cc327d785446\") " pod="kube-flannel/kube-flannel-ds-z8xpq" Jan 13 20:08:33.327842 kubelet[3164]: I0113 20:08:33.327649 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/ce79dad4-602d-45d1-8d51-cc327d785446-flannel-cfg\") pod \"kube-flannel-ds-z8xpq\" (UID: \"ce79dad4-602d-45d1-8d51-cc327d785446\") " pod="kube-flannel/kube-flannel-ds-z8xpq" Jan 13 20:08:33.329707 kubelet[3164]: I0113 20:08:33.328805 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/ce79dad4-602d-45d1-8d51-cc327d785446-cni\") pod \"kube-flannel-ds-z8xpq\" (UID: \"ce79dad4-602d-45d1-8d51-cc327d785446\") " pod="kube-flannel/kube-flannel-ds-z8xpq" Jan 13 20:08:33.489807 containerd[1972]: time="2025-01-13T20:08:33.489563476Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zbs4d,Uid:358d6950-e250-47ed-ae1c-f709aab3d186,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:33.531028 containerd[1972]: time="2025-01-13T20:08:33.530904652Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-z8xpq,Uid:ce79dad4-602d-45d1-8d51-cc327d785446,Namespace:kube-flannel,Attempt:0,}" Jan 13 20:08:33.553132 containerd[1972]: time="2025-01-13T20:08:33.552723124Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:08:33.553132 containerd[1972]: time="2025-01-13T20:08:33.552818320Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:08:33.553132 containerd[1972]: time="2025-01-13T20:08:33.552853612Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:33.553132 containerd[1972]: time="2025-01-13T20:08:33.553008748Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:33.627199 systemd[1]: Started cri-containerd-deb1e8fe741c1e06c94521425d05e8123af7976c8b042021747c77fc2aea761b.scope - libcontainer container deb1e8fe741c1e06c94521425d05e8123af7976c8b042021747c77fc2aea761b. Jan 13 20:08:33.640632 containerd[1972]: time="2025-01-13T20:08:33.640373344Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:08:33.640632 containerd[1972]: time="2025-01-13T20:08:33.640517320Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:08:33.641177 containerd[1972]: time="2025-01-13T20:08:33.640554460Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:33.641662 containerd[1972]: time="2025-01-13T20:08:33.641491432Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:33.684292 containerd[1972]: time="2025-01-13T20:08:33.684226277Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zbs4d,Uid:358d6950-e250-47ed-ae1c-f709aab3d186,Namespace:kube-system,Attempt:0,} returns sandbox id \"deb1e8fe741c1e06c94521425d05e8123af7976c8b042021747c77fc2aea761b\"" Jan 13 20:08:33.686843 systemd[1]: Started cri-containerd-f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044.scope - libcontainer container f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044. Jan 13 20:08:33.695973 containerd[1972]: time="2025-01-13T20:08:33.695894789Z" level=info msg="CreateContainer within sandbox \"deb1e8fe741c1e06c94521425d05e8123af7976c8b042021747c77fc2aea761b\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 13 20:08:33.737409 containerd[1972]: time="2025-01-13T20:08:33.737216153Z" level=info msg="CreateContainer within sandbox \"deb1e8fe741c1e06c94521425d05e8123af7976c8b042021747c77fc2aea761b\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"099e30cc32d0260d8d756dcc484c884c0dd42ee4c74082bc5950a2fdbec1d613\"" Jan 13 20:08:33.739589 containerd[1972]: time="2025-01-13T20:08:33.738477965Z" level=info msg="StartContainer for \"099e30cc32d0260d8d756dcc484c884c0dd42ee4c74082bc5950a2fdbec1d613\"" Jan 13 20:08:33.787554 containerd[1972]: time="2025-01-13T20:08:33.787335953Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-z8xpq,Uid:ce79dad4-602d-45d1-8d51-cc327d785446,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044\"" Jan 13 20:08:33.795350 containerd[1972]: time="2025-01-13T20:08:33.795281681Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\"" Jan 13 20:08:33.802767 systemd[1]: Started cri-containerd-099e30cc32d0260d8d756dcc484c884c0dd42ee4c74082bc5950a2fdbec1d613.scope - libcontainer container 099e30cc32d0260d8d756dcc484c884c0dd42ee4c74082bc5950a2fdbec1d613. Jan 13 20:08:33.896403 containerd[1972]: time="2025-01-13T20:08:33.896318610Z" level=info msg="StartContainer for \"099e30cc32d0260d8d756dcc484c884c0dd42ee4c74082bc5950a2fdbec1d613\" returns successfully" Jan 13 20:08:34.309776 kubelet[3164]: I0113 20:08:34.309159 3164 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-zbs4d" podStartSLOduration=1.309135604 podStartE2EDuration="1.309135604s" podCreationTimestamp="2025-01-13 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-13 20:08:34.307631524 +0000 UTC m=+5.348022556" watchObservedRunningTime="2025-01-13 20:08:34.309135604 +0000 UTC m=+5.349526612" Jan 13 20:08:35.694472 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3821464566.mount: Deactivated successfully. Jan 13 20:08:35.785624 containerd[1972]: time="2025-01-13T20:08:35.785546983Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:35.787347 containerd[1972]: time="2025-01-13T20:08:35.787223755Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3673531" Jan 13 20:08:35.789368 containerd[1972]: time="2025-01-13T20:08:35.789325159Z" level=info msg="ImageCreate event name:\"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:35.794146 containerd[1972]: time="2025-01-13T20:08:35.794049343Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:35.796767 containerd[1972]: time="2025-01-13T20:08:35.795685543Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3662650\" in 2.000341498s" Jan 13 20:08:35.796767 containerd[1972]: time="2025-01-13T20:08:35.795741715Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\"" Jan 13 20:08:35.801360 containerd[1972]: time="2025-01-13T20:08:35.801289627Z" level=info msg="CreateContainer within sandbox \"f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Jan 13 20:08:35.830376 containerd[1972]: time="2025-01-13T20:08:35.830304547Z" level=info msg="CreateContainer within sandbox \"f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122\"" Jan 13 20:08:35.832057 containerd[1972]: time="2025-01-13T20:08:35.831990127Z" level=info msg="StartContainer for \"8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122\"" Jan 13 20:08:35.879772 systemd[1]: Started cri-containerd-8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122.scope - libcontainer container 8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122. Jan 13 20:08:35.927418 containerd[1972]: time="2025-01-13T20:08:35.927141752Z" level=info msg="StartContainer for \"8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122\" returns successfully" Jan 13 20:08:35.931013 systemd[1]: cri-containerd-8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122.scope: Deactivated successfully. Jan 13 20:08:36.066500 containerd[1972]: time="2025-01-13T20:08:36.065992312Z" level=info msg="shim disconnected" id=8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122 namespace=k8s.io Jan 13 20:08:36.066500 containerd[1972]: time="2025-01-13T20:08:36.066109432Z" level=warning msg="cleaning up after shim disconnected" id=8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122 namespace=k8s.io Jan 13 20:08:36.066500 containerd[1972]: time="2025-01-13T20:08:36.066132196Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:08:36.296575 containerd[1972]: time="2025-01-13T20:08:36.296378538Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\"" Jan 13 20:08:36.558830 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8c0973670f2653f3b62967f529f6c86a3a7417d616c3a4ca80ed0e354f2e9122-rootfs.mount: Deactivated successfully. Jan 13 20:08:38.345581 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3277036495.mount: Deactivated successfully. Jan 13 20:08:39.620223 containerd[1972]: time="2025-01-13T20:08:39.620162794Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:39.622925 containerd[1972]: time="2025-01-13T20:08:39.622852690Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26874260" Jan 13 20:08:39.623790 containerd[1972]: time="2025-01-13T20:08:39.623720134Z" level=info msg="ImageCreate event name:\"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:39.631039 containerd[1972]: time="2025-01-13T20:08:39.630963670Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 13 20:08:39.632628 containerd[1972]: time="2025-01-13T20:08:39.632363158Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26863435\" in 3.335894032s" Jan 13 20:08:39.632628 containerd[1972]: time="2025-01-13T20:08:39.632417074Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\"" Jan 13 20:08:39.637981 containerd[1972]: time="2025-01-13T20:08:39.637920742Z" level=info msg="CreateContainer within sandbox \"f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 13 20:08:39.666326 containerd[1972]: time="2025-01-13T20:08:39.666182710Z" level=info msg="CreateContainer within sandbox \"f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f\"" Jan 13 20:08:39.666982 containerd[1972]: time="2025-01-13T20:08:39.666933202Z" level=info msg="StartContainer for \"258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f\"" Jan 13 20:08:39.725718 systemd[1]: Started cri-containerd-258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f.scope - libcontainer container 258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f. Jan 13 20:08:39.770585 systemd[1]: cri-containerd-258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f.scope: Deactivated successfully. Jan 13 20:08:39.773067 containerd[1972]: time="2025-01-13T20:08:39.772908803Z" level=info msg="StartContainer for \"258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f\" returns successfully" Jan 13 20:08:39.820033 kubelet[3164]: I0113 20:08:39.818700 3164 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Jan 13 20:08:39.882851 systemd[1]: Created slice kubepods-burstable-pod7b5ee5ce_cce3_4c1e_958a_b29a634549b4.slice - libcontainer container kubepods-burstable-pod7b5ee5ce_cce3_4c1e_958a_b29a634549b4.slice. Jan 13 20:08:39.903091 systemd[1]: Created slice kubepods-burstable-podfc1b70e9_2dac_457c_ab75_d830f1e0f93a.slice - libcontainer container kubepods-burstable-podfc1b70e9_2dac_457c_ab75_d830f1e0f93a.slice. Jan 13 20:08:39.970654 kubelet[3164]: I0113 20:08:39.970604 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2wxt\" (UniqueName: \"kubernetes.io/projected/7b5ee5ce-cce3-4c1e-958a-b29a634549b4-kube-api-access-h2wxt\") pod \"coredns-6f6b679f8f-tmjvp\" (UID: \"7b5ee5ce-cce3-4c1e-958a-b29a634549b4\") " pod="kube-system/coredns-6f6b679f8f-tmjvp" Jan 13 20:08:39.971190 kubelet[3164]: I0113 20:08:39.970989 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc1b70e9-2dac-457c-ab75-d830f1e0f93a-config-volume\") pod \"coredns-6f6b679f8f-289p9\" (UID: \"fc1b70e9-2dac-457c-ab75-d830f1e0f93a\") " pod="kube-system/coredns-6f6b679f8f-289p9" Jan 13 20:08:39.971190 kubelet[3164]: I0113 20:08:39.971036 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4knx\" (UniqueName: \"kubernetes.io/projected/fc1b70e9-2dac-457c-ab75-d830f1e0f93a-kube-api-access-s4knx\") pod \"coredns-6f6b679f8f-289p9\" (UID: \"fc1b70e9-2dac-457c-ab75-d830f1e0f93a\") " pod="kube-system/coredns-6f6b679f8f-289p9" Jan 13 20:08:39.971190 kubelet[3164]: I0113 20:08:39.971107 3164 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7b5ee5ce-cce3-4c1e-958a-b29a634549b4-config-volume\") pod \"coredns-6f6b679f8f-tmjvp\" (UID: \"7b5ee5ce-cce3-4c1e-958a-b29a634549b4\") " pod="kube-system/coredns-6f6b679f8f-tmjvp" Jan 13 20:08:40.197550 containerd[1972]: time="2025-01-13T20:08:40.196657161Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tmjvp,Uid:7b5ee5ce-cce3-4c1e-958a-b29a634549b4,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:40.210499 containerd[1972]: time="2025-01-13T20:08:40.210395337Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-289p9,Uid:fc1b70e9-2dac-457c-ab75-d830f1e0f93a,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:40.345457 containerd[1972]: time="2025-01-13T20:08:40.345139426Z" level=info msg="shim disconnected" id=258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f namespace=k8s.io Jan 13 20:08:40.345457 containerd[1972]: time="2025-01-13T20:08:40.345214990Z" level=warning msg="cleaning up after shim disconnected" id=258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f namespace=k8s.io Jan 13 20:08:40.345457 containerd[1972]: time="2025-01-13T20:08:40.345238702Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:08:40.393120 containerd[1972]: time="2025-01-13T20:08:40.393020038Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tmjvp,Uid:7b5ee5ce-cce3-4c1e-958a-b29a634549b4,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"80a5a9b95be2b52171bdf446265e967429061a4bfd2a770818675d813911dd73\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 13 20:08:40.393888 kubelet[3164]: E0113 20:08:40.393767 3164 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"80a5a9b95be2b52171bdf446265e967429061a4bfd2a770818675d813911dd73\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 13 20:08:40.394164 kubelet[3164]: E0113 20:08:40.394069 3164 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"80a5a9b95be2b52171bdf446265e967429061a4bfd2a770818675d813911dd73\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-tmjvp" Jan 13 20:08:40.394164 kubelet[3164]: E0113 20:08:40.394128 3164 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"80a5a9b95be2b52171bdf446265e967429061a4bfd2a770818675d813911dd73\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-tmjvp" Jan 13 20:08:40.394364 kubelet[3164]: E0113 20:08:40.394323 3164 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-tmjvp_kube-system(7b5ee5ce-cce3-4c1e-958a-b29a634549b4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-tmjvp_kube-system(7b5ee5ce-cce3-4c1e-958a-b29a634549b4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"80a5a9b95be2b52171bdf446265e967429061a4bfd2a770818675d813911dd73\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-6f6b679f8f-tmjvp" podUID="7b5ee5ce-cce3-4c1e-958a-b29a634549b4" Jan 13 20:08:40.403770 containerd[1972]: time="2025-01-13T20:08:40.403681810Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-289p9,Uid:fc1b70e9-2dac-457c-ab75-d830f1e0f93a,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"59a1fbf718d66c366f4a0a1d7410c5c1901e53bfe28253d1ede43622f2e582e9\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 13 20:08:40.404390 kubelet[3164]: E0113 20:08:40.404190 3164 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59a1fbf718d66c366f4a0a1d7410c5c1901e53bfe28253d1ede43622f2e582e9\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Jan 13 20:08:40.404390 kubelet[3164]: E0113 20:08:40.404285 3164 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59a1fbf718d66c366f4a0a1d7410c5c1901e53bfe28253d1ede43622f2e582e9\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-289p9" Jan 13 20:08:40.404390 kubelet[3164]: E0113 20:08:40.404318 3164 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59a1fbf718d66c366f4a0a1d7410c5c1901e53bfe28253d1ede43622f2e582e9\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-6f6b679f8f-289p9" Jan 13 20:08:40.405085 kubelet[3164]: E0113 20:08:40.404994 3164 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-289p9_kube-system(fc1b70e9-2dac-457c-ab75-d830f1e0f93a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-289p9_kube-system(fc1b70e9-2dac-457c-ab75-d830f1e0f93a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"59a1fbf718d66c366f4a0a1d7410c5c1901e53bfe28253d1ede43622f2e582e9\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-6f6b679f8f-289p9" podUID="fc1b70e9-2dac-457c-ab75-d830f1e0f93a" Jan 13 20:08:40.658744 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-258a1fd9352312b3285627465a9ddf9a59440c68dcb9b7760c4d6845c439071f-rootfs.mount: Deactivated successfully. Jan 13 20:08:41.316214 containerd[1972]: time="2025-01-13T20:08:41.316005035Z" level=info msg="CreateContainer within sandbox \"f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Jan 13 20:08:41.345098 containerd[1972]: time="2025-01-13T20:08:41.344973995Z" level=info msg="CreateContainer within sandbox \"f27765869c67bd44ef1574347af33aec9f48286a002480921f4b9546bd8f5044\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"2a66c6c4cb72bf9be1bcda45b5519f21045c118d9d916c17aef52a4bcb6b28a2\"" Jan 13 20:08:41.346867 containerd[1972]: time="2025-01-13T20:08:41.346672415Z" level=info msg="StartContainer for \"2a66c6c4cb72bf9be1bcda45b5519f21045c118d9d916c17aef52a4bcb6b28a2\"" Jan 13 20:08:41.402745 systemd[1]: Started cri-containerd-2a66c6c4cb72bf9be1bcda45b5519f21045c118d9d916c17aef52a4bcb6b28a2.scope - libcontainer container 2a66c6c4cb72bf9be1bcda45b5519f21045c118d9d916c17aef52a4bcb6b28a2. Jan 13 20:08:41.450761 containerd[1972]: time="2025-01-13T20:08:41.450604811Z" level=info msg="StartContainer for \"2a66c6c4cb72bf9be1bcda45b5519f21045c118d9d916c17aef52a4bcb6b28a2\" returns successfully" Jan 13 20:08:42.337361 kubelet[3164]: I0113 20:08:42.336296 3164 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-z8xpq" podStartSLOduration=3.494122843 podStartE2EDuration="9.336273288s" podCreationTimestamp="2025-01-13 20:08:33 +0000 UTC" firstStartedPulling="2025-01-13 20:08:33.792628205 +0000 UTC m=+4.833019213" lastFinishedPulling="2025-01-13 20:08:39.63477865 +0000 UTC m=+10.675169658" observedRunningTime="2025-01-13 20:08:42.336060768 +0000 UTC m=+13.376451788" watchObservedRunningTime="2025-01-13 20:08:42.336273288 +0000 UTC m=+13.376664296" Jan 13 20:08:42.569474 (udev-worker)[3896]: Network interface NamePolicy= disabled on kernel command line. Jan 13 20:08:42.597007 systemd-networkd[1836]: flannel.1: Link UP Jan 13 20:08:42.597021 systemd-networkd[1836]: flannel.1: Gained carrier Jan 13 20:08:44.098715 systemd-networkd[1836]: flannel.1: Gained IPv6LL Jan 13 20:08:46.393685 ntpd[1937]: Listen normally on 7 flannel.1 192.168.0.0:123 Jan 13 20:08:46.393844 ntpd[1937]: Listen normally on 8 flannel.1 [fe80::5c48:1bff:fe8a:58d4%4]:123 Jan 13 20:08:46.394298 ntpd[1937]: 13 Jan 20:08:46 ntpd[1937]: Listen normally on 7 flannel.1 192.168.0.0:123 Jan 13 20:08:46.394298 ntpd[1937]: 13 Jan 20:08:46 ntpd[1937]: Listen normally on 8 flannel.1 [fe80::5c48:1bff:fe8a:58d4%4]:123 Jan 13 20:08:52.232983 containerd[1972]: time="2025-01-13T20:08:52.232889349Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-289p9,Uid:fc1b70e9-2dac-457c-ab75-d830f1e0f93a,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:52.233662 containerd[1972]: time="2025-01-13T20:08:52.232889289Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tmjvp,Uid:7b5ee5ce-cce3-4c1e-958a-b29a634549b4,Namespace:kube-system,Attempt:0,}" Jan 13 20:08:52.283465 systemd-networkd[1836]: cni0: Link UP Jan 13 20:08:52.283480 systemd-networkd[1836]: cni0: Gained carrier Jan 13 20:08:52.301491 (udev-worker)[4032]: Network interface NamePolicy= disabled on kernel command line. Jan 13 20:08:52.302594 systemd-networkd[1836]: cni0: Lost carrier Jan 13 20:08:52.306568 systemd-networkd[1836]: veth0113236b: Link UP Jan 13 20:08:52.314015 kernel: cni0: port 1(vethb41cfc85) entered blocking state Jan 13 20:08:52.314128 kernel: cni0: port 1(vethb41cfc85) entered disabled state Jan 13 20:08:52.314169 kernel: vethb41cfc85: entered allmulticast mode Jan 13 20:08:52.313250 systemd-networkd[1836]: vethb41cfc85: Link UP Jan 13 20:08:52.317845 kernel: vethb41cfc85: entered promiscuous mode Jan 13 20:08:52.317941 kernel: cni0: port 1(vethb41cfc85) entered blocking state Jan 13 20:08:52.317980 kernel: cni0: port 1(vethb41cfc85) entered forwarding state Jan 13 20:08:52.320047 kernel: cni0: port 2(veth0113236b) entered blocking state Jan 13 20:08:52.322520 kernel: cni0: port 2(veth0113236b) entered disabled state Jan 13 20:08:52.323891 kernel: veth0113236b: entered allmulticast mode Jan 13 20:08:52.325496 kernel: veth0113236b: entered promiscuous mode Jan 13 20:08:52.328227 kernel: cni0: port 2(veth0113236b) entered blocking state Jan 13 20:08:52.328290 kernel: cni0: port 2(veth0113236b) entered forwarding state Jan 13 20:08:52.331079 kernel: cni0: port 2(veth0113236b) entered disabled state Jan 13 20:08:52.331188 kernel: cni0: port 1(vethb41cfc85) entered disabled state Jan 13 20:08:52.333036 (udev-worker)[4033]: Network interface NamePolicy= disabled on kernel command line. Jan 13 20:08:52.340288 (udev-worker)[4037]: Network interface NamePolicy= disabled on kernel command line. Jan 13 20:08:52.360213 kernel: cni0: port 1(vethb41cfc85) entered blocking state Jan 13 20:08:52.360338 kernel: cni0: port 1(vethb41cfc85) entered forwarding state Jan 13 20:08:52.359525 systemd-networkd[1836]: vethb41cfc85: Gained carrier Jan 13 20:08:52.361591 systemd-networkd[1836]: cni0: Gained carrier Jan 13 20:08:52.368916 kernel: cni0: port 2(veth0113236b) entered blocking state Jan 13 20:08:52.369056 kernel: cni0: port 2(veth0113236b) entered forwarding state Jan 13 20:08:52.370371 systemd-networkd[1836]: veth0113236b: Gained carrier Jan 13 20:08:52.371691 containerd[1972]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x4000012938), "name":"cbr0", "type":"bridge"} Jan 13 20:08:52.371691 containerd[1972]: delegateAdd: netconf sent to delegate plugin: Jan 13 20:08:52.385703 containerd[1972]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":8951,"name":"cbr0","type":"bridge"} Jan 13 20:08:52.385703 containerd[1972]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x40000928e8), "name":"cbr0", "type":"bridge"} Jan 13 20:08:52.385703 containerd[1972]: delegateAdd: netconf sent to delegate plugin: Jan 13 20:08:52.418005 containerd[1972]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":8951,"name":"cbr0","type":"bridge"}time="2025-01-13T20:08:52.417651718Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:08:52.419322 containerd[1972]: time="2025-01-13T20:08:52.417974062Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:08:52.419322 containerd[1972]: time="2025-01-13T20:08:52.419031322Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:52.419580 containerd[1972]: time="2025-01-13T20:08:52.419236750Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:52.435853 containerd[1972]: time="2025-01-13T20:08:52.435699286Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 13 20:08:52.437189 containerd[1972]: time="2025-01-13T20:08:52.436969630Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 13 20:08:52.437521 containerd[1972]: time="2025-01-13T20:08:52.437258434Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:52.437970 containerd[1972]: time="2025-01-13T20:08:52.437767390Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 13 20:08:52.475768 systemd[1]: Started cri-containerd-c65832cdc084a80388d27d4173b23304496869280d029d1cae63ae407a881d8c.scope - libcontainer container c65832cdc084a80388d27d4173b23304496869280d029d1cae63ae407a881d8c. Jan 13 20:08:52.491759 systemd[1]: Started cri-containerd-beec17200e4f79ee22eb350b418c9b2c161138c016a3fe75461fd9f27f77c635.scope - libcontainer container beec17200e4f79ee22eb350b418c9b2c161138c016a3fe75461fd9f27f77c635. Jan 13 20:08:52.581515 containerd[1972]: time="2025-01-13T20:08:52.581376142Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-289p9,Uid:fc1b70e9-2dac-457c-ab75-d830f1e0f93a,Namespace:kube-system,Attempt:0,} returns sandbox id \"c65832cdc084a80388d27d4173b23304496869280d029d1cae63ae407a881d8c\"" Jan 13 20:08:52.590123 containerd[1972]: time="2025-01-13T20:08:52.589853926Z" level=info msg="CreateContainer within sandbox \"c65832cdc084a80388d27d4173b23304496869280d029d1cae63ae407a881d8c\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 13 20:08:52.600644 containerd[1972]: time="2025-01-13T20:08:52.600575279Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tmjvp,Uid:7b5ee5ce-cce3-4c1e-958a-b29a634549b4,Namespace:kube-system,Attempt:0,} returns sandbox id \"beec17200e4f79ee22eb350b418c9b2c161138c016a3fe75461fd9f27f77c635\"" Jan 13 20:08:52.607555 containerd[1972]: time="2025-01-13T20:08:52.607233719Z" level=info msg="CreateContainer within sandbox \"beec17200e4f79ee22eb350b418c9b2c161138c016a3fe75461fd9f27f77c635\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 13 20:08:52.624559 containerd[1972]: time="2025-01-13T20:08:52.624497027Z" level=info msg="CreateContainer within sandbox \"c65832cdc084a80388d27d4173b23304496869280d029d1cae63ae407a881d8c\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a781ff95c6b5015bcbe67ef59b07100db685f27dfba5caef98b00bf76612b7f6\"" Jan 13 20:08:52.626760 containerd[1972]: time="2025-01-13T20:08:52.625181255Z" level=info msg="StartContainer for \"a781ff95c6b5015bcbe67ef59b07100db685f27dfba5caef98b00bf76612b7f6\"" Jan 13 20:08:52.641341 containerd[1972]: time="2025-01-13T20:08:52.641279063Z" level=info msg="CreateContainer within sandbox \"beec17200e4f79ee22eb350b418c9b2c161138c016a3fe75461fd9f27f77c635\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"5295989a199a564b9cba5a5752da73fe1f2a7c35c2aad94d0bafd650feb86543\"" Jan 13 20:08:52.642939 containerd[1972]: time="2025-01-13T20:08:52.642861755Z" level=info msg="StartContainer for \"5295989a199a564b9cba5a5752da73fe1f2a7c35c2aad94d0bafd650feb86543\"" Jan 13 20:08:52.683761 systemd[1]: Started cri-containerd-a781ff95c6b5015bcbe67ef59b07100db685f27dfba5caef98b00bf76612b7f6.scope - libcontainer container a781ff95c6b5015bcbe67ef59b07100db685f27dfba5caef98b00bf76612b7f6. Jan 13 20:08:52.705724 systemd[1]: Started cri-containerd-5295989a199a564b9cba5a5752da73fe1f2a7c35c2aad94d0bafd650feb86543.scope - libcontainer container 5295989a199a564b9cba5a5752da73fe1f2a7c35c2aad94d0bafd650feb86543. Jan 13 20:08:52.769138 containerd[1972]: time="2025-01-13T20:08:52.768583667Z" level=info msg="StartContainer for \"a781ff95c6b5015bcbe67ef59b07100db685f27dfba5caef98b00bf76612b7f6\" returns successfully" Jan 13 20:08:52.801113 containerd[1972]: time="2025-01-13T20:08:52.801043968Z" level=info msg="StartContainer for \"5295989a199a564b9cba5a5752da73fe1f2a7c35c2aad94d0bafd650feb86543\" returns successfully" Jan 13 20:08:53.406917 kubelet[3164]: I0113 20:08:53.405279 3164 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-289p9" podStartSLOduration=20.405255071 podStartE2EDuration="20.405255071s" podCreationTimestamp="2025-01-13 20:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-13 20:08:53.382564354 +0000 UTC m=+24.422955374" watchObservedRunningTime="2025-01-13 20:08:53.405255071 +0000 UTC m=+24.445646079" Jan 13 20:08:53.954823 systemd-networkd[1836]: cni0: Gained IPv6LL Jan 13 20:08:54.082756 systemd-networkd[1836]: veth0113236b: Gained IPv6LL Jan 13 20:08:54.210753 systemd-networkd[1836]: vethb41cfc85: Gained IPv6LL Jan 13 20:08:56.393653 ntpd[1937]: Listen normally on 9 cni0 192.168.0.1:123 Jan 13 20:08:56.393821 ntpd[1937]: Listen normally on 10 cni0 [fe80::ccd0:68ff:fe85:989d%5]:123 Jan 13 20:08:56.394231 ntpd[1937]: 13 Jan 20:08:56 ntpd[1937]: Listen normally on 9 cni0 192.168.0.1:123 Jan 13 20:08:56.394231 ntpd[1937]: 13 Jan 20:08:56 ntpd[1937]: Listen normally on 10 cni0 [fe80::ccd0:68ff:fe85:989d%5]:123 Jan 13 20:08:56.394231 ntpd[1937]: 13 Jan 20:08:56 ntpd[1937]: Listen normally on 11 veth0113236b [fe80::70d6:17ff:fee8:d5e1%6]:123 Jan 13 20:08:56.394231 ntpd[1937]: 13 Jan 20:08:56 ntpd[1937]: Listen normally on 12 vethb41cfc85 [fe80::b84c:6fff:fe2e:c727%7]:123 Jan 13 20:08:56.393905 ntpd[1937]: Listen normally on 11 veth0113236b [fe80::70d6:17ff:fee8:d5e1%6]:123 Jan 13 20:08:56.393974 ntpd[1937]: Listen normally on 12 vethb41cfc85 [fe80::b84c:6fff:fe2e:c727%7]:123 Jan 13 20:09:12.550955 systemd[1]: Started sshd@5-172.31.20.118:22-147.75.109.163:56748.service - OpenSSH per-connection server daemon (147.75.109.163:56748). Jan 13 20:09:12.737106 sshd[4309]: Accepted publickey for core from 147.75.109.163 port 56748 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:12.739614 sshd-session[4309]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:12.748100 systemd-logind[1942]: New session 6 of user core. Jan 13 20:09:12.755673 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 13 20:09:13.019093 sshd[4311]: Connection closed by 147.75.109.163 port 56748 Jan 13 20:09:13.019368 sshd-session[4309]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:13.027028 systemd[1]: sshd@5-172.31.20.118:22-147.75.109.163:56748.service: Deactivated successfully. Jan 13 20:09:13.031999 systemd[1]: session-6.scope: Deactivated successfully. Jan 13 20:09:13.035418 systemd-logind[1942]: Session 6 logged out. Waiting for processes to exit. Jan 13 20:09:13.038167 systemd-logind[1942]: Removed session 6. Jan 13 20:09:18.057057 systemd[1]: Started sshd@6-172.31.20.118:22-147.75.109.163:52562.service - OpenSSH per-connection server daemon (147.75.109.163:52562). Jan 13 20:09:18.247011 sshd[4359]: Accepted publickey for core from 147.75.109.163 port 52562 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:18.249522 sshd-session[4359]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:18.257247 systemd-logind[1942]: New session 7 of user core. Jan 13 20:09:18.262693 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 13 20:09:18.509163 sshd[4367]: Connection closed by 147.75.109.163 port 52562 Jan 13 20:09:18.509043 sshd-session[4359]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:18.514312 systemd[1]: sshd@6-172.31.20.118:22-147.75.109.163:52562.service: Deactivated successfully. Jan 13 20:09:18.514877 systemd-logind[1942]: Session 7 logged out. Waiting for processes to exit. Jan 13 20:09:18.519335 systemd[1]: session-7.scope: Deactivated successfully. Jan 13 20:09:18.523011 systemd-logind[1942]: Removed session 7. Jan 13 20:09:23.549958 systemd[1]: Started sshd@7-172.31.20.118:22-147.75.109.163:52568.service - OpenSSH per-connection server daemon (147.75.109.163:52568). Jan 13 20:09:23.733047 sshd[4404]: Accepted publickey for core from 147.75.109.163 port 52568 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:23.735541 sshd-session[4404]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:23.743889 systemd-logind[1942]: New session 8 of user core. Jan 13 20:09:23.749745 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 13 20:09:23.990458 sshd[4406]: Connection closed by 147.75.109.163 port 52568 Jan 13 20:09:23.991271 sshd-session[4404]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:23.997231 systemd-logind[1942]: Session 8 logged out. Waiting for processes to exit. Jan 13 20:09:23.999150 systemd[1]: sshd@7-172.31.20.118:22-147.75.109.163:52568.service: Deactivated successfully. Jan 13 20:09:24.002635 systemd[1]: session-8.scope: Deactivated successfully. Jan 13 20:09:24.005705 systemd-logind[1942]: Removed session 8. Jan 13 20:09:24.028665 systemd[1]: Started sshd@8-172.31.20.118:22-147.75.109.163:52574.service - OpenSSH per-connection server daemon (147.75.109.163:52574). Jan 13 20:09:24.219663 sshd[4418]: Accepted publickey for core from 147.75.109.163 port 52574 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:24.222223 sshd-session[4418]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:24.229556 systemd-logind[1942]: New session 9 of user core. Jan 13 20:09:24.241741 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 13 20:09:24.565564 sshd[4420]: Connection closed by 147.75.109.163 port 52574 Jan 13 20:09:24.566024 sshd-session[4418]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:24.574043 systemd[1]: sshd@8-172.31.20.118:22-147.75.109.163:52574.service: Deactivated successfully. Jan 13 20:09:24.580820 systemd[1]: session-9.scope: Deactivated successfully. Jan 13 20:09:24.587894 systemd-logind[1942]: Session 9 logged out. Waiting for processes to exit. Jan 13 20:09:24.608171 systemd[1]: Started sshd@9-172.31.20.118:22-147.75.109.163:52584.service - OpenSSH per-connection server daemon (147.75.109.163:52584). Jan 13 20:09:24.609517 systemd-logind[1942]: Removed session 9. Jan 13 20:09:24.815228 sshd[4429]: Accepted publickey for core from 147.75.109.163 port 52584 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:24.817745 sshd-session[4429]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:24.826540 systemd-logind[1942]: New session 10 of user core. Jan 13 20:09:24.833708 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 13 20:09:25.080838 sshd[4431]: Connection closed by 147.75.109.163 port 52584 Jan 13 20:09:25.080541 sshd-session[4429]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:25.086506 systemd-logind[1942]: Session 10 logged out. Waiting for processes to exit. Jan 13 20:09:25.086818 systemd[1]: sshd@9-172.31.20.118:22-147.75.109.163:52584.service: Deactivated successfully. Jan 13 20:09:25.090491 systemd[1]: session-10.scope: Deactivated successfully. Jan 13 20:09:25.095067 systemd-logind[1942]: Removed session 10. Jan 13 20:09:30.122009 systemd[1]: Started sshd@10-172.31.20.118:22-147.75.109.163:52740.service - OpenSSH per-connection server daemon (147.75.109.163:52740). Jan 13 20:09:30.311472 sshd[4465]: Accepted publickey for core from 147.75.109.163 port 52740 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:30.313942 sshd-session[4465]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:30.321504 systemd-logind[1942]: New session 11 of user core. Jan 13 20:09:30.328700 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 13 20:09:30.575999 sshd[4468]: Connection closed by 147.75.109.163 port 52740 Jan 13 20:09:30.576786 sshd-session[4465]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:30.582623 systemd-logind[1942]: Session 11 logged out. Waiting for processes to exit. Jan 13 20:09:30.584518 systemd[1]: sshd@10-172.31.20.118:22-147.75.109.163:52740.service: Deactivated successfully. Jan 13 20:09:30.589514 systemd[1]: session-11.scope: Deactivated successfully. Jan 13 20:09:30.591659 systemd-logind[1942]: Removed session 11. Jan 13 20:09:35.611929 systemd[1]: Started sshd@11-172.31.20.118:22-147.75.109.163:52754.service - OpenSSH per-connection server daemon (147.75.109.163:52754). Jan 13 20:09:35.802299 sshd[4503]: Accepted publickey for core from 147.75.109.163 port 52754 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:35.804922 sshd-session[4503]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:35.813270 systemd-logind[1942]: New session 12 of user core. Jan 13 20:09:35.821696 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 13 20:09:36.058127 sshd[4505]: Connection closed by 147.75.109.163 port 52754 Jan 13 20:09:36.059315 sshd-session[4503]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:36.065579 systemd[1]: sshd@11-172.31.20.118:22-147.75.109.163:52754.service: Deactivated successfully. Jan 13 20:09:36.068673 systemd[1]: session-12.scope: Deactivated successfully. Jan 13 20:09:36.071081 systemd-logind[1942]: Session 12 logged out. Waiting for processes to exit. Jan 13 20:09:36.073422 systemd-logind[1942]: Removed session 12. Jan 13 20:09:41.098943 systemd[1]: Started sshd@12-172.31.20.118:22-147.75.109.163:39428.service - OpenSSH per-connection server daemon (147.75.109.163:39428). Jan 13 20:09:41.290694 sshd[4537]: Accepted publickey for core from 147.75.109.163 port 39428 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:41.294076 sshd-session[4537]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:41.304096 systemd-logind[1942]: New session 13 of user core. Jan 13 20:09:41.315680 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 13 20:09:41.561846 sshd[4539]: Connection closed by 147.75.109.163 port 39428 Jan 13 20:09:41.562876 sshd-session[4537]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:41.567846 systemd-logind[1942]: Session 13 logged out. Waiting for processes to exit. Jan 13 20:09:41.569094 systemd[1]: sshd@12-172.31.20.118:22-147.75.109.163:39428.service: Deactivated successfully. Jan 13 20:09:41.573135 systemd[1]: session-13.scope: Deactivated successfully. Jan 13 20:09:41.577185 systemd-logind[1942]: Removed session 13. Jan 13 20:09:46.601976 systemd[1]: Started sshd@13-172.31.20.118:22-147.75.109.163:39430.service - OpenSSH per-connection server daemon (147.75.109.163:39430). Jan 13 20:09:46.800062 sshd[4571]: Accepted publickey for core from 147.75.109.163 port 39430 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:46.803315 sshd-session[4571]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:46.811579 systemd-logind[1942]: New session 14 of user core. Jan 13 20:09:46.822678 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 13 20:09:47.068489 sshd[4573]: Connection closed by 147.75.109.163 port 39430 Jan 13 20:09:47.069282 sshd-session[4571]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:47.075911 systemd[1]: sshd@13-172.31.20.118:22-147.75.109.163:39430.service: Deactivated successfully. Jan 13 20:09:47.079238 systemd[1]: session-14.scope: Deactivated successfully. Jan 13 20:09:47.081197 systemd-logind[1942]: Session 14 logged out. Waiting for processes to exit. Jan 13 20:09:47.083303 systemd-logind[1942]: Removed session 14. Jan 13 20:09:47.106660 systemd[1]: Started sshd@14-172.31.20.118:22-147.75.109.163:39432.service - OpenSSH per-connection server daemon (147.75.109.163:39432). Jan 13 20:09:47.309395 sshd[4584]: Accepted publickey for core from 147.75.109.163 port 39432 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:47.311811 sshd-session[4584]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:47.319109 systemd-logind[1942]: New session 15 of user core. Jan 13 20:09:47.331698 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 13 20:09:47.622158 sshd[4586]: Connection closed by 147.75.109.163 port 39432 Jan 13 20:09:47.622714 sshd-session[4584]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:47.629758 systemd[1]: sshd@14-172.31.20.118:22-147.75.109.163:39432.service: Deactivated successfully. Jan 13 20:09:47.633471 systemd[1]: session-15.scope: Deactivated successfully. Jan 13 20:09:47.636325 systemd-logind[1942]: Session 15 logged out. Waiting for processes to exit. Jan 13 20:09:47.638260 systemd-logind[1942]: Removed session 15. Jan 13 20:09:47.657861 systemd[1]: Started sshd@15-172.31.20.118:22-147.75.109.163:42086.service - OpenSSH per-connection server daemon (147.75.109.163:42086). Jan 13 20:09:47.855465 sshd[4595]: Accepted publickey for core from 147.75.109.163 port 42086 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:47.857960 sshd-session[4595]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:47.865898 systemd-logind[1942]: New session 16 of user core. Jan 13 20:09:47.875744 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 13 20:09:50.165180 sshd[4597]: Connection closed by 147.75.109.163 port 42086 Jan 13 20:09:50.167944 sshd-session[4595]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:50.176118 systemd[1]: sshd@15-172.31.20.118:22-147.75.109.163:42086.service: Deactivated successfully. Jan 13 20:09:50.183046 systemd[1]: session-16.scope: Deactivated successfully. Jan 13 20:09:50.189508 systemd-logind[1942]: Session 16 logged out. Waiting for processes to exit. Jan 13 20:09:50.217970 systemd[1]: Started sshd@16-172.31.20.118:22-147.75.109.163:42090.service - OpenSSH per-connection server daemon (147.75.109.163:42090). Jan 13 20:09:50.220149 systemd-logind[1942]: Removed session 16. Jan 13 20:09:50.412638 sshd[4634]: Accepted publickey for core from 147.75.109.163 port 42090 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:50.415093 sshd-session[4634]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:50.423301 systemd-logind[1942]: New session 17 of user core. Jan 13 20:09:50.429680 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 13 20:09:50.905098 sshd[4636]: Connection closed by 147.75.109.163 port 42090 Jan 13 20:09:50.905625 sshd-session[4634]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:50.911389 systemd-logind[1942]: Session 17 logged out. Waiting for processes to exit. Jan 13 20:09:50.912010 systemd[1]: sshd@16-172.31.20.118:22-147.75.109.163:42090.service: Deactivated successfully. Jan 13 20:09:50.916846 systemd[1]: session-17.scope: Deactivated successfully. Jan 13 20:09:50.920551 systemd-logind[1942]: Removed session 17. Jan 13 20:09:50.944987 systemd[1]: Started sshd@17-172.31.20.118:22-147.75.109.163:42106.service - OpenSSH per-connection server daemon (147.75.109.163:42106). Jan 13 20:09:51.133898 sshd[4644]: Accepted publickey for core from 147.75.109.163 port 42106 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:51.136371 sshd-session[4644]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:51.144754 systemd-logind[1942]: New session 18 of user core. Jan 13 20:09:51.154684 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 13 20:09:51.389411 sshd[4646]: Connection closed by 147.75.109.163 port 42106 Jan 13 20:09:51.390473 sshd-session[4644]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:51.395967 systemd[1]: sshd@17-172.31.20.118:22-147.75.109.163:42106.service: Deactivated successfully. Jan 13 20:09:51.401725 systemd[1]: session-18.scope: Deactivated successfully. Jan 13 20:09:51.404491 systemd-logind[1942]: Session 18 logged out. Waiting for processes to exit. Jan 13 20:09:51.406533 systemd-logind[1942]: Removed session 18. Jan 13 20:09:56.433928 systemd[1]: Started sshd@18-172.31.20.118:22-147.75.109.163:42110.service - OpenSSH per-connection server daemon (147.75.109.163:42110). Jan 13 20:09:56.620879 sshd[4679]: Accepted publickey for core from 147.75.109.163 port 42110 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:09:56.623618 sshd-session[4679]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:09:56.634363 systemd-logind[1942]: New session 19 of user core. Jan 13 20:09:56.641767 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 13 20:09:56.878300 sshd[4681]: Connection closed by 147.75.109.163 port 42110 Jan 13 20:09:56.879422 sshd-session[4679]: pam_unix(sshd:session): session closed for user core Jan 13 20:09:56.885032 systemd[1]: sshd@18-172.31.20.118:22-147.75.109.163:42110.service: Deactivated successfully. Jan 13 20:09:56.889643 systemd[1]: session-19.scope: Deactivated successfully. Jan 13 20:09:56.892634 systemd-logind[1942]: Session 19 logged out. Waiting for processes to exit. Jan 13 20:09:56.894908 systemd-logind[1942]: Removed session 19. Jan 13 20:10:01.920935 systemd[1]: Started sshd@19-172.31.20.118:22-147.75.109.163:35444.service - OpenSSH per-connection server daemon (147.75.109.163:35444). Jan 13 20:10:02.109856 sshd[4716]: Accepted publickey for core from 147.75.109.163 port 35444 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:10:02.112314 sshd-session[4716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:10:02.119331 systemd-logind[1942]: New session 20 of user core. Jan 13 20:10:02.126695 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 13 20:10:02.376635 sshd[4718]: Connection closed by 147.75.109.163 port 35444 Jan 13 20:10:02.376516 sshd-session[4716]: pam_unix(sshd:session): session closed for user core Jan 13 20:10:02.385417 systemd[1]: sshd@19-172.31.20.118:22-147.75.109.163:35444.service: Deactivated successfully. Jan 13 20:10:02.391803 systemd[1]: session-20.scope: Deactivated successfully. Jan 13 20:10:02.394960 systemd-logind[1942]: Session 20 logged out. Waiting for processes to exit. Jan 13 20:10:02.397443 systemd-logind[1942]: Removed session 20. Jan 13 20:10:07.417101 systemd[1]: Started sshd@20-172.31.20.118:22-147.75.109.163:36166.service - OpenSSH per-connection server daemon (147.75.109.163:36166). Jan 13 20:10:07.596887 sshd[4752]: Accepted publickey for core from 147.75.109.163 port 36166 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:10:07.598824 sshd-session[4752]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:10:07.606792 systemd-logind[1942]: New session 21 of user core. Jan 13 20:10:07.618676 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 13 20:10:07.849227 sshd[4754]: Connection closed by 147.75.109.163 port 36166 Jan 13 20:10:07.850108 sshd-session[4752]: pam_unix(sshd:session): session closed for user core Jan 13 20:10:07.855994 systemd[1]: sshd@20-172.31.20.118:22-147.75.109.163:36166.service: Deactivated successfully. Jan 13 20:10:07.861603 systemd[1]: session-21.scope: Deactivated successfully. Jan 13 20:10:07.863375 systemd-logind[1942]: Session 21 logged out. Waiting for processes to exit. Jan 13 20:10:07.866052 systemd-logind[1942]: Removed session 21. Jan 13 20:10:12.894911 systemd[1]: Started sshd@21-172.31.20.118:22-147.75.109.163:36172.service - OpenSSH per-connection server daemon (147.75.109.163:36172). Jan 13 20:10:13.074217 sshd[4785]: Accepted publickey for core from 147.75.109.163 port 36172 ssh2: RSA SHA256:IRHkteilZRLg/mCVEzdResksy7NfUBDRRywgALKaHg0 Jan 13 20:10:13.076742 sshd-session[4785]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 13 20:10:13.085210 systemd-logind[1942]: New session 22 of user core. Jan 13 20:10:13.094670 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 13 20:10:13.341500 sshd[4793]: Connection closed by 147.75.109.163 port 36172 Jan 13 20:10:13.342339 sshd-session[4785]: pam_unix(sshd:session): session closed for user core Jan 13 20:10:13.348556 systemd[1]: sshd@21-172.31.20.118:22-147.75.109.163:36172.service: Deactivated successfully. Jan 13 20:10:13.352746 systemd[1]: session-22.scope: Deactivated successfully. Jan 13 20:10:13.354404 systemd-logind[1942]: Session 22 logged out. Waiting for processes to exit. Jan 13 20:10:13.356337 systemd-logind[1942]: Removed session 22. Jan 13 20:10:28.028230 systemd[1]: cri-containerd-5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc.scope: Deactivated successfully. Jan 13 20:10:28.028917 systemd[1]: cri-containerd-5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc.scope: Consumed 3.693s CPU time, 17.8M memory peak, 0B memory swap peak. Jan 13 20:10:28.075862 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc-rootfs.mount: Deactivated successfully. Jan 13 20:10:28.085632 containerd[1972]: time="2025-01-13T20:10:28.085555901Z" level=info msg="shim disconnected" id=5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc namespace=k8s.io Jan 13 20:10:28.086566 containerd[1972]: time="2025-01-13T20:10:28.086224757Z" level=warning msg="cleaning up after shim disconnected" id=5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc namespace=k8s.io Jan 13 20:10:28.086566 containerd[1972]: time="2025-01-13T20:10:28.086259437Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:10:28.597398 kubelet[3164]: I0113 20:10:28.597285 3164 scope.go:117] "RemoveContainer" containerID="5a32f5328df93b074d18ac1613cadf07deba2dfdc97c0890778f1de28e92d9bc" Jan 13 20:10:28.601554 containerd[1972]: time="2025-01-13T20:10:28.601403347Z" level=info msg="CreateContainer within sandbox \"2dc0640630805fff70f7416dbcd18199d313f521d58b929e5844dfd5cafef48e\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Jan 13 20:10:28.636449 containerd[1972]: time="2025-01-13T20:10:28.636350300Z" level=info msg="CreateContainer within sandbox \"2dc0640630805fff70f7416dbcd18199d313f521d58b929e5844dfd5cafef48e\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"cc69e6e2301be8437b6937c6c7c1c93a04dc12a03b7d2392c588952846f54d2a\"" Jan 13 20:10:28.637509 containerd[1972]: time="2025-01-13T20:10:28.637056368Z" level=info msg="StartContainer for \"cc69e6e2301be8437b6937c6c7c1c93a04dc12a03b7d2392c588952846f54d2a\"" Jan 13 20:10:28.689753 systemd[1]: Started cri-containerd-cc69e6e2301be8437b6937c6c7c1c93a04dc12a03b7d2392c588952846f54d2a.scope - libcontainer container cc69e6e2301be8437b6937c6c7c1c93a04dc12a03b7d2392c588952846f54d2a. Jan 13 20:10:28.756728 containerd[1972]: time="2025-01-13T20:10:28.756653708Z" level=info msg="StartContainer for \"cc69e6e2301be8437b6937c6c7c1c93a04dc12a03b7d2392c588952846f54d2a\" returns successfully" Jan 13 20:10:31.501619 kubelet[3164]: E0113 20:10:31.501166 3164 controller.go:195] "Failed to update lease" err="Put \"https://172.31.20.118:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-118?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 13 20:10:31.817755 systemd[1]: cri-containerd-79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5.scope: Deactivated successfully. Jan 13 20:10:31.818237 systemd[1]: cri-containerd-79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5.scope: Consumed 3.126s CPU time, 16.0M memory peak, 0B memory swap peak. Jan 13 20:10:31.860153 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5-rootfs.mount: Deactivated successfully. Jan 13 20:10:31.897011 containerd[1972]: time="2025-01-13T20:10:31.896932812Z" level=info msg="shim disconnected" id=79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5 namespace=k8s.io Jan 13 20:10:31.897011 containerd[1972]: time="2025-01-13T20:10:31.897005976Z" level=warning msg="cleaning up after shim disconnected" id=79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5 namespace=k8s.io Jan 13 20:10:31.898145 containerd[1972]: time="2025-01-13T20:10:31.897028284Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 13 20:10:32.612074 kubelet[3164]: I0113 20:10:32.612012 3164 scope.go:117] "RemoveContainer" containerID="79ee69d83c61c20af911406a63ca9599a80f2cccbb77b5e0bd0bfb420c8adbc5" Jan 13 20:10:32.615072 containerd[1972]: time="2025-01-13T20:10:32.614853167Z" level=info msg="CreateContainer within sandbox \"cf5ce157ce50049b9cc6feef6f92074be3eb00eea3888cd80fbbe130c3e77a28\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Jan 13 20:10:32.641582 containerd[1972]: time="2025-01-13T20:10:32.641458427Z" level=info msg="CreateContainer within sandbox \"cf5ce157ce50049b9cc6feef6f92074be3eb00eea3888cd80fbbe130c3e77a28\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"a004eba19350beb128867162d09c12f786d3fcbbf735080ccd700976e336e7dc\"" Jan 13 20:10:32.642718 containerd[1972]: time="2025-01-13T20:10:32.642072983Z" level=info msg="StartContainer for \"a004eba19350beb128867162d09c12f786d3fcbbf735080ccd700976e336e7dc\"" Jan 13 20:10:32.698754 systemd[1]: Started cri-containerd-a004eba19350beb128867162d09c12f786d3fcbbf735080ccd700976e336e7dc.scope - libcontainer container a004eba19350beb128867162d09c12f786d3fcbbf735080ccd700976e336e7dc. Jan 13 20:10:32.762525 containerd[1972]: time="2025-01-13T20:10:32.762388032Z" level=info msg="StartContainer for \"a004eba19350beb128867162d09c12f786d3fcbbf735080ccd700976e336e7dc\" returns successfully" Jan 13 20:10:41.501889 kubelet[3164]: E0113 20:10:41.501570 3164 controller.go:195] "Failed to update lease" err="Put \"https://172.31.20.118:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-118?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"