Feb 13 19:48:40.172753 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd083] Feb 13 19:48:40.172799 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Thu Feb 13 18:13:29 -00 2025 Feb 13 19:48:40.172824 kernel: KASLR disabled due to lack of seed Feb 13 19:48:40.172841 kernel: efi: EFI v2.7 by EDK II Feb 13 19:48:40.172857 kernel: efi: SMBIOS=0x7bed0000 SMBIOS 3.0=0x7beb0000 ACPI=0x786e0000 ACPI 2.0=0x786e0014 MEMATTR=0x7b003a98 MEMRESERVE=0x7852ee18 Feb 13 19:48:40.172872 kernel: ACPI: Early table checksum verification disabled Feb 13 19:48:40.172913 kernel: ACPI: RSDP 0x00000000786E0014 000024 (v02 AMAZON) Feb 13 19:48:40.172932 kernel: ACPI: XSDT 0x00000000786D00E8 000064 (v01 AMAZON AMZNFACP 00000001 01000013) Feb 13 19:48:40.172949 kernel: ACPI: FACP 0x00000000786B0000 000114 (v06 AMAZON AMZNFACP 00000001 AMZN 00000001) Feb 13 19:48:40.172965 kernel: ACPI: DSDT 0x0000000078640000 00159D (v02 AMAZON AMZNDSDT 00000001 INTL 20160527) Feb 13 19:48:40.172987 kernel: ACPI: APIC 0x00000000786C0000 000108 (v04 AMAZON AMZNAPIC 00000001 AMZN 00000001) Feb 13 19:48:40.173003 kernel: ACPI: SPCR 0x00000000786A0000 000050 (v02 AMAZON AMZNSPCR 00000001 AMZN 00000001) Feb 13 19:48:40.173018 kernel: ACPI: GTDT 0x0000000078690000 000060 (v02 AMAZON AMZNGTDT 00000001 AMZN 00000001) Feb 13 19:48:40.173034 kernel: ACPI: MCFG 0x0000000078680000 00003C (v02 AMAZON AMZNMCFG 00000001 AMZN 00000001) Feb 13 19:48:40.173053 kernel: ACPI: SLIT 0x0000000078670000 00002D (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Feb 13 19:48:40.173074 kernel: ACPI: IORT 0x0000000078660000 000078 (v01 AMAZON AMZNIORT 00000001 AMZN 00000001) Feb 13 19:48:40.173091 kernel: ACPI: PPTT 0x0000000078650000 0000EC (v01 AMAZON AMZNPPTT 00000001 AMZN 00000001) Feb 13 19:48:40.173108 kernel: ACPI: SPCR: console: uart,mmio,0x90a0000,115200 Feb 13 19:48:40.173124 kernel: earlycon: uart0 at MMIO 0x00000000090a0000 (options '115200') Feb 13 19:48:40.173140 kernel: printk: bootconsole [uart0] enabled Feb 13 19:48:40.173157 kernel: NUMA: Failed to initialise from firmware Feb 13 19:48:40.173174 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000004b5ffffff] Feb 13 19:48:40.173190 kernel: NUMA: NODE_DATA [mem 0x4b583f800-0x4b5844fff] Feb 13 19:48:40.173206 kernel: Zone ranges: Feb 13 19:48:40.173262 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Feb 13 19:48:40.173282 kernel: DMA32 empty Feb 13 19:48:40.173305 kernel: Normal [mem 0x0000000100000000-0x00000004b5ffffff] Feb 13 19:48:40.173322 kernel: Movable zone start for each node Feb 13 19:48:40.173338 kernel: Early memory node ranges Feb 13 19:48:40.173354 kernel: node 0: [mem 0x0000000040000000-0x000000007862ffff] Feb 13 19:48:40.173371 kernel: node 0: [mem 0x0000000078630000-0x000000007863ffff] Feb 13 19:48:40.173387 kernel: node 0: [mem 0x0000000078640000-0x00000000786effff] Feb 13 19:48:40.173403 kernel: node 0: [mem 0x00000000786f0000-0x000000007872ffff] Feb 13 19:48:40.173419 kernel: node 0: [mem 0x0000000078730000-0x000000007bbfffff] Feb 13 19:48:40.173435 kernel: node 0: [mem 0x000000007bc00000-0x000000007bfdffff] Feb 13 19:48:40.173452 kernel: node 0: [mem 0x000000007bfe0000-0x000000007fffffff] Feb 13 19:48:40.173468 kernel: node 0: [mem 0x0000000400000000-0x00000004b5ffffff] Feb 13 19:48:40.173485 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000004b5ffffff] Feb 13 19:48:40.173505 kernel: On node 0, zone Normal: 8192 pages in unavailable ranges Feb 13 19:48:40.173522 kernel: psci: probing for conduit method from ACPI. Feb 13 19:48:40.173546 kernel: psci: PSCIv1.0 detected in firmware. Feb 13 19:48:40.173564 kernel: psci: Using standard PSCI v0.2 function IDs Feb 13 19:48:40.173582 kernel: psci: Trusted OS migration not required Feb 13 19:48:40.173603 kernel: psci: SMC Calling Convention v1.1 Feb 13 19:48:40.173621 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Feb 13 19:48:40.173638 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Feb 13 19:48:40.173656 kernel: pcpu-alloc: [0] 0 [0] 1 Feb 13 19:48:40.173674 kernel: Detected PIPT I-cache on CPU0 Feb 13 19:48:40.173691 kernel: CPU features: detected: GIC system register CPU interface Feb 13 19:48:40.173709 kernel: CPU features: detected: Spectre-v2 Feb 13 19:48:40.173726 kernel: CPU features: detected: Spectre-v3a Feb 13 19:48:40.173744 kernel: CPU features: detected: Spectre-BHB Feb 13 19:48:40.173761 kernel: CPU features: detected: ARM erratum 1742098 Feb 13 19:48:40.173779 kernel: CPU features: detected: ARM errata 1165522, 1319367, or 1530923 Feb 13 19:48:40.173800 kernel: alternatives: applying boot alternatives Feb 13 19:48:40.173821 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=c15c751c06cfb933aa98417326b93d899c08a83ce060a940cd01082629c201a7 Feb 13 19:48:40.173840 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 19:48:40.173857 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 19:48:40.173875 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 19:48:40.173892 kernel: Fallback order for Node 0: 0 Feb 13 19:48:40.173910 kernel: Built 1 zonelists, mobility grouping on. Total pages: 991872 Feb 13 19:48:40.173927 kernel: Policy zone: Normal Feb 13 19:48:40.173944 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 19:48:40.173961 kernel: software IO TLB: area num 2. Feb 13 19:48:40.173979 kernel: software IO TLB: mapped [mem 0x000000007c000000-0x0000000080000000] (64MB) Feb 13 19:48:40.174002 kernel: Memory: 3820216K/4030464K available (10240K kernel code, 2186K rwdata, 8096K rodata, 39360K init, 897K bss, 210248K reserved, 0K cma-reserved) Feb 13 19:48:40.174020 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 13 19:48:40.174037 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 19:48:40.174055 kernel: rcu: RCU event tracing is enabled. Feb 13 19:48:40.174073 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 13 19:48:40.174091 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 19:48:40.174110 kernel: Tracing variant of Tasks RCU enabled. Feb 13 19:48:40.174128 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 19:48:40.174145 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 13 19:48:40.174162 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Feb 13 19:48:40.174180 kernel: GICv3: 96 SPIs implemented Feb 13 19:48:40.174201 kernel: GICv3: 0 Extended SPIs implemented Feb 13 19:48:40.175257 kernel: Root IRQ handler: gic_handle_irq Feb 13 19:48:40.175283 kernel: GICv3: GICv3 features: 16 PPIs Feb 13 19:48:40.175301 kernel: GICv3: CPU0: found redistributor 0 region 0:0x0000000010200000 Feb 13 19:48:40.175318 kernel: ITS [mem 0x10080000-0x1009ffff] Feb 13 19:48:40.175336 kernel: ITS@0x0000000010080000: allocated 8192 Devices @4000b0000 (indirect, esz 8, psz 64K, shr 1) Feb 13 19:48:40.175354 kernel: ITS@0x0000000010080000: allocated 8192 Interrupt Collections @4000c0000 (flat, esz 8, psz 64K, shr 1) Feb 13 19:48:40.175371 kernel: GICv3: using LPI property table @0x00000004000d0000 Feb 13 19:48:40.175388 kernel: ITS: Using hypervisor restricted LPI range [128] Feb 13 19:48:40.175405 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000004000e0000 Feb 13 19:48:40.175422 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 19:48:40.175440 kernel: arch_timer: cp15 timer(s) running at 83.33MHz (virt). Feb 13 19:48:40.175464 kernel: clocksource: arch_sys_counter: mask: 0x1ffffffffffffff max_cycles: 0x13381ebeec, max_idle_ns: 440795203145 ns Feb 13 19:48:40.175482 kernel: sched_clock: 57 bits at 83MHz, resolution 12ns, wraps every 4398046511100ns Feb 13 19:48:40.175499 kernel: Console: colour dummy device 80x25 Feb 13 19:48:40.175517 kernel: printk: console [tty1] enabled Feb 13 19:48:40.175534 kernel: ACPI: Core revision 20230628 Feb 13 19:48:40.175552 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 166.66 BogoMIPS (lpj=83333) Feb 13 19:48:40.175570 kernel: pid_max: default: 32768 minimum: 301 Feb 13 19:48:40.175588 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 19:48:40.175605 kernel: landlock: Up and running. Feb 13 19:48:40.175627 kernel: SELinux: Initializing. Feb 13 19:48:40.175645 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:48:40.175663 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:48:40.175681 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 19:48:40.175699 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 19:48:40.175716 kernel: rcu: Hierarchical SRCU implementation. Feb 13 19:48:40.175734 kernel: rcu: Max phase no-delay instances is 400. Feb 13 19:48:40.175752 kernel: Platform MSI: ITS@0x10080000 domain created Feb 13 19:48:40.175770 kernel: PCI/MSI: ITS@0x10080000 domain created Feb 13 19:48:40.175791 kernel: Remapping and enabling EFI services. Feb 13 19:48:40.175809 kernel: smp: Bringing up secondary CPUs ... Feb 13 19:48:40.175826 kernel: Detected PIPT I-cache on CPU1 Feb 13 19:48:40.175844 kernel: GICv3: CPU1: found redistributor 1 region 0:0x0000000010220000 Feb 13 19:48:40.175862 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000004000f0000 Feb 13 19:48:40.175879 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd083] Feb 13 19:48:40.175897 kernel: smp: Brought up 1 node, 2 CPUs Feb 13 19:48:40.175914 kernel: SMP: Total of 2 processors activated. Feb 13 19:48:40.175932 kernel: CPU features: detected: 32-bit EL0 Support Feb 13 19:48:40.175953 kernel: CPU features: detected: 32-bit EL1 Support Feb 13 19:48:40.175971 kernel: CPU features: detected: CRC32 instructions Feb 13 19:48:40.175989 kernel: CPU: All CPU(s) started at EL1 Feb 13 19:48:40.176018 kernel: alternatives: applying system-wide alternatives Feb 13 19:48:40.176041 kernel: devtmpfs: initialized Feb 13 19:48:40.176059 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 19:48:40.176078 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 13 19:48:40.176096 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 19:48:40.176114 kernel: SMBIOS 3.0.0 present. Feb 13 19:48:40.176132 kernel: DMI: Amazon EC2 a1.large/, BIOS 1.0 11/1/2018 Feb 13 19:48:40.176155 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 19:48:40.176173 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Feb 13 19:48:40.176192 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Feb 13 19:48:40.176210 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Feb 13 19:48:40.177308 kernel: audit: initializing netlink subsys (disabled) Feb 13 19:48:40.177330 kernel: audit: type=2000 audit(0.287:1): state=initialized audit_enabled=0 res=1 Feb 13 19:48:40.177350 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 19:48:40.177378 kernel: cpuidle: using governor menu Feb 13 19:48:40.177397 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Feb 13 19:48:40.177417 kernel: ASID allocator initialised with 65536 entries Feb 13 19:48:40.177437 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 19:48:40.177456 kernel: Serial: AMBA PL011 UART driver Feb 13 19:48:40.177475 kernel: Modules: 17520 pages in range for non-PLT usage Feb 13 19:48:40.177495 kernel: Modules: 509040 pages in range for PLT usage Feb 13 19:48:40.177516 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 19:48:40.177535 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 19:48:40.177560 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Feb 13 19:48:40.177581 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Feb 13 19:48:40.177599 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 19:48:40.177620 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 19:48:40.177640 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Feb 13 19:48:40.177660 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Feb 13 19:48:40.177678 kernel: ACPI: Added _OSI(Module Device) Feb 13 19:48:40.177697 kernel: ACPI: Added _OSI(Processor Device) Feb 13 19:48:40.177716 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 19:48:40.177741 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 19:48:40.177763 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 19:48:40.177782 kernel: ACPI: Interpreter enabled Feb 13 19:48:40.177800 kernel: ACPI: Using GIC for interrupt routing Feb 13 19:48:40.177820 kernel: ACPI: MCFG table detected, 1 entries Feb 13 19:48:40.177841 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-0f]) Feb 13 19:48:40.178202 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Feb 13 19:48:40.178563 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Feb 13 19:48:40.178772 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Feb 13 19:48:40.178976 kernel: acpi PNP0A08:00: ECAM area [mem 0x20000000-0x20ffffff] reserved by PNP0C02:00 Feb 13 19:48:40.179178 kernel: acpi PNP0A08:00: ECAM at [mem 0x20000000-0x20ffffff] for [bus 00-0f] Feb 13 19:48:40.179204 kernel: ACPI: Remapped I/O 0x000000001fff0000 to [io 0x0000-0xffff window] Feb 13 19:48:40.180342 kernel: acpiphp: Slot [1] registered Feb 13 19:48:40.180376 kernel: acpiphp: Slot [2] registered Feb 13 19:48:40.180395 kernel: acpiphp: Slot [3] registered Feb 13 19:48:40.180413 kernel: acpiphp: Slot [4] registered Feb 13 19:48:40.180443 kernel: acpiphp: Slot [5] registered Feb 13 19:48:40.180462 kernel: acpiphp: Slot [6] registered Feb 13 19:48:40.180480 kernel: acpiphp: Slot [7] registered Feb 13 19:48:40.180499 kernel: acpiphp: Slot [8] registered Feb 13 19:48:40.180517 kernel: acpiphp: Slot [9] registered Feb 13 19:48:40.180536 kernel: acpiphp: Slot [10] registered Feb 13 19:48:40.180554 kernel: acpiphp: Slot [11] registered Feb 13 19:48:40.180573 kernel: acpiphp: Slot [12] registered Feb 13 19:48:40.180592 kernel: acpiphp: Slot [13] registered Feb 13 19:48:40.180610 kernel: acpiphp: Slot [14] registered Feb 13 19:48:40.180634 kernel: acpiphp: Slot [15] registered Feb 13 19:48:40.180652 kernel: acpiphp: Slot [16] registered Feb 13 19:48:40.180671 kernel: acpiphp: Slot [17] registered Feb 13 19:48:40.180689 kernel: acpiphp: Slot [18] registered Feb 13 19:48:40.180707 kernel: acpiphp: Slot [19] registered Feb 13 19:48:40.180726 kernel: acpiphp: Slot [20] registered Feb 13 19:48:40.180744 kernel: acpiphp: Slot [21] registered Feb 13 19:48:40.180763 kernel: acpiphp: Slot [22] registered Feb 13 19:48:40.180781 kernel: acpiphp: Slot [23] registered Feb 13 19:48:40.180804 kernel: acpiphp: Slot [24] registered Feb 13 19:48:40.180824 kernel: acpiphp: Slot [25] registered Feb 13 19:48:40.180844 kernel: acpiphp: Slot [26] registered Feb 13 19:48:40.180863 kernel: acpiphp: Slot [27] registered Feb 13 19:48:40.180898 kernel: acpiphp: Slot [28] registered Feb 13 19:48:40.180923 kernel: acpiphp: Slot [29] registered Feb 13 19:48:40.180942 kernel: acpiphp: Slot [30] registered Feb 13 19:48:40.180961 kernel: acpiphp: Slot [31] registered Feb 13 19:48:40.180979 kernel: PCI host bridge to bus 0000:00 Feb 13 19:48:40.181269 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xffffffff window] Feb 13 19:48:40.181470 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Feb 13 19:48:40.181652 kernel: pci_bus 0000:00: root bus resource [mem 0x400000000000-0x407fffffffff window] Feb 13 19:48:40.181832 kernel: pci_bus 0000:00: root bus resource [bus 00-0f] Feb 13 19:48:40.182062 kernel: pci 0000:00:00.0: [1d0f:0200] type 00 class 0x060000 Feb 13 19:48:40.184391 kernel: pci 0000:00:01.0: [1d0f:8250] type 00 class 0x070003 Feb 13 19:48:40.184657 kernel: pci 0000:00:01.0: reg 0x10: [mem 0x80118000-0x80118fff] Feb 13 19:48:40.184913 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Feb 13 19:48:40.185131 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80114000-0x80117fff] Feb 13 19:48:40.185400 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D2 D3hot D3cold Feb 13 19:48:40.185633 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Feb 13 19:48:40.185848 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80110000-0x80113fff] Feb 13 19:48:40.186054 kernel: pci 0000:00:05.0: reg 0x18: [mem 0x80000000-0x800fffff pref] Feb 13 19:48:40.186350 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x80100000-0x8010ffff] Feb 13 19:48:40.186562 kernel: pci 0000:00:05.0: PME# supported from D0 D1 D2 D3hot D3cold Feb 13 19:48:40.186766 kernel: pci 0000:00:05.0: BAR 2: assigned [mem 0x80000000-0x800fffff pref] Feb 13 19:48:40.186978 kernel: pci 0000:00:05.0: BAR 4: assigned [mem 0x80100000-0x8010ffff] Feb 13 19:48:40.187190 kernel: pci 0000:00:04.0: BAR 0: assigned [mem 0x80110000-0x80113fff] Feb 13 19:48:40.188510 kernel: pci 0000:00:05.0: BAR 0: assigned [mem 0x80114000-0x80117fff] Feb 13 19:48:40.188728 kernel: pci 0000:00:01.0: BAR 0: assigned [mem 0x80118000-0x80118fff] Feb 13 19:48:40.188947 kernel: pci_bus 0000:00: resource 4 [mem 0x80000000-0xffffffff window] Feb 13 19:48:40.189131 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Feb 13 19:48:40.189628 kernel: pci_bus 0000:00: resource 6 [mem 0x400000000000-0x407fffffffff window] Feb 13 19:48:40.189658 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Feb 13 19:48:40.189678 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Feb 13 19:48:40.189697 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Feb 13 19:48:40.189716 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Feb 13 19:48:40.189735 kernel: iommu: Default domain type: Translated Feb 13 19:48:40.189753 kernel: iommu: DMA domain TLB invalidation policy: strict mode Feb 13 19:48:40.189780 kernel: efivars: Registered efivars operations Feb 13 19:48:40.189798 kernel: vgaarb: loaded Feb 13 19:48:40.189817 kernel: clocksource: Switched to clocksource arch_sys_counter Feb 13 19:48:40.189835 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 19:48:40.189854 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 19:48:40.189872 kernel: pnp: PnP ACPI init Feb 13 19:48:40.190084 kernel: system 00:00: [mem 0x20000000-0x2fffffff] could not be reserved Feb 13 19:48:40.190112 kernel: pnp: PnP ACPI: found 1 devices Feb 13 19:48:40.190136 kernel: NET: Registered PF_INET protocol family Feb 13 19:48:40.190155 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 19:48:40.190174 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Feb 13 19:48:40.190193 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 19:48:40.190212 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 19:48:40.191728 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Feb 13 19:48:40.191750 kernel: TCP: Hash tables configured (established 32768 bind 32768) Feb 13 19:48:40.191769 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:48:40.191788 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:48:40.191816 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 19:48:40.191835 kernel: PCI: CLS 0 bytes, default 64 Feb 13 19:48:40.191853 kernel: kvm [1]: HYP mode not available Feb 13 19:48:40.191872 kernel: Initialise system trusted keyrings Feb 13 19:48:40.191890 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Feb 13 19:48:40.191909 kernel: Key type asymmetric registered Feb 13 19:48:40.191927 kernel: Asymmetric key parser 'x509' registered Feb 13 19:48:40.191945 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Feb 13 19:48:40.191964 kernel: io scheduler mq-deadline registered Feb 13 19:48:40.191987 kernel: io scheduler kyber registered Feb 13 19:48:40.192005 kernel: io scheduler bfq registered Feb 13 19:48:40.192307 kernel: pl061_gpio ARMH0061:00: PL061 GPIO chip registered Feb 13 19:48:40.192338 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Feb 13 19:48:40.192357 kernel: ACPI: button: Power Button [PWRB] Feb 13 19:48:40.192376 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0E:00/input/input1 Feb 13 19:48:40.192395 kernel: ACPI: button: Sleep Button [SLPB] Feb 13 19:48:40.192414 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 19:48:40.192439 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Feb 13 19:48:40.192650 kernel: serial 0000:00:01.0: enabling device (0010 -> 0012) Feb 13 19:48:40.192676 kernel: printk: console [ttyS0] disabled Feb 13 19:48:40.192696 kernel: 0000:00:01.0: ttyS0 at MMIO 0x80118000 (irq = 14, base_baud = 115200) is a 16550A Feb 13 19:48:40.192714 kernel: printk: console [ttyS0] enabled Feb 13 19:48:40.192733 kernel: printk: bootconsole [uart0] disabled Feb 13 19:48:40.192751 kernel: thunder_xcv, ver 1.0 Feb 13 19:48:40.192770 kernel: thunder_bgx, ver 1.0 Feb 13 19:48:40.192788 kernel: nicpf, ver 1.0 Feb 13 19:48:40.192811 kernel: nicvf, ver 1.0 Feb 13 19:48:40.193051 kernel: rtc-efi rtc-efi.0: registered as rtc0 Feb 13 19:48:40.193614 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-02-13T19:48:39 UTC (1739476119) Feb 13 19:48:40.193650 kernel: hid: raw HID events driver (C) Jiri Kosina Feb 13 19:48:40.193670 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 3 counters available Feb 13 19:48:40.193695 kernel: watchdog: Delayed init of the lockup detector failed: -19 Feb 13 19:48:40.193714 kernel: watchdog: Hard watchdog permanently disabled Feb 13 19:48:40.193732 kernel: NET: Registered PF_INET6 protocol family Feb 13 19:48:40.193759 kernel: Segment Routing with IPv6 Feb 13 19:48:40.193778 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 19:48:40.193796 kernel: NET: Registered PF_PACKET protocol family Feb 13 19:48:40.193815 kernel: Key type dns_resolver registered Feb 13 19:48:40.193833 kernel: registered taskstats version 1 Feb 13 19:48:40.193852 kernel: Loading compiled-in X.509 certificates Feb 13 19:48:40.193871 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 8bd805622262697b24b0fa7c407ae82c4289ceec' Feb 13 19:48:40.193890 kernel: Key type .fscrypt registered Feb 13 19:48:40.193909 kernel: Key type fscrypt-provisioning registered Feb 13 19:48:40.193934 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 19:48:40.193954 kernel: ima: Allocated hash algorithm: sha1 Feb 13 19:48:40.193976 kernel: ima: No architecture policies found Feb 13 19:48:40.193998 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Feb 13 19:48:40.194016 kernel: clk: Disabling unused clocks Feb 13 19:48:40.194035 kernel: Freeing unused kernel memory: 39360K Feb 13 19:48:40.194053 kernel: Run /init as init process Feb 13 19:48:40.194071 kernel: with arguments: Feb 13 19:48:40.194089 kernel: /init Feb 13 19:48:40.194107 kernel: with environment: Feb 13 19:48:40.194130 kernel: HOME=/ Feb 13 19:48:40.194148 kernel: TERM=linux Feb 13 19:48:40.194166 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 19:48:40.194194 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 19:48:40.194381 systemd[1]: Detected virtualization amazon. Feb 13 19:48:40.194409 systemd[1]: Detected architecture arm64. Feb 13 19:48:40.194430 systemd[1]: Running in initrd. Feb 13 19:48:40.194456 systemd[1]: No hostname configured, using default hostname. Feb 13 19:48:40.194477 systemd[1]: Hostname set to . Feb 13 19:48:40.194498 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:48:40.194518 systemd[1]: Queued start job for default target initrd.target. Feb 13 19:48:40.194539 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:48:40.194559 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:48:40.194580 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 19:48:40.194601 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:48:40.194626 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 19:48:40.194647 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 19:48:40.194671 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 19:48:40.194692 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 19:48:40.194713 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:48:40.194733 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:48:40.194754 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:48:40.194779 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:48:40.194799 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:48:40.194819 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:48:40.194840 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:48:40.194860 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:48:40.194881 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 19:48:40.194901 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Feb 13 19:48:40.194921 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:48:40.194941 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:48:40.194966 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:48:40.194986 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:48:40.195007 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 19:48:40.195027 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:48:40.195047 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 19:48:40.195067 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 19:48:40.195087 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:48:40.195108 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:48:40.195132 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:48:40.195153 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 19:48:40.195173 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:48:40.195273 systemd-journald[251]: Collecting audit messages is disabled. Feb 13 19:48:40.195326 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 19:48:40.195349 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:48:40.195370 systemd-journald[251]: Journal started Feb 13 19:48:40.195412 systemd-journald[251]: Runtime Journal (/run/log/journal/ec2df6ce9d9f50082cb28b4722e627dd) is 8.0M, max 75.3M, 67.3M free. Feb 13 19:48:40.167541 systemd-modules-load[252]: Inserted module 'overlay' Feb 13 19:48:40.202266 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:48:40.207270 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 19:48:40.210285 kernel: Bridge firewalling registered Feb 13 19:48:40.210384 systemd-modules-load[252]: Inserted module 'br_netfilter' Feb 13 19:48:40.212161 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:48:40.215775 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:48:40.225083 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:48:40.243724 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:48:40.250079 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:48:40.258440 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:48:40.263433 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:48:40.297712 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:48:40.305693 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:48:40.323519 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:48:40.337585 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:48:40.346732 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:48:40.364486 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 19:48:40.396264 dracut-cmdline[290]: dracut-dracut-053 Feb 13 19:48:40.401670 dracut-cmdline[290]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyS0,115200n8 earlycon flatcar.first_boot=detected acpi=force flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=c15c751c06cfb933aa98417326b93d899c08a83ce060a940cd01082629c201a7 Feb 13 19:48:40.427523 systemd-resolved[287]: Positive Trust Anchors: Feb 13 19:48:40.427559 systemd-resolved[287]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:48:40.427624 systemd-resolved[287]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:48:40.545247 kernel: SCSI subsystem initialized Feb 13 19:48:40.551258 kernel: Loading iSCSI transport class v2.0-870. Feb 13 19:48:40.563253 kernel: iscsi: registered transport (tcp) Feb 13 19:48:40.585517 kernel: iscsi: registered transport (qla4xxx) Feb 13 19:48:40.585589 kernel: QLogic iSCSI HBA Driver Feb 13 19:48:40.665255 kernel: random: crng init done Feb 13 19:48:40.665503 systemd-resolved[287]: Defaulting to hostname 'linux'. Feb 13 19:48:40.669022 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:48:40.673022 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:48:40.694971 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 19:48:40.703513 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 19:48:40.749265 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 19:48:40.749340 kernel: device-mapper: uevent: version 1.0.3 Feb 13 19:48:40.749366 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 19:48:40.815260 kernel: raid6: neonx8 gen() 6785 MB/s Feb 13 19:48:40.832249 kernel: raid6: neonx4 gen() 6616 MB/s Feb 13 19:48:40.849248 kernel: raid6: neonx2 gen() 5477 MB/s Feb 13 19:48:40.866248 kernel: raid6: neonx1 gen() 3959 MB/s Feb 13 19:48:40.883248 kernel: raid6: int64x8 gen() 3826 MB/s Feb 13 19:48:40.900248 kernel: raid6: int64x4 gen() 3728 MB/s Feb 13 19:48:40.917248 kernel: raid6: int64x2 gen() 3616 MB/s Feb 13 19:48:40.934992 kernel: raid6: int64x1 gen() 2756 MB/s Feb 13 19:48:40.935025 kernel: raid6: using algorithm neonx8 gen() 6785 MB/s Feb 13 19:48:40.952982 kernel: raid6: .... xor() 4746 MB/s, rmw enabled Feb 13 19:48:40.953026 kernel: raid6: using neon recovery algorithm Feb 13 19:48:40.961364 kernel: xor: measuring software checksum speed Feb 13 19:48:40.961422 kernel: 8regs : 10970 MB/sec Feb 13 19:48:40.962445 kernel: 32regs : 11943 MB/sec Feb 13 19:48:40.963599 kernel: arm64_neon : 9578 MB/sec Feb 13 19:48:40.963631 kernel: xor: using function: 32regs (11943 MB/sec) Feb 13 19:48:41.047261 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 19:48:41.066178 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:48:41.075557 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:48:41.115154 systemd-udevd[471]: Using default interface naming scheme 'v255'. Feb 13 19:48:41.123858 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:48:41.145622 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 19:48:41.175651 dracut-pre-trigger[482]: rd.md=0: removing MD RAID activation Feb 13 19:48:41.231290 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:48:41.243506 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:48:41.362831 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:48:41.373863 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 19:48:41.420536 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 19:48:41.425670 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:48:41.430390 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:48:41.446325 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:48:41.452064 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 19:48:41.504311 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:48:41.558691 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Feb 13 19:48:41.558763 kernel: ena 0000:00:05.0: enabling device (0010 -> 0012) Feb 13 19:48:41.582489 kernel: ena 0000:00:05.0: ENA device version: 0.10 Feb 13 19:48:41.582743 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Feb 13 19:48:41.582993 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80114000, mac addr 06:9e:4f:2e:9e:83 Feb 13 19:48:41.567781 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:48:41.568232 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:48:41.570887 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:48:41.573010 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:48:41.573313 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:48:41.575526 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:48:41.583812 (udev-worker)[532]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:48:41.588775 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:48:41.635257 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Feb 13 19:48:41.637274 kernel: nvme nvme0: pci function 0000:00:04.0 Feb 13 19:48:41.646271 kernel: nvme nvme0: 2/0/0 default/read/poll queues Feb 13 19:48:41.653767 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 19:48:41.653826 kernel: GPT:9289727 != 16777215 Feb 13 19:48:41.653851 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 19:48:41.653876 kernel: GPT:9289727 != 16777215 Feb 13 19:48:41.653900 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 19:48:41.653924 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:48:41.658697 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:48:41.667538 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:48:41.714448 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:48:41.785529 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 scanned by (udev-worker) (518) Feb 13 19:48:41.802273 kernel: BTRFS: device fsid 4bb2b262-8ef2-48e3-80f4-24f9d7a85bf6 devid 1 transid 40 /dev/nvme0n1p3 scanned by (udev-worker) (520) Feb 13 19:48:41.861346 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Feb 13 19:48:41.904846 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Feb 13 19:48:41.922648 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 19:48:41.936429 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Feb 13 19:48:41.940074 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Feb 13 19:48:41.979885 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 19:48:41.992190 disk-uuid[661]: Primary Header is updated. Feb 13 19:48:41.992190 disk-uuid[661]: Secondary Entries is updated. Feb 13 19:48:41.992190 disk-uuid[661]: Secondary Header is updated. Feb 13 19:48:42.007605 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:48:42.013268 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:48:42.021262 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:48:43.023282 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 19:48:43.024500 disk-uuid[662]: The operation has completed successfully. Feb 13 19:48:43.201462 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 19:48:43.203617 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 19:48:43.250599 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 19:48:43.262410 sh[1004]: Success Feb 13 19:48:43.290248 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Feb 13 19:48:43.399859 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 19:48:43.412460 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 19:48:43.423290 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 19:48:43.458164 kernel: BTRFS info (device dm-0): first mount of filesystem 4bb2b262-8ef2-48e3-80f4-24f9d7a85bf6 Feb 13 19:48:43.458267 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:48:43.459963 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 19:48:43.461237 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 19:48:43.462343 kernel: BTRFS info (device dm-0): using free space tree Feb 13 19:48:43.570247 kernel: BTRFS info (device dm-0): enabling ssd optimizations Feb 13 19:48:43.597837 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 19:48:43.601447 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 19:48:43.611515 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 19:48:43.616976 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 19:48:43.642575 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 896fb6d3-4143-43a6-a44b-ca1ce10817e1 Feb 13 19:48:43.642658 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:48:43.642692 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 19:48:43.650273 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 19:48:43.670766 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 19:48:43.672968 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 896fb6d3-4143-43a6-a44b-ca1ce10817e1 Feb 13 19:48:43.684769 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 19:48:43.695611 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 19:48:43.799838 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:48:43.811558 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:48:43.869193 systemd-networkd[1208]: lo: Link UP Feb 13 19:48:43.870761 systemd-networkd[1208]: lo: Gained carrier Feb 13 19:48:43.874645 systemd-networkd[1208]: Enumeration completed Feb 13 19:48:43.875869 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:48:43.880478 systemd[1]: Reached target network.target - Network. Feb 13 19:48:43.883919 systemd-networkd[1208]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:48:43.883940 systemd-networkd[1208]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:48:43.891945 systemd-networkd[1208]: eth0: Link UP Feb 13 19:48:43.891960 systemd-networkd[1208]: eth0: Gained carrier Feb 13 19:48:43.891977 systemd-networkd[1208]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:48:43.921303 systemd-networkd[1208]: eth0: DHCPv4 address 172.31.25.134/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 19:48:44.104095 ignition[1120]: Ignition 2.19.0 Feb 13 19:48:44.104641 ignition[1120]: Stage: fetch-offline Feb 13 19:48:44.105183 ignition[1120]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:48:44.105207 ignition[1120]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:48:44.105730 ignition[1120]: Ignition finished successfully Feb 13 19:48:44.115321 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:48:44.123707 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Feb 13 19:48:44.158439 ignition[1218]: Ignition 2.19.0 Feb 13 19:48:44.158465 ignition[1218]: Stage: fetch Feb 13 19:48:44.159658 ignition[1218]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:48:44.159683 ignition[1218]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:48:44.159836 ignition[1218]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:48:44.179290 ignition[1218]: PUT result: OK Feb 13 19:48:44.182426 ignition[1218]: parsed url from cmdline: "" Feb 13 19:48:44.182448 ignition[1218]: no config URL provided Feb 13 19:48:44.182464 ignition[1218]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 19:48:44.182517 ignition[1218]: no config at "/usr/lib/ignition/user.ign" Feb 13 19:48:44.182550 ignition[1218]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:48:44.186017 ignition[1218]: PUT result: OK Feb 13 19:48:44.195723 unknown[1218]: fetched base config from "system" Feb 13 19:48:44.186089 ignition[1218]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Feb 13 19:48:44.195739 unknown[1218]: fetched base config from "system" Feb 13 19:48:44.188066 ignition[1218]: GET result: OK Feb 13 19:48:44.195753 unknown[1218]: fetched user config from "aws" Feb 13 19:48:44.188211 ignition[1218]: parsing config with SHA512: 0aeaaec2af808d4aeaa3fffeb6dab55536a4bb5ec78f36f1c9ee76429cb0ee98302322bc825a7b851978b832835d3ef4df0110f7d8d4a89b629f6ab57b1ddeb1 Feb 13 19:48:44.200027 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Feb 13 19:48:44.196385 ignition[1218]: fetch: fetch complete Feb 13 19:48:44.221501 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 19:48:44.196397 ignition[1218]: fetch: fetch passed Feb 13 19:48:44.196474 ignition[1218]: Ignition finished successfully Feb 13 19:48:44.252588 ignition[1224]: Ignition 2.19.0 Feb 13 19:48:44.253099 ignition[1224]: Stage: kargs Feb 13 19:48:44.253752 ignition[1224]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:48:44.253777 ignition[1224]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:48:44.253959 ignition[1224]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:48:44.258417 ignition[1224]: PUT result: OK Feb 13 19:48:44.268925 ignition[1224]: kargs: kargs passed Feb 13 19:48:44.269255 ignition[1224]: Ignition finished successfully Feb 13 19:48:44.274384 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 19:48:44.288624 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 19:48:44.311792 ignition[1230]: Ignition 2.19.0 Feb 13 19:48:44.311815 ignition[1230]: Stage: disks Feb 13 19:48:44.313119 ignition[1230]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:48:44.313146 ignition[1230]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:48:44.313339 ignition[1230]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:48:44.316371 ignition[1230]: PUT result: OK Feb 13 19:48:44.323111 ignition[1230]: disks: disks passed Feb 13 19:48:44.323204 ignition[1230]: Ignition finished successfully Feb 13 19:48:44.331339 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 19:48:44.335412 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 19:48:44.337497 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 19:48:44.339735 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:48:44.341587 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:48:44.343480 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:48:44.360572 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 19:48:44.399681 systemd-fsck[1238]: ROOT: clean, 14/553520 files, 52654/553472 blocks Feb 13 19:48:44.403874 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 19:48:44.417733 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 19:48:44.500248 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 9957d679-c6c4-49f4-b1b2-c3c1f3ba5699 r/w with ordered data mode. Quota mode: none. Feb 13 19:48:44.501071 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 19:48:44.504649 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 19:48:44.530358 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:48:44.537414 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 19:48:44.540754 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Feb 13 19:48:44.540833 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 19:48:44.541958 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:48:44.559150 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/nvme0n1p6 scanned by mount (1257) Feb 13 19:48:44.563578 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 896fb6d3-4143-43a6-a44b-ca1ce10817e1 Feb 13 19:48:44.563636 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:48:44.563663 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 19:48:44.572578 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 19:48:44.584826 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 19:48:44.580686 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 19:48:44.596389 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:48:44.942837 initrd-setup-root[1281]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 19:48:44.963860 initrd-setup-root[1288]: cut: /sysroot/etc/group: No such file or directory Feb 13 19:48:44.983684 initrd-setup-root[1295]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 19:48:44.991908 initrd-setup-root[1302]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 19:48:45.286788 systemd-networkd[1208]: eth0: Gained IPv6LL Feb 13 19:48:45.294082 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 19:48:45.301391 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 19:48:45.305821 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 19:48:45.330180 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 19:48:45.336265 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 896fb6d3-4143-43a6-a44b-ca1ce10817e1 Feb 13 19:48:45.363065 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 19:48:45.377576 ignition[1371]: INFO : Ignition 2.19.0 Feb 13 19:48:45.377576 ignition[1371]: INFO : Stage: mount Feb 13 19:48:45.381688 ignition[1371]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:48:45.381688 ignition[1371]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:48:45.381688 ignition[1371]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:48:45.381688 ignition[1371]: INFO : PUT result: OK Feb 13 19:48:45.392255 ignition[1371]: INFO : mount: mount passed Feb 13 19:48:45.392255 ignition[1371]: INFO : Ignition finished successfully Feb 13 19:48:45.394354 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 19:48:45.407456 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 19:48:45.516548 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:48:45.537256 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by mount (1382) Feb 13 19:48:45.540943 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 896fb6d3-4143-43a6-a44b-ca1ce10817e1 Feb 13 19:48:45.540997 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:48:45.541025 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 19:48:45.547250 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 19:48:45.551063 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:48:45.592469 ignition[1399]: INFO : Ignition 2.19.0 Feb 13 19:48:45.592469 ignition[1399]: INFO : Stage: files Feb 13 19:48:45.595658 ignition[1399]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:48:45.595658 ignition[1399]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:48:45.599707 ignition[1399]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:48:45.602742 ignition[1399]: INFO : PUT result: OK Feb 13 19:48:45.607028 ignition[1399]: DEBUG : files: compiled without relabeling support, skipping Feb 13 19:48:45.610211 ignition[1399]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 19:48:45.610211 ignition[1399]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 19:48:45.638077 ignition[1399]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 19:48:45.640647 ignition[1399]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 19:48:45.640647 ignition[1399]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 19:48:45.638968 unknown[1399]: wrote ssh authorized keys file for user: core Feb 13 19:48:45.654302 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Feb 13 19:48:45.661127 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 Feb 13 19:48:45.745423 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 19:48:45.956339 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Feb 13 19:48:45.959935 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 19:48:45.963502 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 19:48:45.966718 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:48:45.970096 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:48:45.973368 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:48:45.973368 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:48:45.979846 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:48:45.983051 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:48:45.986521 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:48:45.989971 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:48:45.993233 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:48:45.997928 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:48:46.002484 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:48:46.006719 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-arm64.raw: attempt #1 Feb 13 19:48:46.329258 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 13 19:48:46.676902 ignition[1399]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:48:46.676902 ignition[1399]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Feb 13 19:48:46.684344 ignition[1399]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:48:46.684344 ignition[1399]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:48:46.684344 ignition[1399]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Feb 13 19:48:46.684344 ignition[1399]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Feb 13 19:48:46.684344 ignition[1399]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 19:48:46.684344 ignition[1399]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:48:46.702315 ignition[1399]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:48:46.702315 ignition[1399]: INFO : files: files passed Feb 13 19:48:46.702315 ignition[1399]: INFO : Ignition finished successfully Feb 13 19:48:46.710305 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 19:48:46.719524 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 19:48:46.729582 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 19:48:46.741925 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 19:48:46.742170 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 19:48:46.758141 initrd-setup-root-after-ignition[1427]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:48:46.758141 initrd-setup-root-after-ignition[1427]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:48:46.765441 initrd-setup-root-after-ignition[1431]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:48:46.770450 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:48:46.775582 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 19:48:46.791453 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 19:48:46.842606 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 19:48:46.844419 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 19:48:46.847474 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 19:48:46.849592 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 19:48:46.853937 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 19:48:46.866516 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 19:48:46.903265 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:48:46.923640 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 19:48:46.948963 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:48:46.951904 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:48:46.956780 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 19:48:46.961974 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 19:48:46.962207 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:48:46.965074 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 19:48:46.973548 systemd[1]: Stopped target basic.target - Basic System. Feb 13 19:48:46.975862 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 19:48:46.981174 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:48:46.983483 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 19:48:46.986035 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 19:48:46.993294 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:48:46.996477 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 19:48:47.000268 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 19:48:47.003258 systemd[1]: Stopped target swap.target - Swaps. Feb 13 19:48:47.005816 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 19:48:47.006053 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:48:47.014492 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:48:47.016654 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:48:47.018952 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 19:48:47.020878 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:48:47.023362 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 19:48:47.023996 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 19:48:47.027966 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 19:48:47.028212 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:48:47.041648 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 19:48:47.041856 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 19:48:47.059700 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 19:48:47.066585 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 19:48:47.068355 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 19:48:47.068631 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:48:47.071869 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 19:48:47.072107 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:48:47.103246 ignition[1451]: INFO : Ignition 2.19.0 Feb 13 19:48:47.103246 ignition[1451]: INFO : Stage: umount Feb 13 19:48:47.103246 ignition[1451]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:48:47.103246 ignition[1451]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 19:48:47.103246 ignition[1451]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 19:48:47.102465 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 19:48:47.105016 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 19:48:47.121030 ignition[1451]: INFO : PUT result: OK Feb 13 19:48:47.126833 ignition[1451]: INFO : umount: umount passed Feb 13 19:48:47.129392 ignition[1451]: INFO : Ignition finished successfully Feb 13 19:48:47.132856 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 19:48:47.135004 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 19:48:47.140751 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 19:48:47.143165 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 19:48:47.143359 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 19:48:47.150632 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 19:48:47.150754 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 19:48:47.161192 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 13 19:48:47.161324 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Feb 13 19:48:47.168535 systemd[1]: Stopped target network.target - Network. Feb 13 19:48:47.170243 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 19:48:47.170336 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:48:47.172651 systemd[1]: Stopped target paths.target - Path Units. Feb 13 19:48:47.174310 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 19:48:47.182049 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:48:47.192408 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 19:48:47.194313 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 19:48:47.196461 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 19:48:47.196538 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:48:47.198450 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 19:48:47.198516 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:48:47.200438 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 19:48:47.200519 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 19:48:47.202460 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 19:48:47.202536 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 19:48:47.204936 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 19:48:47.207729 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 19:48:47.211409 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 19:48:47.211581 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 19:48:47.215534 systemd-networkd[1208]: eth0: DHCPv6 lease lost Feb 13 19:48:47.233363 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 19:48:47.233574 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 19:48:47.242189 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 19:48:47.243485 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 19:48:47.250401 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 19:48:47.250510 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:48:47.252795 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 19:48:47.252931 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 19:48:47.264361 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 19:48:47.277747 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 19:48:47.277852 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:48:47.278507 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 19:48:47.278583 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:48:47.279183 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 19:48:47.279280 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 19:48:47.279905 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 19:48:47.279976 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:48:47.280780 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:48:47.342846 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 19:48:47.344759 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:48:47.351047 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 19:48:47.351672 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 19:48:47.357152 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 19:48:47.357901 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:48:47.364070 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 19:48:47.364175 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:48:47.366471 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 19:48:47.366553 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 19:48:47.375403 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:48:47.375497 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:48:47.395594 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 19:48:47.397944 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 19:48:47.398054 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:48:47.400415 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Feb 13 19:48:47.400497 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:48:47.402971 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 19:48:47.403055 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:48:47.411438 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:48:47.411548 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:48:47.415167 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 19:48:47.415429 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 19:48:47.423881 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 19:48:47.424405 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 19:48:47.429138 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 19:48:47.451485 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 19:48:47.474991 systemd[1]: Switching root. Feb 13 19:48:47.524011 systemd-journald[251]: Journal stopped Feb 13 19:48:50.159275 systemd-journald[251]: Received SIGTERM from PID 1 (systemd). Feb 13 19:48:50.159409 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 19:48:50.159450 kernel: SELinux: policy capability open_perms=1 Feb 13 19:48:50.159482 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 19:48:50.159512 kernel: SELinux: policy capability always_check_network=0 Feb 13 19:48:50.159542 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 19:48:50.159574 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 19:48:50.159610 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 19:48:50.159642 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 19:48:50.159672 kernel: audit: type=1403 audit(1739476128.201:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 19:48:50.159712 systemd[1]: Successfully loaded SELinux policy in 74.911ms. Feb 13 19:48:50.159764 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 23.024ms. Feb 13 19:48:50.159800 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 19:48:50.159833 systemd[1]: Detected virtualization amazon. Feb 13 19:48:50.159864 systemd[1]: Detected architecture arm64. Feb 13 19:48:50.159895 systemd[1]: Detected first boot. Feb 13 19:48:50.159932 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:48:50.159964 zram_generator::config[1493]: No configuration found. Feb 13 19:48:50.159999 systemd[1]: Populated /etc with preset unit settings. Feb 13 19:48:50.160031 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 19:48:50.160061 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 19:48:50.160094 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 19:48:50.160127 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 19:48:50.160160 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 19:48:50.160194 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 19:48:50.160242 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 19:48:50.160281 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 19:48:50.160314 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 19:48:50.160347 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 19:48:50.160377 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 19:48:50.160410 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:48:50.160440 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:48:50.160470 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 19:48:50.160504 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 19:48:50.160535 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 19:48:50.160565 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:48:50.160596 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Feb 13 19:48:50.160626 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:48:50.160657 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 19:48:50.160691 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 19:48:50.160722 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 19:48:50.160759 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 19:48:50.160792 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:48:50.160843 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:48:50.160878 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:48:50.160911 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:48:50.160941 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 19:48:50.160975 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 19:48:50.161005 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:48:50.161037 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:48:50.161124 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:48:50.165145 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 19:48:50.165189 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 19:48:50.165260 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 19:48:50.165295 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 19:48:50.165326 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 19:48:50.165357 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 19:48:50.165387 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 19:48:50.165418 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 19:48:50.165458 systemd[1]: Reached target machines.target - Containers. Feb 13 19:48:50.165489 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 19:48:50.165520 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:48:50.165549 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:48:50.165579 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 19:48:50.165609 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:48:50.165641 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:48:50.165670 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:48:50.165704 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 19:48:50.165735 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:48:50.165767 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 19:48:50.165797 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 19:48:50.165828 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 19:48:50.165858 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 19:48:50.165888 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 19:48:50.165919 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:48:50.165949 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:48:50.165983 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 19:48:50.166016 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 19:48:50.166046 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:48:50.166078 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 19:48:50.166108 systemd[1]: Stopped verity-setup.service. Feb 13 19:48:50.166140 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 19:48:50.166172 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 19:48:50.166202 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 19:48:50.166251 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 19:48:50.166291 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 19:48:50.166322 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 19:48:50.166353 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:48:50.166382 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 19:48:50.166412 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 19:48:50.166447 kernel: loop: module loaded Feb 13 19:48:50.166477 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:48:50.166506 kernel: fuse: init (API version 7.39) Feb 13 19:48:50.166535 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:48:50.166574 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:48:50.166607 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:48:50.166636 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:48:50.166666 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:48:50.166700 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 19:48:50.166732 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 19:48:50.166767 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:48:50.166799 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 19:48:50.166830 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 19:48:50.166860 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 19:48:50.166938 systemd-journald[1571]: Collecting audit messages is disabled. Feb 13 19:48:50.166995 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 19:48:50.167026 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 19:48:50.167060 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 19:48:50.167091 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:48:50.167123 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Feb 13 19:48:50.167152 systemd-journald[1571]: Journal started Feb 13 19:48:50.167203 systemd-journald[1571]: Runtime Journal (/run/log/journal/ec2df6ce9d9f50082cb28b4722e627dd) is 8.0M, max 75.3M, 67.3M free. Feb 13 19:48:50.178796 kernel: ACPI: bus type drm_connector registered Feb 13 19:48:49.542926 systemd[1]: Queued start job for default target multi-user.target. Feb 13 19:48:50.179301 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 19:48:49.609489 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Feb 13 19:48:49.610297 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 19:48:50.193175 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 19:48:50.194992 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:48:50.209299 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 19:48:50.214283 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:48:50.224507 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 19:48:50.229476 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:48:50.243418 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:48:50.254636 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 19:48:50.265267 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:48:50.271288 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:48:50.275330 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 19:48:50.278187 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:48:50.278508 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:48:50.281443 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 19:48:50.287505 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 19:48:50.292344 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 19:48:50.322526 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 19:48:50.357001 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 19:48:50.378739 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 19:48:50.391526 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Feb 13 19:48:50.416490 kernel: loop0: detected capacity change from 0 to 201592 Feb 13 19:48:50.421095 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:48:50.446945 systemd-journald[1571]: Time spent on flushing to /var/log/journal/ec2df6ce9d9f50082cb28b4722e627dd is 82.934ms for 913 entries. Feb 13 19:48:50.446945 systemd-journald[1571]: System Journal (/var/log/journal/ec2df6ce9d9f50082cb28b4722e627dd) is 8.0M, max 195.6M, 187.6M free. Feb 13 19:48:50.551703 systemd-journald[1571]: Received client request to flush runtime journal. Feb 13 19:48:50.551778 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 19:48:50.551814 kernel: loop1: detected capacity change from 0 to 114432 Feb 13 19:48:50.444462 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 19:48:50.452205 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Feb 13 19:48:50.515366 systemd-tmpfiles[1605]: ACLs are not supported, ignoring. Feb 13 19:48:50.515391 systemd-tmpfiles[1605]: ACLs are not supported, ignoring. Feb 13 19:48:50.531167 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:48:50.543717 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 19:48:50.556260 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 19:48:50.610454 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 19:48:50.623030 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:48:50.644268 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:48:50.653570 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 19:48:50.675271 kernel: loop2: detected capacity change from 0 to 114328 Feb 13 19:48:50.689056 systemd-tmpfiles[1643]: ACLs are not supported, ignoring. Feb 13 19:48:50.689094 systemd-tmpfiles[1643]: ACLs are not supported, ignoring. Feb 13 19:48:50.705845 udevadm[1646]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Feb 13 19:48:50.710886 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:48:50.780276 kernel: loop3: detected capacity change from 0 to 52536 Feb 13 19:48:50.832267 kernel: loop4: detected capacity change from 0 to 201592 Feb 13 19:48:50.865266 kernel: loop5: detected capacity change from 0 to 114432 Feb 13 19:48:50.879252 kernel: loop6: detected capacity change from 0 to 114328 Feb 13 19:48:50.892270 kernel: loop7: detected capacity change from 0 to 52536 Feb 13 19:48:50.909931 (sd-merge)[1651]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Feb 13 19:48:50.910916 (sd-merge)[1651]: Merged extensions into '/usr'. Feb 13 19:48:50.920018 systemd[1]: Reloading requested from client PID 1604 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 19:48:50.920054 systemd[1]: Reloading... Feb 13 19:48:51.098250 zram_generator::config[1677]: No configuration found. Feb 13 19:48:51.399412 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:48:51.517044 systemd[1]: Reloading finished in 595 ms. Feb 13 19:48:51.563335 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 19:48:51.578708 systemd[1]: Starting ensure-sysext.service... Feb 13 19:48:51.587767 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:48:51.610352 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 19:48:51.626648 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:48:51.634825 systemd[1]: Reloading requested from client PID 1728 ('systemctl') (unit ensure-sysext.service)... Feb 13 19:48:51.634851 systemd[1]: Reloading... Feb 13 19:48:51.668776 systemd-tmpfiles[1729]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 19:48:51.669472 systemd-tmpfiles[1729]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 19:48:51.673838 systemd-tmpfiles[1729]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 19:48:51.676537 systemd-tmpfiles[1729]: ACLs are not supported, ignoring. Feb 13 19:48:51.677087 systemd-tmpfiles[1729]: ACLs are not supported, ignoring. Feb 13 19:48:51.701549 systemd-tmpfiles[1729]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:48:51.701570 systemd-tmpfiles[1729]: Skipping /boot Feb 13 19:48:51.732475 systemd-tmpfiles[1729]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:48:51.732653 systemd-tmpfiles[1729]: Skipping /boot Feb 13 19:48:51.772418 systemd-udevd[1731]: Using default interface naming scheme 'v255'. Feb 13 19:48:51.841289 zram_generator::config[1761]: No configuration found. Feb 13 19:48:52.068470 (udev-worker)[1784]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:48:52.201272 ldconfig[1597]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 19:48:52.235828 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:48:52.352416 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (1790) Feb 13 19:48:52.402719 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Feb 13 19:48:52.403420 systemd[1]: Reloading finished in 766 ms. Feb 13 19:48:52.436718 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:48:52.455310 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 19:48:52.459661 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:48:52.546881 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Feb 13 19:48:52.554420 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 19:48:52.556956 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:48:52.563888 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:48:52.570049 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:48:52.577973 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:48:52.580193 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:48:52.583306 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 19:48:52.591876 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:48:52.605928 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:48:52.615918 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 19:48:52.623874 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:48:52.660649 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:48:52.718688 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:48:52.721164 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:48:52.721310 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 19:48:52.734550 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 19:48:52.739484 systemd[1]: Finished ensure-sysext.service. Feb 13 19:48:52.742107 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:48:52.743682 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:48:52.747348 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 19:48:52.765880 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:48:52.766855 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:48:52.772462 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:48:52.772822 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:48:52.776092 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:48:52.776438 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:48:52.836569 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 19:48:52.849212 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 19:48:52.853777 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 19:48:52.877972 augenrules[1963]: No rules Feb 13 19:48:52.878102 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 19:48:52.891706 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Feb 13 19:48:52.905700 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 19:48:52.921521 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 19:48:52.923688 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:48:52.923829 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:48:52.928690 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 19:48:52.933380 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 19:48:52.933868 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 19:48:52.956272 lvm[1969]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:48:53.003988 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 19:48:53.016935 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 19:48:53.029268 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 19:48:53.030175 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:48:53.046722 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 19:48:53.066361 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:48:53.070398 lvm[1983]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:48:53.119376 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 19:48:53.152967 systemd-networkd[1926]: lo: Link UP Feb 13 19:48:53.152983 systemd-networkd[1926]: lo: Gained carrier Feb 13 19:48:53.156558 systemd-networkd[1926]: Enumeration completed Feb 13 19:48:53.156968 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:48:53.160426 systemd-resolved[1927]: Positive Trust Anchors: Feb 13 19:48:53.160460 systemd-resolved[1927]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:48:53.160524 systemd-resolved[1927]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:48:53.162707 systemd-networkd[1926]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:48:53.162723 systemd-networkd[1926]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:48:53.165108 systemd-networkd[1926]: eth0: Link UP Feb 13 19:48:53.165524 systemd-networkd[1926]: eth0: Gained carrier Feb 13 19:48:53.165557 systemd-networkd[1926]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:48:53.168546 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 19:48:53.177804 systemd-resolved[1927]: Defaulting to hostname 'linux'. Feb 13 19:48:53.178334 systemd-networkd[1926]: eth0: DHCPv4 address 172.31.25.134/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 19:48:53.182031 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:48:53.184337 systemd[1]: Reached target network.target - Network. Feb 13 19:48:53.186076 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:48:53.188333 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:48:53.190471 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 19:48:53.192796 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 19:48:53.195386 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 19:48:53.197613 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 19:48:53.199984 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 19:48:53.202309 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 19:48:53.202363 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:48:53.204103 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:48:53.207176 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 19:48:53.213944 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 19:48:53.227531 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 19:48:53.230763 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 19:48:53.233114 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:48:53.234967 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:48:53.237109 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:48:53.237159 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:48:53.252008 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 19:48:53.256690 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Feb 13 19:48:53.262597 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 19:48:53.279121 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 19:48:53.286755 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 19:48:53.294670 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 19:48:53.304580 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 19:48:53.308839 jq[1994]: false Feb 13 19:48:53.320033 systemd[1]: Started ntpd.service - Network Time Service. Feb 13 19:48:53.329112 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 19:48:53.338527 systemd[1]: Starting setup-oem.service - Setup OEM... Feb 13 19:48:53.345434 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 19:48:53.356531 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 19:48:53.381571 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 19:48:53.386826 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 19:48:53.388821 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 19:48:53.390853 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 19:48:53.400450 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 19:48:53.401764 extend-filesystems[1995]: Found loop4 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found loop5 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found loop6 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found loop7 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1p1 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1p2 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1p3 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found usr Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1p4 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1p6 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1p7 Feb 13 19:48:53.421840 extend-filesystems[1995]: Found nvme0n1p9 Feb 13 19:48:53.421840 extend-filesystems[1995]: Checking size of /dev/nvme0n1p9 Feb 13 19:48:53.408956 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 19:48:53.476005 dbus-daemon[1993]: [system] SELinux support is enabled Feb 13 19:48:53.409338 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 19:48:53.494895 dbus-daemon[1993]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1926 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Feb 13 19:48:53.476315 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 19:48:53.487325 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 19:48:53.487375 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 19:48:53.528687 extend-filesystems[1995]: Resized partition /dev/nvme0n1p9 Feb 13 19:48:53.501483 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 19:48:53.554619 jq[2010]: true Feb 13 19:48:53.547467 dbus-daemon[1993]: [system] Successfully activated service 'org.freedesktop.systemd1' Feb 13 19:48:53.555629 extend-filesystems[2026]: resize2fs 1.47.1 (20-May-2024) Feb 13 19:48:53.501525 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 19:48:53.516180 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 19:48:53.517332 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 19:48:53.522756 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 19:48:53.523142 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 19:48:53.569827 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Feb 13 19:48:53.600545 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 17:35:09 UTC 2025 (1): Starting Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: ---------------------------------------------------- Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: ntp-4 is maintained by Network Time Foundation, Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: corporation. Support and training for ntp-4 are Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: available at https://www.nwtime.org/support Feb 13 19:48:53.601900 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: ---------------------------------------------------- Feb 13 19:48:53.600699 ntpd[1998]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 17:35:09 UTC 2025 (1): Starting Feb 13 19:48:53.600785 ntpd[1998]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 19:48:53.600809 ntpd[1998]: ---------------------------------------------------- Feb 13 19:48:53.600828 ntpd[1998]: ntp-4 is maintained by Network Time Foundation, Feb 13 19:48:53.600848 ntpd[1998]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 19:48:53.600867 ntpd[1998]: corporation. Support and training for ntp-4 are Feb 13 19:48:53.621588 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: proto: precision = 0.096 usec (-23) Feb 13 19:48:53.600886 ntpd[1998]: available at https://www.nwtime.org/support Feb 13 19:48:53.600904 ntpd[1998]: ---------------------------------------------------- Feb 13 19:48:53.619381 ntpd[1998]: proto: precision = 0.096 usec (-23) Feb 13 19:48:53.622271 ntpd[1998]: basedate set to 2025-02-01 Feb 13 19:48:53.639850 tar[2019]: linux-arm64/LICENSE Feb 13 19:48:53.639850 tar[2019]: linux-arm64/helm Feb 13 19:48:53.649664 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: basedate set to 2025-02-01 Feb 13 19:48:53.649664 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: gps base set to 2025-02-02 (week 2352) Feb 13 19:48:53.649664 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 19:48:53.649664 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 19:48:53.622310 ntpd[1998]: gps base set to 2025-02-02 (week 2352) Feb 13 19:48:53.643749 ntpd[1998]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 19:48:53.643831 ntpd[1998]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 19:48:53.658463 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 19:48:53.658463 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Listen normally on 3 eth0 172.31.25.134:123 Feb 13 19:48:53.658463 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Listen normally on 4 lo [::1]:123 Feb 13 19:48:53.658463 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: bind(21) AF_INET6 fe80::49e:4fff:fe2e:9e83%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 19:48:53.658463 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: unable to create socket on eth0 (5) for fe80::49e:4fff:fe2e:9e83%2#123 Feb 13 19:48:53.658463 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: failed to init interface for address fe80::49e:4fff:fe2e:9e83%2 Feb 13 19:48:53.658463 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: Listening on routing socket on fd #21 for interface updates Feb 13 19:48:53.657515 ntpd[1998]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 19:48:53.659076 update_engine[2008]: I20250213 19:48:53.653820 2008 main.cc:92] Flatcar Update Engine starting Feb 13 19:48:53.657589 ntpd[1998]: Listen normally on 3 eth0 172.31.25.134:123 Feb 13 19:48:53.657661 ntpd[1998]: Listen normally on 4 lo [::1]:123 Feb 13 19:48:53.657737 ntpd[1998]: bind(21) AF_INET6 fe80::49e:4fff:fe2e:9e83%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 19:48:53.657777 ntpd[1998]: unable to create socket on eth0 (5) for fe80::49e:4fff:fe2e:9e83%2#123 Feb 13 19:48:53.657806 ntpd[1998]: failed to init interface for address fe80::49e:4fff:fe2e:9e83%2 Feb 13 19:48:53.657861 ntpd[1998]: Listening on routing socket on fd #21 for interface updates Feb 13 19:48:53.682379 update_engine[2008]: I20250213 19:48:53.679594 2008 update_check_scheduler.cc:74] Next update check in 6m22s Feb 13 19:48:53.667994 (ntainerd)[2031]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 19:48:53.676932 systemd[1]: Started update-engine.service - Update Engine. Feb 13 19:48:53.702680 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Feb 13 19:48:53.702785 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:48:53.702785 ntpd[1998]: 13 Feb 19:48:53 ntpd[1998]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:48:53.695639 ntpd[1998]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:48:53.693777 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 19:48:53.695691 ntpd[1998]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 19:48:53.711765 jq[2032]: true Feb 13 19:48:53.726465 extend-filesystems[2026]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Feb 13 19:48:53.726465 extend-filesystems[2026]: old_desc_blocks = 1, new_desc_blocks = 1 Feb 13 19:48:53.726465 extend-filesystems[2026]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Feb 13 19:48:53.724694 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 19:48:53.748585 extend-filesystems[1995]: Resized filesystem in /dev/nvme0n1p9 Feb 13 19:48:53.725984 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 19:48:53.762629 systemd[1]: Finished setup-oem.service - Setup OEM. Feb 13 19:48:53.856642 coreos-metadata[1992]: Feb 13 19:48:53.856 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 19:48:53.859038 systemd-logind[2004]: Watching system buttons on /dev/input/event0 (Power Button) Feb 13 19:48:53.859147 systemd-logind[2004]: Watching system buttons on /dev/input/event1 (Sleep Button) Feb 13 19:48:53.860489 coreos-metadata[1992]: Feb 13 19:48:53.860 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Feb 13 19:48:53.861387 coreos-metadata[1992]: Feb 13 19:48:53.861 INFO Fetch successful Feb 13 19:48:53.861387 coreos-metadata[1992]: Feb 13 19:48:53.861 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Feb 13 19:48:53.861557 systemd-logind[2004]: New seat seat0. Feb 13 19:48:53.862255 coreos-metadata[1992]: Feb 13 19:48:53.862 INFO Fetch successful Feb 13 19:48:53.862255 coreos-metadata[1992]: Feb 13 19:48:53.862 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Feb 13 19:48:53.863196 coreos-metadata[1992]: Feb 13 19:48:53.862 INFO Fetch successful Feb 13 19:48:53.863196 coreos-metadata[1992]: Feb 13 19:48:53.863 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Feb 13 19:48:53.864023 coreos-metadata[1992]: Feb 13 19:48:53.863 INFO Fetch successful Feb 13 19:48:53.864023 coreos-metadata[1992]: Feb 13 19:48:53.863 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Feb 13 19:48:53.864785 coreos-metadata[1992]: Feb 13 19:48:53.864 INFO Fetch failed with 404: resource not found Feb 13 19:48:53.864785 coreos-metadata[1992]: Feb 13 19:48:53.864 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Feb 13 19:48:53.867947 coreos-metadata[1992]: Feb 13 19:48:53.867 INFO Fetch successful Feb 13 19:48:53.869698 coreos-metadata[1992]: Feb 13 19:48:53.868 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Feb 13 19:48:53.870023 coreos-metadata[1992]: Feb 13 19:48:53.869 INFO Fetch successful Feb 13 19:48:53.870023 coreos-metadata[1992]: Feb 13 19:48:53.869 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Feb 13 19:48:53.872236 coreos-metadata[1992]: Feb 13 19:48:53.871 INFO Fetch successful Feb 13 19:48:53.872236 coreos-metadata[1992]: Feb 13 19:48:53.871 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Feb 13 19:48:53.872463 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 19:48:53.881557 coreos-metadata[1992]: Feb 13 19:48:53.881 INFO Fetch successful Feb 13 19:48:53.881557 coreos-metadata[1992]: Feb 13 19:48:53.881 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Feb 13 19:48:53.884260 coreos-metadata[1992]: Feb 13 19:48:53.882 INFO Fetch successful Feb 13 19:48:53.894451 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (1783) Feb 13 19:48:54.033791 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 19:48:54.065377 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Feb 13 19:48:54.071049 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 19:48:54.100738 bash[2093]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:48:54.105343 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 19:48:54.125692 systemd[1]: Starting sshkeys.service... Feb 13 19:48:54.183301 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Feb 13 19:48:54.216236 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Feb 13 19:48:54.286181 dbus-daemon[1993]: [system] Successfully activated service 'org.freedesktop.hostname1' Feb 13 19:48:54.287570 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Feb 13 19:48:54.304002 dbus-daemon[1993]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.6' (uid=0 pid=2039 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Feb 13 19:48:54.326476 systemd[1]: Starting polkit.service - Authorization Manager... Feb 13 19:48:54.418555 polkitd[2167]: Started polkitd version 121 Feb 13 19:48:54.458188 polkitd[2167]: Loading rules from directory /etc/polkit-1/rules.d Feb 13 19:48:54.458332 polkitd[2167]: Loading rules from directory /usr/share/polkit-1/rules.d Feb 13 19:48:54.461388 coreos-metadata[2149]: Feb 13 19:48:54.460 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 19:48:54.462714 coreos-metadata[2149]: Feb 13 19:48:54.462 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Feb 13 19:48:54.463908 coreos-metadata[2149]: Feb 13 19:48:54.463 INFO Fetch successful Feb 13 19:48:54.464372 coreos-metadata[2149]: Feb 13 19:48:54.464 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Feb 13 19:48:54.465549 coreos-metadata[2149]: Feb 13 19:48:54.465 INFO Fetch successful Feb 13 19:48:54.468255 polkitd[2167]: Finished loading, compiling and executing 2 rules Feb 13 19:48:54.472514 dbus-daemon[1993]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Feb 13 19:48:54.473433 polkitd[2167]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Feb 13 19:48:54.473942 locksmithd[2044]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 19:48:54.483741 unknown[2149]: wrote ssh authorized keys file for user: core Feb 13 19:48:54.525643 systemd[1]: Started polkit.service - Authorization Manager. Feb 13 19:48:54.566426 systemd-networkd[1926]: eth0: Gained IPv6LL Feb 13 19:48:54.577277 update-ssh-keys[2188]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:48:54.584359 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 19:48:54.591757 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Feb 13 19:48:54.596714 systemd-hostnamed[2039]: Hostname set to (transient) Feb 13 19:48:54.604956 systemd[1]: Finished sshkeys.service. Feb 13 19:48:54.605910 systemd-resolved[1927]: System hostname changed to 'ip-172-31-25-134'. Feb 13 19:48:54.614518 sshd_keygen[2036]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 19:48:54.616472 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 19:48:54.631789 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Feb 13 19:48:54.647570 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:48:54.663693 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 19:48:54.699704 containerd[2031]: time="2025-02-13T19:48:54.699547463Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Feb 13 19:48:54.744567 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 19:48:54.773375 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 19:48:54.791767 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 19:48:54.806573 amazon-ssm-agent[2199]: Initializing new seelog logger Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: New Seelog Logger Creation Complete Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 processing appconfig overrides Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 processing appconfig overrides Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 processing appconfig overrides Feb 13 19:48:54.813270 amazon-ssm-agent[2199]: 2025-02-13 19:48:54 INFO Proxy environment variables: Feb 13 19:48:54.811046 systemd[1]: Started sshd@0-172.31.25.134:22-139.178.89.65:58276.service - OpenSSH per-connection server daemon (139.178.89.65:58276). Feb 13 19:48:54.827658 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.828298 amazon-ssm-agent[2199]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 19:48:54.830266 amazon-ssm-agent[2199]: 2025/02/13 19:48:54 processing appconfig overrides Feb 13 19:48:54.856022 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 19:48:54.856940 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 19:48:54.870500 containerd[2031]: time="2025-02-13T19:48:54.870435756Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:48:54.873751 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 19:48:54.880679 containerd[2031]: time="2025-02-13T19:48:54.880324440Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:48:54.881057 containerd[2031]: time="2025-02-13T19:48:54.881021676Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 19:48:54.881700 containerd[2031]: time="2025-02-13T19:48:54.881451192Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 19:48:54.885277 containerd[2031]: time="2025-02-13T19:48:54.883282116Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.885465564Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.885672288Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.885715788Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.886032456Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.886068732Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.886120248Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.886146888Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.886380588Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.886780548Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.887016084Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:48:54.887260 containerd[2031]: time="2025-02-13T19:48:54.887046684Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 19:48:54.892793 containerd[2031]: time="2025-02-13T19:48:54.892317396Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 19:48:54.892793 containerd[2031]: time="2025-02-13T19:48:54.892530192Z" level=info msg="metadata content store policy set" policy=shared Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.903068904Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.903183048Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.903249408Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.903291444Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.903324624Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.903574368Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904010484Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904248660Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904287972Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904332264Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904379388Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904418880Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904452936Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905272 containerd[2031]: time="2025-02-13T19:48:54.904484412Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904517076Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904547808Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904580124Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904607880Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904647000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904678116Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904708104Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904739052Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904795764Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904830960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904861212Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904893024Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904924500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.905910 containerd[2031]: time="2025-02-13T19:48:54.904959252Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.906512 containerd[2031]: time="2025-02-13T19:48:54.904992732Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.906512 containerd[2031]: time="2025-02-13T19:48:54.905023824Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.906512 containerd[2031]: time="2025-02-13T19:48:54.905053788Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.906512 containerd[2031]: time="2025-02-13T19:48:54.905090772Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 19:48:54.906512 containerd[2031]: time="2025-02-13T19:48:54.905136336Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.906512 containerd[2031]: time="2025-02-13T19:48:54.905165436Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.906512 containerd[2031]: time="2025-02-13T19:48:54.905192124Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910612680Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910690104Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910718004Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910747404Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910773228Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910821432Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910848888Z" level=info msg="NRI interface is disabled by configuration." Feb 13 19:48:54.912342 containerd[2031]: time="2025-02-13T19:48:54.910882008Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 19:48:54.913283 containerd[2031]: time="2025-02-13T19:48:54.913073544Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 19:48:54.914132 containerd[2031]: time="2025-02-13T19:48:54.913204200Z" level=info msg="Connect containerd service" Feb 13 19:48:54.914132 containerd[2031]: time="2025-02-13T19:48:54.913606260Z" level=info msg="using legacy CRI server" Feb 13 19:48:54.914132 containerd[2031]: time="2025-02-13T19:48:54.913628460Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 19:48:54.914132 containerd[2031]: time="2025-02-13T19:48:54.913798044Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 19:48:54.914375 amazon-ssm-agent[2199]: 2025-02-13 19:48:54 INFO http_proxy: Feb 13 19:48:54.918135 containerd[2031]: time="2025-02-13T19:48:54.917432628Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:48:54.918135 containerd[2031]: time="2025-02-13T19:48:54.917993208Z" level=info msg="Start subscribing containerd event" Feb 13 19:48:54.918135 containerd[2031]: time="2025-02-13T19:48:54.918094620Z" level=info msg="Start recovering state" Feb 13 19:48:54.918408 containerd[2031]: time="2025-02-13T19:48:54.918234756Z" level=info msg="Start event monitor" Feb 13 19:48:54.918408 containerd[2031]: time="2025-02-13T19:48:54.918265500Z" level=info msg="Start snapshots syncer" Feb 13 19:48:54.918408 containerd[2031]: time="2025-02-13T19:48:54.918288312Z" level=info msg="Start cni network conf syncer for default" Feb 13 19:48:54.918408 containerd[2031]: time="2025-02-13T19:48:54.918307236Z" level=info msg="Start streaming server" Feb 13 19:48:54.918580 containerd[2031]: time="2025-02-13T19:48:54.918096624Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 19:48:54.918580 containerd[2031]: time="2025-02-13T19:48:54.918556788Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 19:48:54.918994 containerd[2031]: time="2025-02-13T19:48:54.918653952Z" level=info msg="containerd successfully booted in 0.228022s" Feb 13 19:48:54.918786 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 19:48:54.961647 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 19:48:54.975862 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 19:48:54.990997 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Feb 13 19:48:54.996072 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 19:48:55.014248 amazon-ssm-agent[2199]: 2025-02-13 19:48:54 INFO no_proxy: Feb 13 19:48:55.114380 amazon-ssm-agent[2199]: 2025-02-13 19:48:54 INFO https_proxy: Feb 13 19:48:55.159033 sshd[2218]: Accepted publickey for core from 139.178.89.65 port 58276 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:48:55.164561 sshd[2218]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:48:55.192623 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 19:48:55.204831 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 19:48:55.212499 amazon-ssm-agent[2199]: 2025-02-13 19:48:54 INFO Checking if agent identity type OnPrem can be assumed Feb 13 19:48:55.221544 systemd-logind[2004]: New session 1 of user core. Feb 13 19:48:55.246791 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 19:48:55.266776 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 19:48:55.287183 (systemd)[2234]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 19:48:55.311676 amazon-ssm-agent[2199]: 2025-02-13 19:48:54 INFO Checking if agent identity type EC2 can be assumed Feb 13 19:48:55.411075 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO Agent will take identity from EC2 Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [amazon-ssm-agent] OS: linux, Arch: arm64 Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [amazon-ssm-agent] Starting Core Agent Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [amazon-ssm-agent] registrar detected. Attempting registration Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [Registrar] Starting registrar module Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [EC2Identity] EC2 registration was successful. Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [CredentialRefresher] credentialRefresher has started Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [CredentialRefresher] Starting credentials refresher loop Feb 13 19:48:55.470780 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO EC2RoleProvider Successfully connected with instance profile role credentials Feb 13 19:48:55.510892 amazon-ssm-agent[2199]: 2025-02-13 19:48:55 INFO [CredentialRefresher] Next credential rotation will be in 30.266634853066666 minutes Feb 13 19:48:55.613548 systemd[2234]: Queued start job for default target default.target. Feb 13 19:48:55.622701 systemd[2234]: Created slice app.slice - User Application Slice. Feb 13 19:48:55.622920 systemd[2234]: Reached target paths.target - Paths. Feb 13 19:48:55.623069 systemd[2234]: Reached target timers.target - Timers. Feb 13 19:48:55.627451 systemd[2234]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 19:48:55.658602 systemd[2234]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 19:48:55.658847 systemd[2234]: Reached target sockets.target - Sockets. Feb 13 19:48:55.658881 systemd[2234]: Reached target basic.target - Basic System. Feb 13 19:48:55.658980 systemd[2234]: Reached target default.target - Main User Target. Feb 13 19:48:55.659048 systemd[2234]: Startup finished in 348ms. Feb 13 19:48:55.659192 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 19:48:55.670756 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 19:48:55.840772 systemd[1]: Started sshd@1-172.31.25.134:22-139.178.89.65:36382.service - OpenSSH per-connection server daemon (139.178.89.65:36382). Feb 13 19:48:55.898386 tar[2019]: linux-arm64/README.md Feb 13 19:48:55.921989 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 19:48:56.029957 sshd[2246]: Accepted publickey for core from 139.178.89.65 port 36382 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:48:56.033160 sshd[2246]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:48:56.043005 systemd-logind[2004]: New session 2 of user core. Feb 13 19:48:56.053851 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 19:48:56.183385 sshd[2246]: pam_unix(sshd:session): session closed for user core Feb 13 19:48:56.189212 systemd[1]: sshd@1-172.31.25.134:22-139.178.89.65:36382.service: Deactivated successfully. Feb 13 19:48:56.193845 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 19:48:56.197829 systemd-logind[2004]: Session 2 logged out. Waiting for processes to exit. Feb 13 19:48:56.200457 systemd-logind[2004]: Removed session 2. Feb 13 19:48:56.220119 systemd[1]: Started sshd@2-172.31.25.134:22-139.178.89.65:36394.service - OpenSSH per-connection server daemon (139.178.89.65:36394). Feb 13 19:48:56.402941 sshd[2256]: Accepted publickey for core from 139.178.89.65 port 36394 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:48:56.405594 sshd[2256]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:48:56.415317 systemd-logind[2004]: New session 3 of user core. Feb 13 19:48:56.423584 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 19:48:56.502075 amazon-ssm-agent[2199]: 2025-02-13 19:48:56 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Feb 13 19:48:56.552502 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:48:56.557287 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 19:48:56.559680 systemd[1]: Startup finished in 1.149s (kernel) + 8.379s (initrd) + 8.431s (userspace) = 17.960s. Feb 13 19:48:56.568515 sshd[2256]: pam_unix(sshd:session): session closed for user core Feb 13 19:48:56.576087 (kubelet)[2269]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:48:56.583121 systemd[1]: sshd@2-172.31.25.134:22-139.178.89.65:36394.service: Deactivated successfully. Feb 13 19:48:56.589788 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 19:48:56.596588 systemd-logind[2004]: Session 3 logged out. Waiting for processes to exit. Feb 13 19:48:56.602033 ntpd[1998]: Listen normally on 6 eth0 [fe80::49e:4fff:fe2e:9e83%2]:123 Feb 13 19:48:56.602557 ntpd[1998]: 13 Feb 19:48:56 ntpd[1998]: Listen normally on 6 eth0 [fe80::49e:4fff:fe2e:9e83%2]:123 Feb 13 19:48:56.603688 systemd-logind[2004]: Removed session 3. Feb 13 19:48:56.605369 amazon-ssm-agent[2199]: 2025-02-13 19:48:56 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2262) started Feb 13 19:48:56.707293 amazon-ssm-agent[2199]: 2025-02-13 19:48:56 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Feb 13 19:48:57.430278 kubelet[2269]: E0213 19:48:57.430074 2269 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:48:57.433262 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:48:57.433591 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:48:57.434146 systemd[1]: kubelet.service: Consumed 1.302s CPU time. Feb 13 19:49:00.862121 systemd-resolved[1927]: Clock change detected. Flushing caches. Feb 13 19:49:06.864218 systemd[1]: Started sshd@3-172.31.25.134:22-139.178.89.65:55186.service - OpenSSH per-connection server daemon (139.178.89.65:55186). Feb 13 19:49:07.041830 sshd[2290]: Accepted publickey for core from 139.178.89.65 port 55186 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:49:07.044471 sshd[2290]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:49:07.051959 systemd-logind[2004]: New session 4 of user core. Feb 13 19:49:07.061913 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 19:49:07.188536 sshd[2290]: pam_unix(sshd:session): session closed for user core Feb 13 19:49:07.194979 systemd[1]: sshd@3-172.31.25.134:22-139.178.89.65:55186.service: Deactivated successfully. Feb 13 19:49:07.199369 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 19:49:07.200487 systemd-logind[2004]: Session 4 logged out. Waiting for processes to exit. Feb 13 19:49:07.202605 systemd-logind[2004]: Removed session 4. Feb 13 19:49:07.224207 systemd[1]: Started sshd@4-172.31.25.134:22-139.178.89.65:55192.service - OpenSSH per-connection server daemon (139.178.89.65:55192). Feb 13 19:49:07.402037 sshd[2297]: Accepted publickey for core from 139.178.89.65 port 55192 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:49:07.404556 sshd[2297]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:49:07.412753 systemd-logind[2004]: New session 5 of user core. Feb 13 19:49:07.418965 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 19:49:07.536516 sshd[2297]: pam_unix(sshd:session): session closed for user core Feb 13 19:49:07.542244 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 19:49:07.542283 systemd-logind[2004]: Session 5 logged out. Waiting for processes to exit. Feb 13 19:49:07.545350 systemd[1]: sshd@4-172.31.25.134:22-139.178.89.65:55192.service: Deactivated successfully. Feb 13 19:49:07.550221 systemd-logind[2004]: Removed session 5. Feb 13 19:49:07.579182 systemd[1]: Started sshd@5-172.31.25.134:22-139.178.89.65:55204.service - OpenSSH per-connection server daemon (139.178.89.65:55204). Feb 13 19:49:07.716161 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 19:49:07.725057 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:49:07.747557 sshd[2304]: Accepted publickey for core from 139.178.89.65 port 55204 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:49:07.750525 sshd[2304]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:49:07.761801 systemd-logind[2004]: New session 6 of user core. Feb 13 19:49:07.776039 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 19:49:07.912990 sshd[2304]: pam_unix(sshd:session): session closed for user core Feb 13 19:49:07.920670 systemd[1]: sshd@5-172.31.25.134:22-139.178.89.65:55204.service: Deactivated successfully. Feb 13 19:49:07.924774 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 19:49:07.927525 systemd-logind[2004]: Session 6 logged out. Waiting for processes to exit. Feb 13 19:49:07.932524 systemd-logind[2004]: Removed session 6. Feb 13 19:49:07.953373 systemd[1]: Started sshd@6-172.31.25.134:22-139.178.89.65:55220.service - OpenSSH per-connection server daemon (139.178.89.65:55220). Feb 13 19:49:08.045300 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:08.060167 (kubelet)[2321]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:49:08.136766 sshd[2314]: Accepted publickey for core from 139.178.89.65 port 55220 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:49:08.140028 sshd[2314]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:49:08.144604 kubelet[2321]: E0213 19:49:08.144159 2321 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:49:08.152544 systemd-logind[2004]: New session 7 of user core. Feb 13 19:49:08.157965 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 19:49:08.158583 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:49:08.158935 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:49:08.293569 sudo[2329]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 19:49:08.294209 sudo[2329]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:49:08.860588 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 19:49:08.877209 (dockerd)[2344]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 19:49:09.331728 dockerd[2344]: time="2025-02-13T19:49:09.329988578Z" level=info msg="Starting up" Feb 13 19:49:09.532353 dockerd[2344]: time="2025-02-13T19:49:09.532061475Z" level=info msg="Loading containers: start." Feb 13 19:49:09.725727 kernel: Initializing XFRM netlink socket Feb 13 19:49:09.786564 (udev-worker)[2365]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:49:09.885102 systemd-networkd[1926]: docker0: Link UP Feb 13 19:49:09.909143 dockerd[2344]: time="2025-02-13T19:49:09.909075304Z" level=info msg="Loading containers: done." Feb 13 19:49:09.931080 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2119485424-merged.mount: Deactivated successfully. Feb 13 19:49:09.934516 dockerd[2344]: time="2025-02-13T19:49:09.934443881Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 19:49:09.934675 dockerd[2344]: time="2025-02-13T19:49:09.934630973Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Feb 13 19:49:09.934912 dockerd[2344]: time="2025-02-13T19:49:09.934859369Z" level=info msg="Daemon has completed initialization" Feb 13 19:49:09.992466 dockerd[2344]: time="2025-02-13T19:49:09.991477229Z" level=info msg="API listen on /run/docker.sock" Feb 13 19:49:09.991793 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 19:49:10.881283 containerd[2031]: time="2025-02-13T19:49:10.880920857Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\"" Feb 13 19:49:11.495900 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1177854695.mount: Deactivated successfully. Feb 13 19:49:12.812896 containerd[2031]: time="2025-02-13T19:49:12.812826931Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:12.815073 containerd[2031]: time="2025-02-13T19:49:12.815000767Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.2: active requests=0, bytes read=26218236" Feb 13 19:49:12.816172 containerd[2031]: time="2025-02-13T19:49:12.816103447Z" level=info msg="ImageCreate event name:\"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:12.821707 containerd[2031]: time="2025-02-13T19:49:12.821612203Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:12.824608 containerd[2031]: time="2025-02-13T19:49:12.824018503Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.2\" with image id \"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\", size \"26215036\" in 1.943038858s" Feb 13 19:49:12.824608 containerd[2031]: time="2025-02-13T19:49:12.824078347Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\" returns image reference \"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\"" Feb 13 19:49:12.825197 containerd[2031]: time="2025-02-13T19:49:12.825159055Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\"" Feb 13 19:49:14.201271 containerd[2031]: time="2025-02-13T19:49:14.201190554Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:14.203309 containerd[2031]: time="2025-02-13T19:49:14.203243502Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.2: active requests=0, bytes read=22528145" Feb 13 19:49:14.204279 containerd[2031]: time="2025-02-13T19:49:14.203871366Z" level=info msg="ImageCreate event name:\"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:14.209480 containerd[2031]: time="2025-02-13T19:49:14.209378586Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:14.215714 containerd[2031]: time="2025-02-13T19:49:14.214512246Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.2\" with image id \"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\", size \"23968941\" in 1.389145303s" Feb 13 19:49:14.215714 containerd[2031]: time="2025-02-13T19:49:14.214592154Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\" returns image reference \"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\"" Feb 13 19:49:14.218661 containerd[2031]: time="2025-02-13T19:49:14.218614074Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\"" Feb 13 19:49:15.478885 containerd[2031]: time="2025-02-13T19:49:15.478808996Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:15.480620 containerd[2031]: time="2025-02-13T19:49:15.480540008Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.2: active requests=0, bytes read=17480800" Feb 13 19:49:15.482414 containerd[2031]: time="2025-02-13T19:49:15.482342372Z" level=info msg="ImageCreate event name:\"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:15.491213 containerd[2031]: time="2025-02-13T19:49:15.491131760Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:15.493642 containerd[2031]: time="2025-02-13T19:49:15.493395032Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.2\" with image id \"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\", size \"18921614\" in 1.27449597s" Feb 13 19:49:15.493642 containerd[2031]: time="2025-02-13T19:49:15.493451024Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\" returns image reference \"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\"" Feb 13 19:49:15.494425 containerd[2031]: time="2025-02-13T19:49:15.494378912Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\"" Feb 13 19:49:16.792358 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1132871189.mount: Deactivated successfully. Feb 13 19:49:17.353102 containerd[2031]: time="2025-02-13T19:49:17.352927977Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:17.354366 containerd[2031]: time="2025-02-13T19:49:17.354312117Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.2: active requests=0, bytes read=27363382" Feb 13 19:49:17.355167 containerd[2031]: time="2025-02-13T19:49:17.355080381Z" level=info msg="ImageCreate event name:\"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:17.358621 containerd[2031]: time="2025-02-13T19:49:17.358540929Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:17.360265 containerd[2031]: time="2025-02-13T19:49:17.360062985Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.2\" with image id \"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\", repo tag \"registry.k8s.io/kube-proxy:v1.32.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\", size \"27362401\" in 1.865628405s" Feb 13 19:49:17.360265 containerd[2031]: time="2025-02-13T19:49:17.360115605Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\" returns image reference \"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\"" Feb 13 19:49:17.361079 containerd[2031]: time="2025-02-13T19:49:17.360999093Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Feb 13 19:49:17.885565 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3611853598.mount: Deactivated successfully. Feb 13 19:49:18.335025 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 13 19:49:18.345979 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:49:18.678956 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:18.692309 (kubelet)[2608]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:49:18.800388 kubelet[2608]: E0213 19:49:18.800106 2608 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:49:18.807001 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:49:18.807383 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:49:19.177817 containerd[2031]: time="2025-02-13T19:49:19.176806150Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:19.179309 containerd[2031]: time="2025-02-13T19:49:19.179230546Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951622" Feb 13 19:49:19.181656 containerd[2031]: time="2025-02-13T19:49:19.181582294Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:19.188023 containerd[2031]: time="2025-02-13T19:49:19.187925531Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:19.190578 containerd[2031]: time="2025-02-13T19:49:19.190367555Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.829304718s" Feb 13 19:49:19.190578 containerd[2031]: time="2025-02-13T19:49:19.190427891Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Feb 13 19:49:19.191964 containerd[2031]: time="2025-02-13T19:49:19.191587031Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Feb 13 19:49:19.716190 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2270092682.mount: Deactivated successfully. Feb 13 19:49:19.730452 containerd[2031]: time="2025-02-13T19:49:19.730391917Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:19.732419 containerd[2031]: time="2025-02-13T19:49:19.732376573Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Feb 13 19:49:19.734920 containerd[2031]: time="2025-02-13T19:49:19.734878429Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:19.740959 containerd[2031]: time="2025-02-13T19:49:19.740894041Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:19.742741 containerd[2031]: time="2025-02-13T19:49:19.742672129Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 551.026442ms" Feb 13 19:49:19.742913 containerd[2031]: time="2025-02-13T19:49:19.742878997Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Feb 13 19:49:19.746020 containerd[2031]: time="2025-02-13T19:49:19.745965805Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Feb 13 19:49:20.372478 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1004543660.mount: Deactivated successfully. Feb 13 19:49:22.455697 containerd[2031]: time="2025-02-13T19:49:22.455593023Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:22.462110 containerd[2031]: time="2025-02-13T19:49:22.462038535Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67812429" Feb 13 19:49:22.465899 containerd[2031]: time="2025-02-13T19:49:22.465758763Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:22.475632 containerd[2031]: time="2025-02-13T19:49:22.475538955Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:22.479387 containerd[2031]: time="2025-02-13T19:49:22.478881243Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 2.731761842s" Feb 13 19:49:22.479387 containerd[2031]: time="2025-02-13T19:49:22.478947819Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" Feb 13 19:49:24.872543 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Feb 13 19:49:28.835676 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Feb 13 19:49:28.845058 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:49:29.188101 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:29.198403 (kubelet)[2709]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:49:29.282718 kubelet[2709]: E0213 19:49:29.279921 2709 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:49:29.284524 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:49:29.284910 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:49:29.990179 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:30.001207 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:49:30.057843 systemd[1]: Reloading requested from client PID 2723 ('systemctl') (unit session-7.scope)... Feb 13 19:49:30.057876 systemd[1]: Reloading... Feb 13 19:49:30.295729 zram_generator::config[2772]: No configuration found. Feb 13 19:49:30.519634 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:49:30.688524 systemd[1]: Reloading finished in 629 ms. Feb 13 19:49:30.774833 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Feb 13 19:49:30.775020 systemd[1]: kubelet.service: Failed with result 'signal'. Feb 13 19:49:30.776016 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:30.785360 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:49:31.088219 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:31.104229 (kubelet)[2827]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:49:31.175511 kubelet[2827]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:49:31.175511 kubelet[2827]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:49:31.175511 kubelet[2827]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:49:31.176106 kubelet[2827]: I0213 19:49:31.175639 2827 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:49:32.042831 kubelet[2827]: I0213 19:49:32.042764 2827 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:49:32.042831 kubelet[2827]: I0213 19:49:32.042814 2827 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:49:32.043316 kubelet[2827]: I0213 19:49:32.043271 2827 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:49:32.091995 kubelet[2827]: E0213 19:49:32.091926 2827 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.25.134:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:32.095110 kubelet[2827]: I0213 19:49:32.094460 2827 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:49:32.107732 kubelet[2827]: E0213 19:49:32.105654 2827 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:49:32.107732 kubelet[2827]: I0213 19:49:32.105731 2827 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:49:32.110601 kubelet[2827]: I0213 19:49:32.110567 2827 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:49:32.112310 kubelet[2827]: I0213 19:49:32.112261 2827 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:49:32.112816 kubelet[2827]: I0213 19:49:32.112455 2827 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-25-134","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:49:32.113075 kubelet[2827]: I0213 19:49:32.113053 2827 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:49:32.113205 kubelet[2827]: I0213 19:49:32.113158 2827 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:49:32.113647 kubelet[2827]: I0213 19:49:32.113604 2827 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:49:32.120458 kubelet[2827]: I0213 19:49:32.120422 2827 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:49:32.120654 kubelet[2827]: I0213 19:49:32.120632 2827 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:49:32.120786 kubelet[2827]: I0213 19:49:32.120767 2827 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:49:32.120887 kubelet[2827]: I0213 19:49:32.120868 2827 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:49:32.129367 kubelet[2827]: W0213 19:49:32.129270 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.25.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-134&limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:32.130614 kubelet[2827]: E0213 19:49:32.129371 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.25.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-134&limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:32.130614 kubelet[2827]: I0213 19:49:32.129899 2827 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Feb 13 19:49:32.131157 kubelet[2827]: I0213 19:49:32.130650 2827 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:49:32.131157 kubelet[2827]: W0213 19:49:32.130813 2827 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 19:49:32.132434 kubelet[2827]: W0213 19:49:32.132185 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.25.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:32.132434 kubelet[2827]: E0213 19:49:32.132283 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.25.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:32.134546 kubelet[2827]: I0213 19:49:32.134481 2827 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:49:32.134546 kubelet[2827]: I0213 19:49:32.134550 2827 server.go:1287] "Started kubelet" Feb 13 19:49:32.144575 kubelet[2827]: E0213 19:49:32.144306 2827 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.25.134:6443/api/v1/namespaces/default/events\": dial tcp 172.31.25.134:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-25-134.1823dc5cc7559807 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-25-134,UID:ip-172-31-25-134,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-25-134,},FirstTimestamp:2025-02-13 19:49:32.134520839 +0000 UTC m=+1.024298622,LastTimestamp:2025-02-13 19:49:32.134520839 +0000 UTC m=+1.024298622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-25-134,}" Feb 13 19:49:32.146725 kubelet[2827]: I0213 19:49:32.144901 2827 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:49:32.146725 kubelet[2827]: I0213 19:49:32.145355 2827 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:49:32.146725 kubelet[2827]: I0213 19:49:32.145839 2827 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:49:32.148223 kubelet[2827]: I0213 19:49:32.148180 2827 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:49:32.148486 kubelet[2827]: I0213 19:49:32.148443 2827 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:49:32.153248 kubelet[2827]: I0213 19:49:32.153211 2827 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:49:32.156447 kubelet[2827]: I0213 19:49:32.156413 2827 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:49:32.161561 kubelet[2827]: I0213 19:49:32.156647 2827 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:49:32.161561 kubelet[2827]: E0213 19:49:32.156987 2827 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-134\" not found" Feb 13 19:49:32.161561 kubelet[2827]: W0213 19:49:32.159604 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.25.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:32.161561 kubelet[2827]: E0213 19:49:32.160722 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.25.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:32.161561 kubelet[2827]: E0213 19:49:32.159770 2827 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-134?timeout=10s\": dial tcp 172.31.25.134:6443: connect: connection refused" interval="200ms" Feb 13 19:49:32.161561 kubelet[2827]: I0213 19:49:32.161435 2827 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:49:32.163234 kubelet[2827]: I0213 19:49:32.163142 2827 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:49:32.166904 kubelet[2827]: I0213 19:49:32.166868 2827 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:49:32.167078 kubelet[2827]: I0213 19:49:32.167059 2827 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:49:32.187640 kubelet[2827]: I0213 19:49:32.187581 2827 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:49:32.191029 kubelet[2827]: I0213 19:49:32.190488 2827 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:49:32.191029 kubelet[2827]: I0213 19:49:32.190532 2827 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:49:32.191029 kubelet[2827]: I0213 19:49:32.190564 2827 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:49:32.191029 kubelet[2827]: I0213 19:49:32.190579 2827 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:49:32.191029 kubelet[2827]: E0213 19:49:32.190648 2827 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:49:32.197517 kubelet[2827]: W0213 19:49:32.197446 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.25.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:32.197778 kubelet[2827]: E0213 19:49:32.197745 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.25.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:32.202092 kubelet[2827]: E0213 19:49:32.201999 2827 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:49:32.210498 kubelet[2827]: I0213 19:49:32.210459 2827 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:49:32.210498 kubelet[2827]: I0213 19:49:32.210491 2827 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:49:32.210762 kubelet[2827]: I0213 19:49:32.210525 2827 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:49:32.213843 kubelet[2827]: I0213 19:49:32.213795 2827 policy_none.go:49] "None policy: Start" Feb 13 19:49:32.213843 kubelet[2827]: I0213 19:49:32.213836 2827 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:49:32.214013 kubelet[2827]: I0213 19:49:32.213861 2827 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:49:32.223857 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 19:49:32.244656 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 19:49:32.250825 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 19:49:32.261431 kubelet[2827]: E0213 19:49:32.261374 2827 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-134\" not found" Feb 13 19:49:32.265238 kubelet[2827]: I0213 19:49:32.264302 2827 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:49:32.265238 kubelet[2827]: I0213 19:49:32.264598 2827 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:49:32.265238 kubelet[2827]: I0213 19:49:32.264618 2827 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:49:32.265238 kubelet[2827]: I0213 19:49:32.265109 2827 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:49:32.270237 kubelet[2827]: E0213 19:49:32.269893 2827 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:49:32.270237 kubelet[2827]: E0213 19:49:32.269983 2827 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-25-134\" not found" Feb 13 19:49:32.310841 systemd[1]: Created slice kubepods-burstable-podd9653220b8fd2de0a92a4b39100e1523.slice - libcontainer container kubepods-burstable-podd9653220b8fd2de0a92a4b39100e1523.slice. Feb 13 19:49:32.326735 kubelet[2827]: E0213 19:49:32.326363 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:32.331073 systemd[1]: Created slice kubepods-burstable-pod62c9281e6f849641bb63a6569393ce9f.slice - libcontainer container kubepods-burstable-pod62c9281e6f849641bb63a6569393ce9f.slice. Feb 13 19:49:32.343353 kubelet[2827]: E0213 19:49:32.343296 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:32.348400 systemd[1]: Created slice kubepods-burstable-pod2cb667127e87f8cb2db9a2d44ee9aff3.slice - libcontainer container kubepods-burstable-pod2cb667127e87f8cb2db9a2d44ee9aff3.slice. Feb 13 19:49:32.352423 kubelet[2827]: E0213 19:49:32.352362 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:32.361778 kubelet[2827]: E0213 19:49:32.361665 2827 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-134?timeout=10s\": dial tcp 172.31.25.134:6443: connect: connection refused" interval="400ms" Feb 13 19:49:32.362844 kubelet[2827]: I0213 19:49:32.362754 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d9653220b8fd2de0a92a4b39100e1523-ca-certs\") pod \"kube-apiserver-ip-172-31-25-134\" (UID: \"d9653220b8fd2de0a92a4b39100e1523\") " pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:32.362844 kubelet[2827]: I0213 19:49:32.362806 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d9653220b8fd2de0a92a4b39100e1523-k8s-certs\") pod \"kube-apiserver-ip-172-31-25-134\" (UID: \"d9653220b8fd2de0a92a4b39100e1523\") " pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:32.363082 kubelet[2827]: I0213 19:49:32.362844 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-ca-certs\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:32.363082 kubelet[2827]: I0213 19:49:32.362883 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-k8s-certs\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:32.363082 kubelet[2827]: I0213 19:49:32.362924 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d9653220b8fd2de0a92a4b39100e1523-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-25-134\" (UID: \"d9653220b8fd2de0a92a4b39100e1523\") " pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:32.363082 kubelet[2827]: I0213 19:49:32.362975 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:32.363082 kubelet[2827]: I0213 19:49:32.363019 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-kubeconfig\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:32.363362 kubelet[2827]: I0213 19:49:32.363055 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:32.363362 kubelet[2827]: I0213 19:49:32.363094 2827 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2cb667127e87f8cb2db9a2d44ee9aff3-kubeconfig\") pod \"kube-scheduler-ip-172-31-25-134\" (UID: \"2cb667127e87f8cb2db9a2d44ee9aff3\") " pod="kube-system/kube-scheduler-ip-172-31-25-134" Feb 13 19:49:32.366886 kubelet[2827]: I0213 19:49:32.366834 2827 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-134" Feb 13 19:49:32.367463 kubelet[2827]: E0213 19:49:32.367420 2827 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.134:6443/api/v1/nodes\": dial tcp 172.31.25.134:6443: connect: connection refused" node="ip-172-31-25-134" Feb 13 19:49:32.569823 kubelet[2827]: I0213 19:49:32.569509 2827 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-134" Feb 13 19:49:32.570144 kubelet[2827]: E0213 19:49:32.570032 2827 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.134:6443/api/v1/nodes\": dial tcp 172.31.25.134:6443: connect: connection refused" node="ip-172-31-25-134" Feb 13 19:49:32.628818 containerd[2031]: time="2025-02-13T19:49:32.628733413Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-25-134,Uid:d9653220b8fd2de0a92a4b39100e1523,Namespace:kube-system,Attempt:0,}" Feb 13 19:49:32.645446 containerd[2031]: time="2025-02-13T19:49:32.645265069Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-25-134,Uid:62c9281e6f849641bb63a6569393ce9f,Namespace:kube-system,Attempt:0,}" Feb 13 19:49:32.654848 containerd[2031]: time="2025-02-13T19:49:32.654351445Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-25-134,Uid:2cb667127e87f8cb2db9a2d44ee9aff3,Namespace:kube-system,Attempt:0,}" Feb 13 19:49:32.763308 kubelet[2827]: E0213 19:49:32.763256 2827 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-134?timeout=10s\": dial tcp 172.31.25.134:6443: connect: connection refused" interval="800ms" Feb 13 19:49:32.973185 kubelet[2827]: I0213 19:49:32.973089 2827 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-134" Feb 13 19:49:32.974270 kubelet[2827]: E0213 19:49:32.973609 2827 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.134:6443/api/v1/nodes\": dial tcp 172.31.25.134:6443: connect: connection refused" node="ip-172-31-25-134" Feb 13 19:49:33.073862 kubelet[2827]: W0213 19:49:33.073794 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.25.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-134&limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:33.074142 kubelet[2827]: E0213 19:49:33.074080 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.25.134:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-134&limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:33.152951 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1843670303.mount: Deactivated successfully. Feb 13 19:49:33.272053 containerd[2031]: time="2025-02-13T19:49:33.270814704Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:49:33.272924 containerd[2031]: time="2025-02-13T19:49:33.272858076Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:49:33.273960 containerd[2031]: time="2025-02-13T19:49:33.273810612Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Feb 13 19:49:33.274834 containerd[2031]: time="2025-02-13T19:49:33.274607712Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:49:33.276237 containerd[2031]: time="2025-02-13T19:49:33.276102205Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:49:33.276349 containerd[2031]: time="2025-02-13T19:49:33.276294445Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:49:33.277512 containerd[2031]: time="2025-02-13T19:49:33.277454413Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:49:33.285005 containerd[2031]: time="2025-02-13T19:49:33.284879293Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:49:33.289344 containerd[2031]: time="2025-02-13T19:49:33.288661153Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 659.801284ms" Feb 13 19:49:33.293343 containerd[2031]: time="2025-02-13T19:49:33.293273281Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 647.898964ms" Feb 13 19:49:33.295463 containerd[2031]: time="2025-02-13T19:49:33.295346641Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 640.890028ms" Feb 13 19:49:33.344637 kubelet[2827]: W0213 19:49:33.344296 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.25.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:33.344637 kubelet[2827]: E0213 19:49:33.344371 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.25.134:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:33.479932 containerd[2031]: time="2025-02-13T19:49:33.479167862Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:49:33.479932 containerd[2031]: time="2025-02-13T19:49:33.479288270Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:49:33.479932 containerd[2031]: time="2025-02-13T19:49:33.479318186Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:33.479932 containerd[2031]: time="2025-02-13T19:49:33.479469086Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:33.486059 containerd[2031]: time="2025-02-13T19:49:33.485438546Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:49:33.486059 containerd[2031]: time="2025-02-13T19:49:33.485554058Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:49:33.486059 containerd[2031]: time="2025-02-13T19:49:33.485591438Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:33.486059 containerd[2031]: time="2025-02-13T19:49:33.485767550Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:33.488983 containerd[2031]: time="2025-02-13T19:49:33.488658854Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:49:33.488983 containerd[2031]: time="2025-02-13T19:49:33.488771570Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:49:33.488983 containerd[2031]: time="2025-02-13T19:49:33.488797754Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:33.489391 containerd[2031]: time="2025-02-13T19:49:33.488932742Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:33.533281 systemd[1]: Started cri-containerd-2ccaf8e9ee13bf78cb12468771597afbb4be4de2b8826db1fc90b98dc7a76d15.scope - libcontainer container 2ccaf8e9ee13bf78cb12468771597afbb4be4de2b8826db1fc90b98dc7a76d15. Feb 13 19:49:33.550050 systemd[1]: Started cri-containerd-450ec195605da19dd4f9f3b322a18b31602f3c90ac971390a678b12f204aac82.scope - libcontainer container 450ec195605da19dd4f9f3b322a18b31602f3c90ac971390a678b12f204aac82. Feb 13 19:49:33.553633 systemd[1]: Started cri-containerd-c6475971ecc09d0ddb241f7a4290493070613d48637dd00a9864d72ae8bbf546.scope - libcontainer container c6475971ecc09d0ddb241f7a4290493070613d48637dd00a9864d72ae8bbf546. Feb 13 19:49:33.565249 kubelet[2827]: E0213 19:49:33.564966 2827 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-134?timeout=10s\": dial tcp 172.31.25.134:6443: connect: connection refused" interval="1.6s" Feb 13 19:49:33.566556 kubelet[2827]: W0213 19:49:33.566304 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.25.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:33.566556 kubelet[2827]: E0213 19:49:33.566412 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.25.134:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:33.654375 containerd[2031]: time="2025-02-13T19:49:33.654298718Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-25-134,Uid:62c9281e6f849641bb63a6569393ce9f,Namespace:kube-system,Attempt:0,} returns sandbox id \"c6475971ecc09d0ddb241f7a4290493070613d48637dd00a9864d72ae8bbf546\"" Feb 13 19:49:33.678265 containerd[2031]: time="2025-02-13T19:49:33.677022086Z" level=info msg="CreateContainer within sandbox \"c6475971ecc09d0ddb241f7a4290493070613d48637dd00a9864d72ae8bbf546\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 19:49:33.683795 containerd[2031]: time="2025-02-13T19:49:33.683723643Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-25-134,Uid:d9653220b8fd2de0a92a4b39100e1523,Namespace:kube-system,Attempt:0,} returns sandbox id \"2ccaf8e9ee13bf78cb12468771597afbb4be4de2b8826db1fc90b98dc7a76d15\"" Feb 13 19:49:33.691163 containerd[2031]: time="2025-02-13T19:49:33.691111359Z" level=info msg="CreateContainer within sandbox \"2ccaf8e9ee13bf78cb12468771597afbb4be4de2b8826db1fc90b98dc7a76d15\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 19:49:33.701553 containerd[2031]: time="2025-02-13T19:49:33.701474739Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-25-134,Uid:2cb667127e87f8cb2db9a2d44ee9aff3,Namespace:kube-system,Attempt:0,} returns sandbox id \"450ec195605da19dd4f9f3b322a18b31602f3c90ac971390a678b12f204aac82\"" Feb 13 19:49:33.710383 containerd[2031]: time="2025-02-13T19:49:33.710325903Z" level=info msg="CreateContainer within sandbox \"450ec195605da19dd4f9f3b322a18b31602f3c90ac971390a678b12f204aac82\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 19:49:33.714666 containerd[2031]: time="2025-02-13T19:49:33.714607251Z" level=info msg="CreateContainer within sandbox \"c6475971ecc09d0ddb241f7a4290493070613d48637dd00a9864d72ae8bbf546\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f\"" Feb 13 19:49:33.716369 containerd[2031]: time="2025-02-13T19:49:33.716106291Z" level=info msg="StartContainer for \"8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f\"" Feb 13 19:49:33.721539 containerd[2031]: time="2025-02-13T19:49:33.721470687Z" level=info msg="CreateContainer within sandbox \"2ccaf8e9ee13bf78cb12468771597afbb4be4de2b8826db1fc90b98dc7a76d15\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"53ea088cead746a28cd916b7707435dbd65d1d7923241d274464b990c961cddb\"" Feb 13 19:49:33.725063 kubelet[2827]: W0213 19:49:33.724862 2827 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.25.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.25.134:6443: connect: connection refused Feb 13 19:49:33.725063 kubelet[2827]: E0213 19:49:33.724982 2827 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.25.134:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.25.134:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:49:33.725706 containerd[2031]: time="2025-02-13T19:49:33.725331747Z" level=info msg="StartContainer for \"53ea088cead746a28cd916b7707435dbd65d1d7923241d274464b990c961cddb\"" Feb 13 19:49:33.737409 containerd[2031]: time="2025-02-13T19:49:33.737337699Z" level=info msg="CreateContainer within sandbox \"450ec195605da19dd4f9f3b322a18b31602f3c90ac971390a678b12f204aac82\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da\"" Feb 13 19:49:33.739133 containerd[2031]: time="2025-02-13T19:49:33.739077171Z" level=info msg="StartContainer for \"4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da\"" Feb 13 19:49:33.775517 systemd[1]: Started cri-containerd-8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f.scope - libcontainer container 8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f. Feb 13 19:49:33.778502 kubelet[2827]: I0213 19:49:33.778465 2827 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-134" Feb 13 19:49:33.780291 kubelet[2827]: E0213 19:49:33.780220 2827 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.134:6443/api/v1/nodes\": dial tcp 172.31.25.134:6443: connect: connection refused" node="ip-172-31-25-134" Feb 13 19:49:33.823104 systemd[1]: Started cri-containerd-53ea088cead746a28cd916b7707435dbd65d1d7923241d274464b990c961cddb.scope - libcontainer container 53ea088cead746a28cd916b7707435dbd65d1d7923241d274464b990c961cddb. Feb 13 19:49:33.844201 systemd[1]: Started cri-containerd-4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da.scope - libcontainer container 4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da. Feb 13 19:49:33.927469 containerd[2031]: time="2025-02-13T19:49:33.927394708Z" level=info msg="StartContainer for \"8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f\" returns successfully" Feb 13 19:49:33.941273 containerd[2031]: time="2025-02-13T19:49:33.941164888Z" level=info msg="StartContainer for \"53ea088cead746a28cd916b7707435dbd65d1d7923241d274464b990c961cddb\" returns successfully" Feb 13 19:49:33.984857 containerd[2031]: time="2025-02-13T19:49:33.984062656Z" level=info msg="StartContainer for \"4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da\" returns successfully" Feb 13 19:49:34.218372 kubelet[2827]: E0213 19:49:34.218236 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:34.225343 kubelet[2827]: E0213 19:49:34.225282 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:34.232121 kubelet[2827]: E0213 19:49:34.232086 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:35.232747 kubelet[2827]: E0213 19:49:35.232066 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:35.234307 kubelet[2827]: E0213 19:49:35.233885 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:35.384710 kubelet[2827]: I0213 19:49:35.384024 2827 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-134" Feb 13 19:49:36.235737 kubelet[2827]: E0213 19:49:36.235509 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:36.291636 kubelet[2827]: E0213 19:49:36.291225 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:37.240121 kubelet[2827]: E0213 19:49:37.239192 2827 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:37.519374 kubelet[2827]: E0213 19:49:37.519094 2827 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-25-134\" not found" node="ip-172-31-25-134" Feb 13 19:49:37.591045 kubelet[2827]: I0213 19:49:37.590644 2827 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-25-134" Feb 13 19:49:37.631914 kubelet[2827]: E0213 19:49:37.631762 2827 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ip-172-31-25-134.1823dc5cc7559807 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-25-134,UID:ip-172-31-25-134,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-25-134,},FirstTimestamp:2025-02-13 19:49:32.134520839 +0000 UTC m=+1.024298622,LastTimestamp:2025-02-13 19:49:32.134520839 +0000 UTC m=+1.024298622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-25-134,}" Feb 13 19:49:37.657954 kubelet[2827]: I0213 19:49:37.657862 2827 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:37.678907 kubelet[2827]: E0213 19:49:37.678838 2827 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-25-134\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:37.678907 kubelet[2827]: I0213 19:49:37.678887 2827 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:37.689060 kubelet[2827]: E0213 19:49:37.689001 2827 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-25-134\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:37.689060 kubelet[2827]: I0213 19:49:37.689052 2827 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-25-134" Feb 13 19:49:37.709497 kubelet[2827]: E0213 19:49:37.709444 2827 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-25-134\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-25-134" Feb 13 19:49:37.722733 kubelet[2827]: E0213 19:49:37.722566 2827 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ip-172-31-25-134.1823dc5ccb5ae28f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-25-134,UID:ip-172-31-25-134,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:ip-172-31-25-134,},FirstTimestamp:2025-02-13 19:49:32.201976463 +0000 UTC m=+1.091754234,LastTimestamp:2025-02-13 19:49:32.201976463 +0000 UTC m=+1.091754234,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-25-134,}" Feb 13 19:49:37.793251 kubelet[2827]: E0213 19:49:37.793001 2827 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ip-172-31-25-134.1823dc5ccbcac383 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-25-134,UID:ip-172-31-25-134,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node ip-172-31-25-134 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:ip-172-31-25-134,},FirstTimestamp:2025-02-13 19:49:32.209308547 +0000 UTC m=+1.099086294,LastTimestamp:2025-02-13 19:49:32.209308547 +0000 UTC m=+1.099086294,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-25-134,}" Feb 13 19:49:38.125415 kubelet[2827]: I0213 19:49:38.125353 2827 apiserver.go:52] "Watching apiserver" Feb 13 19:49:38.161501 kubelet[2827]: I0213 19:49:38.161430 2827 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:49:38.229762 kubelet[2827]: I0213 19:49:38.229710 2827 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:38.864835 update_engine[2008]: I20250213 19:49:38.864738 2008 update_attempter.cc:509] Updating boot flags... Feb 13 19:49:39.002412 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (3118) Feb 13 19:49:39.368738 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (3119) Feb 13 19:49:39.656849 systemd[1]: Reloading requested from client PID 3288 ('systemctl') (unit session-7.scope)... Feb 13 19:49:39.657460 systemd[1]: Reloading... Feb 13 19:49:39.748735 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 40 scanned by (udev-worker) (3119) Feb 13 19:49:39.946067 zram_generator::config[3385]: No configuration found. Feb 13 19:49:40.247326 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:49:40.448693 systemd[1]: Reloading finished in 790 ms. Feb 13 19:49:40.578181 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:49:40.610185 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:49:40.611833 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:40.612114 systemd[1]: kubelet.service: Consumed 1.779s CPU time, 125.4M memory peak, 0B memory swap peak. Feb 13 19:49:40.625326 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:49:40.920531 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:49:40.938353 (kubelet)[3472]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:49:41.032946 kubelet[3472]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:49:41.032946 kubelet[3472]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:49:41.032946 kubelet[3472]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:49:41.033567 kubelet[3472]: I0213 19:49:41.033481 3472 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:49:41.047664 kubelet[3472]: I0213 19:49:41.047589 3472 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:49:41.047664 kubelet[3472]: I0213 19:49:41.047644 3472 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:49:41.049249 kubelet[3472]: I0213 19:49:41.048150 3472 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:49:41.050747 kubelet[3472]: I0213 19:49:41.050513 3472 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 19:49:41.055634 kubelet[3472]: I0213 19:49:41.055392 3472 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:49:41.065588 kubelet[3472]: E0213 19:49:41.065534 3472 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:49:41.066568 kubelet[3472]: I0213 19:49:41.065774 3472 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:49:41.070629 kubelet[3472]: I0213 19:49:41.070530 3472 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:49:41.071014 kubelet[3472]: I0213 19:49:41.070963 3472 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:49:41.072935 kubelet[3472]: I0213 19:49:41.071015 3472 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-25-134","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:49:41.073170 kubelet[3472]: I0213 19:49:41.072952 3472 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:49:41.073170 kubelet[3472]: I0213 19:49:41.072976 3472 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:49:41.073170 kubelet[3472]: I0213 19:49:41.073064 3472 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:49:41.073854 kubelet[3472]: I0213 19:49:41.073818 3472 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:49:41.073986 kubelet[3472]: I0213 19:49:41.073860 3472 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:49:41.073986 kubelet[3472]: I0213 19:49:41.073898 3472 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:49:41.073986 kubelet[3472]: I0213 19:49:41.073920 3472 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:49:41.079826 kubelet[3472]: I0213 19:49:41.079766 3472 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Feb 13 19:49:41.080546 kubelet[3472]: I0213 19:49:41.080500 3472 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:49:41.083512 kubelet[3472]: I0213 19:49:41.083452 3472 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:49:41.083512 kubelet[3472]: I0213 19:49:41.083516 3472 server.go:1287] "Started kubelet" Feb 13 19:49:41.090176 kubelet[3472]: I0213 19:49:41.090047 3472 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:49:41.090906 kubelet[3472]: I0213 19:49:41.090856 3472 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:49:41.093772 kubelet[3472]: I0213 19:49:41.093737 3472 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:49:41.095661 kubelet[3472]: I0213 19:49:41.095042 3472 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:49:41.102420 kubelet[3472]: I0213 19:49:41.102367 3472 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:49:41.104935 kubelet[3472]: I0213 19:49:41.104900 3472 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:49:41.105790 kubelet[3472]: E0213 19:49:41.105755 3472 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-134\" not found" Feb 13 19:49:41.108415 kubelet[3472]: I0213 19:49:41.106393 3472 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:49:41.108415 kubelet[3472]: I0213 19:49:41.106620 3472 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:49:41.111710 kubelet[3472]: I0213 19:49:41.109553 3472 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:49:41.126846 kubelet[3472]: I0213 19:49:41.126797 3472 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:49:41.134391 kubelet[3472]: I0213 19:49:41.133879 3472 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:49:41.134391 kubelet[3472]: I0213 19:49:41.133925 3472 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:49:41.134391 kubelet[3472]: I0213 19:49:41.133966 3472 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:49:41.134391 kubelet[3472]: I0213 19:49:41.133980 3472 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:49:41.134391 kubelet[3472]: E0213 19:49:41.134057 3472 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:49:41.161766 kubelet[3472]: I0213 19:49:41.158469 3472 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:49:41.211396 kubelet[3472]: I0213 19:49:41.210995 3472 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:49:41.213713 kubelet[3472]: I0213 19:49:41.211665 3472 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:49:41.214855 kubelet[3472]: E0213 19:49:41.214819 3472 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-134\" not found" Feb 13 19:49:41.220267 kubelet[3472]: E0213 19:49:41.220224 3472 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:49:41.236421 kubelet[3472]: E0213 19:49:41.235781 3472 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 13 19:49:41.360040 kubelet[3472]: I0213 19:49:41.358840 3472 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:49:41.360040 kubelet[3472]: I0213 19:49:41.359750 3472 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:49:41.360040 kubelet[3472]: I0213 19:49:41.359792 3472 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:49:41.360395 kubelet[3472]: I0213 19:49:41.360146 3472 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 19:49:41.360395 kubelet[3472]: I0213 19:49:41.360199 3472 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 19:49:41.360395 kubelet[3472]: I0213 19:49:41.360238 3472 policy_none.go:49] "None policy: Start" Feb 13 19:49:41.360395 kubelet[3472]: I0213 19:49:41.360288 3472 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:49:41.360395 kubelet[3472]: I0213 19:49:41.360313 3472 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:49:41.360710 kubelet[3472]: I0213 19:49:41.360592 3472 state_mem.go:75] "Updated machine memory state" Feb 13 19:49:41.372349 kubelet[3472]: I0213 19:49:41.372297 3472 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:49:41.373230 kubelet[3472]: I0213 19:49:41.372583 3472 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:49:41.373230 kubelet[3472]: I0213 19:49:41.372613 3472 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:49:41.376102 kubelet[3472]: I0213 19:49:41.376059 3472 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:49:41.383329 kubelet[3472]: E0213 19:49:41.382652 3472 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:49:41.436768 kubelet[3472]: I0213 19:49:41.436726 3472 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-25-134" Feb 13 19:49:41.438383 kubelet[3472]: I0213 19:49:41.437561 3472 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:41.438383 kubelet[3472]: I0213 19:49:41.438076 3472 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:41.456047 kubelet[3472]: E0213 19:49:41.455734 3472 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-25-134\" already exists" pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:41.509036 kubelet[3472]: I0213 19:49:41.508711 3472 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-134" Feb 13 19:49:41.514895 kubelet[3472]: I0213 19:49:41.513881 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:41.514895 kubelet[3472]: I0213 19:49:41.513953 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d9653220b8fd2de0a92a4b39100e1523-k8s-certs\") pod \"kube-apiserver-ip-172-31-25-134\" (UID: \"d9653220b8fd2de0a92a4b39100e1523\") " pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:41.514895 kubelet[3472]: I0213 19:49:41.513992 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-k8s-certs\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:41.514895 kubelet[3472]: I0213 19:49:41.514028 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-kubeconfig\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:41.514895 kubelet[3472]: I0213 19:49:41.514069 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2cb667127e87f8cb2db9a2d44ee9aff3-kubeconfig\") pod \"kube-scheduler-ip-172-31-25-134\" (UID: \"2cb667127e87f8cb2db9a2d44ee9aff3\") " pod="kube-system/kube-scheduler-ip-172-31-25-134" Feb 13 19:49:41.515299 kubelet[3472]: I0213 19:49:41.514105 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d9653220b8fd2de0a92a4b39100e1523-ca-certs\") pod \"kube-apiserver-ip-172-31-25-134\" (UID: \"d9653220b8fd2de0a92a4b39100e1523\") " pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:41.515299 kubelet[3472]: I0213 19:49:41.514142 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d9653220b8fd2de0a92a4b39100e1523-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-25-134\" (UID: \"d9653220b8fd2de0a92a4b39100e1523\") " pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:41.515299 kubelet[3472]: I0213 19:49:41.514185 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-ca-certs\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:41.515299 kubelet[3472]: I0213 19:49:41.514221 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/62c9281e6f849641bb63a6569393ce9f-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-25-134\" (UID: \"62c9281e6f849641bb63a6569393ce9f\") " pod="kube-system/kube-controller-manager-ip-172-31-25-134" Feb 13 19:49:41.535810 kubelet[3472]: I0213 19:49:41.534582 3472 kubelet_node_status.go:125] "Node was previously registered" node="ip-172-31-25-134" Feb 13 19:49:41.535810 kubelet[3472]: I0213 19:49:41.534745 3472 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-25-134" Feb 13 19:49:42.076103 kubelet[3472]: I0213 19:49:42.075300 3472 apiserver.go:52] "Watching apiserver" Feb 13 19:49:42.107088 kubelet[3472]: I0213 19:49:42.106943 3472 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:49:42.289533 kubelet[3472]: I0213 19:49:42.289466 3472 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:42.305234 kubelet[3472]: E0213 19:49:42.305041 3472 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-25-134\" already exists" pod="kube-system/kube-apiserver-ip-172-31-25-134" Feb 13 19:49:42.362561 kubelet[3472]: I0213 19:49:42.362461 3472 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-25-134" podStartSLOduration=1.362437798 podStartE2EDuration="1.362437798s" podCreationTimestamp="2025-02-13 19:49:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:49:42.36018589 +0000 UTC m=+1.414149416" watchObservedRunningTime="2025-02-13 19:49:42.362437798 +0000 UTC m=+1.416401312" Feb 13 19:49:42.362817 kubelet[3472]: I0213 19:49:42.362700 3472 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-25-134" podStartSLOduration=1.3626695899999999 podStartE2EDuration="1.36266959s" podCreationTimestamp="2025-02-13 19:49:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:49:42.341613622 +0000 UTC m=+1.395577136" watchObservedRunningTime="2025-02-13 19:49:42.36266959 +0000 UTC m=+1.416633092" Feb 13 19:49:42.406079 kubelet[3472]: I0213 19:49:42.405759 3472 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-25-134" podStartSLOduration=4.405739042 podStartE2EDuration="4.405739042s" podCreationTimestamp="2025-02-13 19:49:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:49:42.389528758 +0000 UTC m=+1.443492260" watchObservedRunningTime="2025-02-13 19:49:42.405739042 +0000 UTC m=+1.459702556" Feb 13 19:49:42.709112 sudo[2329]: pam_unix(sudo:session): session closed for user root Feb 13 19:49:42.733482 sshd[2314]: pam_unix(sshd:session): session closed for user core Feb 13 19:49:42.740110 systemd-logind[2004]: Session 7 logged out. Waiting for processes to exit. Feb 13 19:49:42.741727 systemd[1]: sshd@6-172.31.25.134:22-139.178.89.65:55220.service: Deactivated successfully. Feb 13 19:49:42.746903 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 19:49:42.748206 systemd[1]: session-7.scope: Consumed 9.480s CPU time, 155.1M memory peak, 0B memory swap peak. Feb 13 19:49:42.750659 systemd-logind[2004]: Removed session 7. Feb 13 19:49:45.849520 kubelet[3472]: I0213 19:49:45.849445 3472 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 19:49:45.850650 containerd[2031]: time="2025-02-13T19:49:45.850492119Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 19:49:45.851218 kubelet[3472]: I0213 19:49:45.850861 3472 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 19:49:46.825929 systemd[1]: Created slice kubepods-besteffort-pod8d40e832_4b64_4224_8756_62e708cde6df.slice - libcontainer container kubepods-besteffort-pod8d40e832_4b64_4224_8756_62e708cde6df.slice. Feb 13 19:49:46.849855 kubelet[3472]: I0213 19:49:46.849575 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/8d40e832-4b64-4224-8756-62e708cde6df-kube-proxy\") pod \"kube-proxy-4h77j\" (UID: \"8d40e832-4b64-4224-8756-62e708cde6df\") " pod="kube-system/kube-proxy-4h77j" Feb 13 19:49:46.849855 kubelet[3472]: I0213 19:49:46.849635 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8d40e832-4b64-4224-8756-62e708cde6df-lib-modules\") pod \"kube-proxy-4h77j\" (UID: \"8d40e832-4b64-4224-8756-62e708cde6df\") " pod="kube-system/kube-proxy-4h77j" Feb 13 19:49:46.849855 kubelet[3472]: I0213 19:49:46.849694 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8d40e832-4b64-4224-8756-62e708cde6df-xtables-lock\") pod \"kube-proxy-4h77j\" (UID: \"8d40e832-4b64-4224-8756-62e708cde6df\") " pod="kube-system/kube-proxy-4h77j" Feb 13 19:49:46.849855 kubelet[3472]: I0213 19:49:46.849741 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56l4f\" (UniqueName: \"kubernetes.io/projected/8d40e832-4b64-4224-8756-62e708cde6df-kube-api-access-56l4f\") pod \"kube-proxy-4h77j\" (UID: \"8d40e832-4b64-4224-8756-62e708cde6df\") " pod="kube-system/kube-proxy-4h77j" Feb 13 19:49:46.866734 kubelet[3472]: W0213 19:49:46.866592 3472 reflector.go:569] object-"kube-flannel"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ip-172-31-25-134" cannot list resource "configmaps" in API group "" in the namespace "kube-flannel": no relationship found between node 'ip-172-31-25-134' and this object Feb 13 19:49:46.866734 kubelet[3472]: E0213 19:49:46.866662 3472 reflector.go:166] "Unhandled Error" err="object-\"kube-flannel\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:ip-172-31-25-134\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-flannel\": no relationship found between node 'ip-172-31-25-134' and this object" logger="UnhandledError" Feb 13 19:49:46.867187 systemd[1]: Created slice kubepods-burstable-podacec9f18_3d83_43c1_8a36_8c78143cfe25.slice - libcontainer container kubepods-burstable-podacec9f18_3d83_43c1_8a36_8c78143cfe25.slice. Feb 13 19:49:46.872283 kubelet[3472]: W0213 19:49:46.872177 3472 reflector.go:569] object-"kube-flannel"/"kube-flannel-cfg": failed to list *v1.ConfigMap: configmaps "kube-flannel-cfg" is forbidden: User "system:node:ip-172-31-25-134" cannot list resource "configmaps" in API group "" in the namespace "kube-flannel": no relationship found between node 'ip-172-31-25-134' and this object Feb 13 19:49:46.872283 kubelet[3472]: E0213 19:49:46.872238 3472 reflector.go:166] "Unhandled Error" err="object-\"kube-flannel\"/\"kube-flannel-cfg\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-flannel-cfg\" is forbidden: User \"system:node:ip-172-31-25-134\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-flannel\": no relationship found between node 'ip-172-31-25-134' and this object" logger="UnhandledError" Feb 13 19:49:46.950388 kubelet[3472]: I0213 19:49:46.950270 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/acec9f18-3d83-43c1-8a36-8c78143cfe25-run\") pod \"kube-flannel-ds-qgxkx\" (UID: \"acec9f18-3d83-43c1-8a36-8c78143cfe25\") " pod="kube-flannel/kube-flannel-ds-qgxkx" Feb 13 19:49:46.951920 kubelet[3472]: I0213 19:49:46.950908 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/acec9f18-3d83-43c1-8a36-8c78143cfe25-cni-plugin\") pod \"kube-flannel-ds-qgxkx\" (UID: \"acec9f18-3d83-43c1-8a36-8c78143cfe25\") " pod="kube-flannel/kube-flannel-ds-qgxkx" Feb 13 19:49:46.951920 kubelet[3472]: I0213 19:49:46.950959 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/acec9f18-3d83-43c1-8a36-8c78143cfe25-flannel-cfg\") pod \"kube-flannel-ds-qgxkx\" (UID: \"acec9f18-3d83-43c1-8a36-8c78143cfe25\") " pod="kube-flannel/kube-flannel-ds-qgxkx" Feb 13 19:49:46.951920 kubelet[3472]: I0213 19:49:46.950997 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c55vj\" (UniqueName: \"kubernetes.io/projected/acec9f18-3d83-43c1-8a36-8c78143cfe25-kube-api-access-c55vj\") pod \"kube-flannel-ds-qgxkx\" (UID: \"acec9f18-3d83-43c1-8a36-8c78143cfe25\") " pod="kube-flannel/kube-flannel-ds-qgxkx" Feb 13 19:49:46.951920 kubelet[3472]: I0213 19:49:46.951063 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/acec9f18-3d83-43c1-8a36-8c78143cfe25-cni\") pod \"kube-flannel-ds-qgxkx\" (UID: \"acec9f18-3d83-43c1-8a36-8c78143cfe25\") " pod="kube-flannel/kube-flannel-ds-qgxkx" Feb 13 19:49:46.951920 kubelet[3472]: I0213 19:49:46.951135 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/acec9f18-3d83-43c1-8a36-8c78143cfe25-xtables-lock\") pod \"kube-flannel-ds-qgxkx\" (UID: \"acec9f18-3d83-43c1-8a36-8c78143cfe25\") " pod="kube-flannel/kube-flannel-ds-qgxkx" Feb 13 19:49:47.141078 containerd[2031]: time="2025-02-13T19:49:47.140998657Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4h77j,Uid:8d40e832-4b64-4224-8756-62e708cde6df,Namespace:kube-system,Attempt:0,}" Feb 13 19:49:47.178500 containerd[2031]: time="2025-02-13T19:49:47.178327562Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:49:47.178916 containerd[2031]: time="2025-02-13T19:49:47.178837682Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:49:47.179001 containerd[2031]: time="2025-02-13T19:49:47.178965938Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:47.179254 containerd[2031]: time="2025-02-13T19:49:47.179188646Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:47.218461 systemd[1]: run-containerd-runc-k8s.io-e6b282b2008ae5923835935cff8faf3441fb4d12145ae632de91dcdad0e75ba8-runc.nxVLjv.mount: Deactivated successfully. Feb 13 19:49:47.232014 systemd[1]: Started cri-containerd-e6b282b2008ae5923835935cff8faf3441fb4d12145ae632de91dcdad0e75ba8.scope - libcontainer container e6b282b2008ae5923835935cff8faf3441fb4d12145ae632de91dcdad0e75ba8. Feb 13 19:49:47.272803 containerd[2031]: time="2025-02-13T19:49:47.272745818Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-4h77j,Uid:8d40e832-4b64-4224-8756-62e708cde6df,Namespace:kube-system,Attempt:0,} returns sandbox id \"e6b282b2008ae5923835935cff8faf3441fb4d12145ae632de91dcdad0e75ba8\"" Feb 13 19:49:47.281221 containerd[2031]: time="2025-02-13T19:49:47.281156534Z" level=info msg="CreateContainer within sandbox \"e6b282b2008ae5923835935cff8faf3441fb4d12145ae632de91dcdad0e75ba8\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 19:49:47.301889 containerd[2031]: time="2025-02-13T19:49:47.301300238Z" level=info msg="CreateContainer within sandbox \"e6b282b2008ae5923835935cff8faf3441fb4d12145ae632de91dcdad0e75ba8\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4cc075a5dce2d29b5d4f1a672db10d7905285aa3c39cc2b456b2c3ce97c6c7ed\"" Feb 13 19:49:47.302459 containerd[2031]: time="2025-02-13T19:49:47.302413190Z" level=info msg="StartContainer for \"4cc075a5dce2d29b5d4f1a672db10d7905285aa3c39cc2b456b2c3ce97c6c7ed\"" Feb 13 19:49:47.356008 systemd[1]: Started cri-containerd-4cc075a5dce2d29b5d4f1a672db10d7905285aa3c39cc2b456b2c3ce97c6c7ed.scope - libcontainer container 4cc075a5dce2d29b5d4f1a672db10d7905285aa3c39cc2b456b2c3ce97c6c7ed. Feb 13 19:49:47.407307 containerd[2031]: time="2025-02-13T19:49:47.406997127Z" level=info msg="StartContainer for \"4cc075a5dce2d29b5d4f1a672db10d7905285aa3c39cc2b456b2c3ce97c6c7ed\" returns successfully" Feb 13 19:49:48.073324 containerd[2031]: time="2025-02-13T19:49:48.073265810Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-qgxkx,Uid:acec9f18-3d83-43c1-8a36-8c78143cfe25,Namespace:kube-flannel,Attempt:0,}" Feb 13 19:49:48.112203 containerd[2031]: time="2025-02-13T19:49:48.112027934Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:49:48.112203 containerd[2031]: time="2025-02-13T19:49:48.112120790Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:49:48.112203 containerd[2031]: time="2025-02-13T19:49:48.112161326Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:48.113306 containerd[2031]: time="2025-02-13T19:49:48.112315478Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:49:48.152647 systemd[1]: Started cri-containerd-017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec.scope - libcontainer container 017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec. Feb 13 19:49:48.217272 containerd[2031]: time="2025-02-13T19:49:48.217207719Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-qgxkx,Uid:acec9f18-3d83-43c1-8a36-8c78143cfe25,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec\"" Feb 13 19:49:48.219955 containerd[2031]: time="2025-02-13T19:49:48.219896523Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\"" Feb 13 19:49:48.321948 kubelet[3472]: I0213 19:49:48.321774 3472 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-4h77j" podStartSLOduration=2.3217478910000002 podStartE2EDuration="2.321747891s" podCreationTimestamp="2025-02-13 19:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:49:48.321098943 +0000 UTC m=+7.375062457" watchObservedRunningTime="2025-02-13 19:49:48.321747891 +0000 UTC m=+7.375711417" Feb 13 19:49:50.392184 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2617187671.mount: Deactivated successfully. Feb 13 19:49:50.438541 containerd[2031]: time="2025-02-13T19:49:50.438322770Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:50.440086 containerd[2031]: time="2025-02-13T19:49:50.440033010Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3673532" Feb 13 19:49:50.440995 containerd[2031]: time="2025-02-13T19:49:50.440909490Z" level=info msg="ImageCreate event name:\"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:50.445350 containerd[2031]: time="2025-02-13T19:49:50.445223718Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:50.448501 containerd[2031]: time="2025-02-13T19:49:50.447536094Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3662650\" in 2.227588463s" Feb 13 19:49:50.448501 containerd[2031]: time="2025-02-13T19:49:50.448418130Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\"" Feb 13 19:49:50.453794 containerd[2031]: time="2025-02-13T19:49:50.453454458Z" level=info msg="CreateContainer within sandbox \"017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}" Feb 13 19:49:50.472899 containerd[2031]: time="2025-02-13T19:49:50.472827150Z" level=info msg="CreateContainer within sandbox \"017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096\"" Feb 13 19:49:50.474730 containerd[2031]: time="2025-02-13T19:49:50.473647542Z" level=info msg="StartContainer for \"88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096\"" Feb 13 19:49:50.518992 systemd[1]: Started cri-containerd-88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096.scope - libcontainer container 88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096. Feb 13 19:49:50.564257 containerd[2031]: time="2025-02-13T19:49:50.564061482Z" level=info msg="StartContainer for \"88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096\" returns successfully" Feb 13 19:49:50.565246 systemd[1]: cri-containerd-88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096.scope: Deactivated successfully. Feb 13 19:49:50.638567 containerd[2031]: time="2025-02-13T19:49:50.638404423Z" level=info msg="shim disconnected" id=88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096 namespace=k8s.io Feb 13 19:49:50.638567 containerd[2031]: time="2025-02-13T19:49:50.638485243Z" level=warning msg="cleaning up after shim disconnected" id=88c0244fc4b5255dc9cbee7a18b4f59edeac133f1efbabea0d0b2051a1568096 namespace=k8s.io Feb 13 19:49:50.638567 containerd[2031]: time="2025-02-13T19:49:50.638505403Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:49:51.319385 containerd[2031]: time="2025-02-13T19:49:51.318492714Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\"" Feb 13 19:49:53.507319 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount659666431.mount: Deactivated successfully. Feb 13 19:49:54.654645 containerd[2031]: time="2025-02-13T19:49:54.654579899Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:54.656773 containerd[2031]: time="2025-02-13T19:49:54.656700791Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26874261" Feb 13 19:49:54.657709 containerd[2031]: time="2025-02-13T19:49:54.657159371Z" level=info msg="ImageCreate event name:\"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:54.663076 containerd[2031]: time="2025-02-13T19:49:54.662978735Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:49:54.666087 containerd[2031]: time="2025-02-13T19:49:54.665890715Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26863435\" in 3.347336825s" Feb 13 19:49:54.666087 containerd[2031]: time="2025-02-13T19:49:54.665947559Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\"" Feb 13 19:49:54.672131 containerd[2031]: time="2025-02-13T19:49:54.672073655Z" level=info msg="CreateContainer within sandbox \"017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Feb 13 19:49:54.693382 containerd[2031]: time="2025-02-13T19:49:54.693179939Z" level=info msg="CreateContainer within sandbox \"017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34\"" Feb 13 19:49:54.695862 containerd[2031]: time="2025-02-13T19:49:54.695777699Z" level=info msg="StartContainer for \"af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34\"" Feb 13 19:49:54.753008 systemd[1]: Started cri-containerd-af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34.scope - libcontainer container af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34. Feb 13 19:49:54.794813 systemd[1]: cri-containerd-af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34.scope: Deactivated successfully. Feb 13 19:49:54.799376 containerd[2031]: time="2025-02-13T19:49:54.799321379Z" level=info msg="StartContainer for \"af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34\" returns successfully" Feb 13 19:49:54.850176 kubelet[3472]: I0213 19:49:54.850114 3472 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Feb 13 19:49:54.950768 systemd[1]: Created slice kubepods-burstable-pod5c26cb20_5ab9_4510_8c96_41e66a4cb07e.slice - libcontainer container kubepods-burstable-pod5c26cb20_5ab9_4510_8c96_41e66a4cb07e.slice. Feb 13 19:49:54.976486 systemd[1]: Created slice kubepods-burstable-podd9e882a4_ba13_47f9_89d3_954cc087e3db.slice - libcontainer container kubepods-burstable-podd9e882a4_ba13_47f9_89d3_954cc087e3db.slice. Feb 13 19:49:54.988818 containerd[2031]: time="2025-02-13T19:49:54.988671456Z" level=info msg="shim disconnected" id=af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34 namespace=k8s.io Feb 13 19:49:54.988818 containerd[2031]: time="2025-02-13T19:49:54.988812348Z" level=warning msg="cleaning up after shim disconnected" id=af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34 namespace=k8s.io Feb 13 19:49:54.989100 containerd[2031]: time="2025-02-13T19:49:54.988835100Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:49:55.012928 kubelet[3472]: I0213 19:49:55.012798 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c26cb20-5ab9-4510-8c96-41e66a4cb07e-config-volume\") pod \"coredns-668d6bf9bc-fzmgc\" (UID: \"5c26cb20-5ab9-4510-8c96-41e66a4cb07e\") " pod="kube-system/coredns-668d6bf9bc-fzmgc" Feb 13 19:49:55.012928 kubelet[3472]: I0213 19:49:55.012879 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqmmw\" (UniqueName: \"kubernetes.io/projected/5c26cb20-5ab9-4510-8c96-41e66a4cb07e-kube-api-access-mqmmw\") pod \"coredns-668d6bf9bc-fzmgc\" (UID: \"5c26cb20-5ab9-4510-8c96-41e66a4cb07e\") " pod="kube-system/coredns-668d6bf9bc-fzmgc" Feb 13 19:49:55.012928 kubelet[3472]: I0213 19:49:55.012928 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5w8j\" (UniqueName: \"kubernetes.io/projected/d9e882a4-ba13-47f9-89d3-954cc087e3db-kube-api-access-m5w8j\") pod \"coredns-668d6bf9bc-tqz6p\" (UID: \"d9e882a4-ba13-47f9-89d3-954cc087e3db\") " pod="kube-system/coredns-668d6bf9bc-tqz6p" Feb 13 19:49:55.013327 kubelet[3472]: I0213 19:49:55.012975 3472 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9e882a4-ba13-47f9-89d3-954cc087e3db-config-volume\") pod \"coredns-668d6bf9bc-tqz6p\" (UID: \"d9e882a4-ba13-47f9-89d3-954cc087e3db\") " pod="kube-system/coredns-668d6bf9bc-tqz6p" Feb 13 19:49:55.263838 containerd[2031]: time="2025-02-13T19:49:55.263390326Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fzmgc,Uid:5c26cb20-5ab9-4510-8c96-41e66a4cb07e,Namespace:kube-system,Attempt:0,}" Feb 13 19:49:55.291481 containerd[2031]: time="2025-02-13T19:49:55.291139042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-tqz6p,Uid:d9e882a4-ba13-47f9-89d3-954cc087e3db,Namespace:kube-system,Attempt:0,}" Feb 13 19:49:55.316394 containerd[2031]: time="2025-02-13T19:49:55.316314970Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fzmgc,Uid:5c26cb20-5ab9-4510-8c96-41e66a4cb07e,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"fa4f55657586a85b99a44af1277281bac814aa53f0f204a2246cd3ff8bd22f5a\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:49:55.317356 kubelet[3472]: E0213 19:49:55.316641 3472 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa4f55657586a85b99a44af1277281bac814aa53f0f204a2246cd3ff8bd22f5a\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:49:55.317356 kubelet[3472]: E0213 19:49:55.317151 3472 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa4f55657586a85b99a44af1277281bac814aa53f0f204a2246cd3ff8bd22f5a\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-fzmgc" Feb 13 19:49:55.317356 kubelet[3472]: E0213 19:49:55.317197 3472 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa4f55657586a85b99a44af1277281bac814aa53f0f204a2246cd3ff8bd22f5a\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-fzmgc" Feb 13 19:49:55.317356 kubelet[3472]: E0213 19:49:55.317300 3472 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-fzmgc_kube-system(5c26cb20-5ab9-4510-8c96-41e66a4cb07e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-fzmgc_kube-system(5c26cb20-5ab9-4510-8c96-41e66a4cb07e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fa4f55657586a85b99a44af1277281bac814aa53f0f204a2246cd3ff8bd22f5a\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-668d6bf9bc-fzmgc" podUID="5c26cb20-5ab9-4510-8c96-41e66a4cb07e" Feb 13 19:49:55.344699 containerd[2031]: time="2025-02-13T19:49:55.344602510Z" level=info msg="CreateContainer within sandbox \"017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}" Feb 13 19:49:55.371458 containerd[2031]: time="2025-02-13T19:49:55.371346610Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-tqz6p,Uid:d9e882a4-ba13-47f9-89d3-954cc087e3db,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c1f66de085d590283b3d5ce854f0d71601ddb2538a16b6bb393e167c2ecc68f6\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:49:55.372915 kubelet[3472]: E0213 19:49:55.371747 3472 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c1f66de085d590283b3d5ce854f0d71601ddb2538a16b6bb393e167c2ecc68f6\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" Feb 13 19:49:55.372915 kubelet[3472]: E0213 19:49:55.371820 3472 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c1f66de085d590283b3d5ce854f0d71601ddb2538a16b6bb393e167c2ecc68f6\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-tqz6p" Feb 13 19:49:55.372915 kubelet[3472]: E0213 19:49:55.371853 3472 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c1f66de085d590283b3d5ce854f0d71601ddb2538a16b6bb393e167c2ecc68f6\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-668d6bf9bc-tqz6p" Feb 13 19:49:55.372915 kubelet[3472]: E0213 19:49:55.371927 3472 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-tqz6p_kube-system(d9e882a4-ba13-47f9-89d3-954cc087e3db)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-tqz6p_kube-system(d9e882a4-ba13-47f9-89d3-954cc087e3db)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c1f66de085d590283b3d5ce854f0d71601ddb2538a16b6bb393e167c2ecc68f6\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-668d6bf9bc-tqz6p" podUID="d9e882a4-ba13-47f9-89d3-954cc087e3db" Feb 13 19:49:55.374883 containerd[2031]: time="2025-02-13T19:49:55.374397814Z" level=info msg="CreateContainer within sandbox \"017a7355efaf2ad0415bddaabb77b1e7a393e78c2369ad3cbc44bf5cbf15ceec\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"8bacf6d2b31849e26c57516ea0064c52691454c36dea8dd27e0dc1199dad6379\"" Feb 13 19:49:55.376112 containerd[2031]: time="2025-02-13T19:49:55.375946126Z" level=info msg="StartContainer for \"8bacf6d2b31849e26c57516ea0064c52691454c36dea8dd27e0dc1199dad6379\"" Feb 13 19:49:55.422064 systemd[1]: Started cri-containerd-8bacf6d2b31849e26c57516ea0064c52691454c36dea8dd27e0dc1199dad6379.scope - libcontainer container 8bacf6d2b31849e26c57516ea0064c52691454c36dea8dd27e0dc1199dad6379. Feb 13 19:49:55.470826 containerd[2031]: time="2025-02-13T19:49:55.470376935Z" level=info msg="StartContainer for \"8bacf6d2b31849e26c57516ea0064c52691454c36dea8dd27e0dc1199dad6379\" returns successfully" Feb 13 19:49:55.693326 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-af8f2c25659720153c71fe10583b41725d5af62027a3f48dbbdfcb87fa1dab34-rootfs.mount: Deactivated successfully. Feb 13 19:49:56.544940 (udev-worker)[4023]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:49:56.567892 systemd-networkd[1926]: flannel.1: Link UP Feb 13 19:49:56.567913 systemd-networkd[1926]: flannel.1: Gained carrier Feb 13 19:49:58.506926 systemd-networkd[1926]: flannel.1: Gained IPv6LL Feb 13 19:50:00.861989 ntpd[1998]: Listen normally on 7 flannel.1 192.168.0.0:123 Feb 13 19:50:00.862121 ntpd[1998]: Listen normally on 8 flannel.1 [fe80::49f:84ff:feca:e57b%4]:123 Feb 13 19:50:00.862932 ntpd[1998]: 13 Feb 19:50:00 ntpd[1998]: Listen normally on 7 flannel.1 192.168.0.0:123 Feb 13 19:50:00.862932 ntpd[1998]: 13 Feb 19:50:00 ntpd[1998]: Listen normally on 8 flannel.1 [fe80::49f:84ff:feca:e57b%4]:123 Feb 13 19:50:07.137272 containerd[2031]: time="2025-02-13T19:50:07.136699797Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-tqz6p,Uid:d9e882a4-ba13-47f9-89d3-954cc087e3db,Namespace:kube-system,Attempt:0,}" Feb 13 19:50:07.185786 systemd-networkd[1926]: cni0: Link UP Feb 13 19:50:07.185801 systemd-networkd[1926]: cni0: Gained carrier Feb 13 19:50:07.192125 (udev-worker)[4162]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:50:07.194722 systemd-networkd[1926]: cni0: Lost carrier Feb 13 19:50:07.200837 systemd-networkd[1926]: veth837aea4e: Link UP Feb 13 19:50:07.204738 kernel: cni0: port 1(veth837aea4e) entered blocking state Feb 13 19:50:07.204868 kernel: cni0: port 1(veth837aea4e) entered disabled state Feb 13 19:50:07.206185 kernel: veth837aea4e: entered allmulticast mode Feb 13 19:50:07.207630 kernel: veth837aea4e: entered promiscuous mode Feb 13 19:50:07.210186 kernel: cni0: port 1(veth837aea4e) entered blocking state Feb 13 19:50:07.210232 kernel: cni0: port 1(veth837aea4e) entered forwarding state Feb 13 19:50:07.214751 kernel: cni0: port 1(veth837aea4e) entered disabled state Feb 13 19:50:07.215439 (udev-worker)[4167]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:50:07.225934 kernel: cni0: port 1(veth837aea4e) entered blocking state Feb 13 19:50:07.226039 kernel: cni0: port 1(veth837aea4e) entered forwarding state Feb 13 19:50:07.226429 systemd-networkd[1926]: veth837aea4e: Gained carrier Feb 13 19:50:07.227946 systemd-networkd[1926]: cni0: Gained carrier Feb 13 19:50:07.234183 containerd[2031]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x40000948e8), "name":"cbr0", "type":"bridge"} Feb 13 19:50:07.234183 containerd[2031]: delegateAdd: netconf sent to delegate plugin: Feb 13 19:50:07.276877 containerd[2031]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":8951,"name":"cbr0","type":"bridge"}time="2025-02-13T19:50:07.276734001Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:50:07.277076 containerd[2031]: time="2025-02-13T19:50:07.276840189Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:50:07.277076 containerd[2031]: time="2025-02-13T19:50:07.276879009Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:07.277076 containerd[2031]: time="2025-02-13T19:50:07.277036797Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:07.321048 systemd[1]: Started cri-containerd-0348424cc4eecbe8f0ed3d54fff8f66a6eb8e1cb8e2ecf8964ede342d6a3abec.scope - libcontainer container 0348424cc4eecbe8f0ed3d54fff8f66a6eb8e1cb8e2ecf8964ede342d6a3abec. Feb 13 19:50:07.381292 containerd[2031]: time="2025-02-13T19:50:07.381233674Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-tqz6p,Uid:d9e882a4-ba13-47f9-89d3-954cc087e3db,Namespace:kube-system,Attempt:0,} returns sandbox id \"0348424cc4eecbe8f0ed3d54fff8f66a6eb8e1cb8e2ecf8964ede342d6a3abec\"" Feb 13 19:50:07.388971 containerd[2031]: time="2025-02-13T19:50:07.388817950Z" level=info msg="CreateContainer within sandbox \"0348424cc4eecbe8f0ed3d54fff8f66a6eb8e1cb8e2ecf8964ede342d6a3abec\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:50:07.420313 containerd[2031]: time="2025-02-13T19:50:07.420237214Z" level=info msg="CreateContainer within sandbox \"0348424cc4eecbe8f0ed3d54fff8f66a6eb8e1cb8e2ecf8964ede342d6a3abec\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a5f8f8435cb8fa1761bc5827351a6588c0cd238c8200f2f4b8cac43eceb70c56\"" Feb 13 19:50:07.421435 containerd[2031]: time="2025-02-13T19:50:07.421378114Z" level=info msg="StartContainer for \"a5f8f8435cb8fa1761bc5827351a6588c0cd238c8200f2f4b8cac43eceb70c56\"" Feb 13 19:50:07.465010 systemd[1]: Started cri-containerd-a5f8f8435cb8fa1761bc5827351a6588c0cd238c8200f2f4b8cac43eceb70c56.scope - libcontainer container a5f8f8435cb8fa1761bc5827351a6588c0cd238c8200f2f4b8cac43eceb70c56. Feb 13 19:50:07.513974 containerd[2031]: time="2025-02-13T19:50:07.513788747Z" level=info msg="StartContainer for \"a5f8f8435cb8fa1761bc5827351a6588c0cd238c8200f2f4b8cac43eceb70c56\" returns successfully" Feb 13 19:50:08.156962 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3372067957.mount: Deactivated successfully. Feb 13 19:50:08.386525 kubelet[3472]: I0213 19:50:08.386436 3472 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-qgxkx" podStartSLOduration=15.937756459 podStartE2EDuration="22.386416139s" podCreationTimestamp="2025-02-13 19:49:46 +0000 UTC" firstStartedPulling="2025-02-13 19:49:48.219155067 +0000 UTC m=+7.273118569" lastFinishedPulling="2025-02-13 19:49:54.667814747 +0000 UTC m=+13.721778249" observedRunningTime="2025-02-13 19:49:56.381977231 +0000 UTC m=+15.435940769" watchObservedRunningTime="2025-02-13 19:50:08.386416139 +0000 UTC m=+27.440379629" Feb 13 19:50:08.409384 kubelet[3472]: I0213 19:50:08.408645 3472 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-tqz6p" podStartSLOduration=22.408622463 podStartE2EDuration="22.408622463s" podCreationTimestamp="2025-02-13 19:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:50:08.387111131 +0000 UTC m=+27.441074633" watchObservedRunningTime="2025-02-13 19:50:08.408622463 +0000 UTC m=+27.462585977" Feb 13 19:50:08.490936 systemd-networkd[1926]: veth837aea4e: Gained IPv6LL Feb 13 19:50:08.555051 systemd-networkd[1926]: cni0: Gained IPv6LL Feb 13 19:50:09.136090 containerd[2031]: time="2025-02-13T19:50:09.135988535Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fzmgc,Uid:5c26cb20-5ab9-4510-8c96-41e66a4cb07e,Namespace:kube-system,Attempt:0,}" Feb 13 19:50:09.187800 systemd-networkd[1926]: vetha03cbff9: Link UP Feb 13 19:50:09.188720 kernel: cni0: port 2(vetha03cbff9) entered blocking state Feb 13 19:50:09.188786 kernel: cni0: port 2(vetha03cbff9) entered disabled state Feb 13 19:50:09.188820 kernel: vetha03cbff9: entered allmulticast mode Feb 13 19:50:09.190785 kernel: vetha03cbff9: entered promiscuous mode Feb 13 19:50:09.191562 (udev-worker)[4166]: Network interface NamePolicy= disabled on kernel command line. Feb 13 19:50:09.202830 kernel: cni0: port 2(vetha03cbff9) entered blocking state Feb 13 19:50:09.202942 kernel: cni0: port 2(vetha03cbff9) entered forwarding state Feb 13 19:50:09.201072 systemd-networkd[1926]: vetha03cbff9: Gained carrier Feb 13 19:50:09.205868 containerd[2031]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x40000948e8), "name":"cbr0", "type":"bridge"} Feb 13 19:50:09.205868 containerd[2031]: delegateAdd: netconf sent to delegate plugin: Feb 13 19:50:09.244043 containerd[2031]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":8951,"name":"cbr0","type":"bridge"}time="2025-02-13T19:50:09.243843455Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:50:09.244043 containerd[2031]: time="2025-02-13T19:50:09.243987887Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:50:09.244405 containerd[2031]: time="2025-02-13T19:50:09.244034447Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:09.244405 containerd[2031]: time="2025-02-13T19:50:09.244218179Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:09.282251 systemd[1]: run-containerd-runc-k8s.io-a10adfec433f5a7112ae0564eabc76aecb3d35dc7b53505f1d1ab3f7aceb064f-runc.Fc6OSf.mount: Deactivated successfully. Feb 13 19:50:09.297989 systemd[1]: Started cri-containerd-a10adfec433f5a7112ae0564eabc76aecb3d35dc7b53505f1d1ab3f7aceb064f.scope - libcontainer container a10adfec433f5a7112ae0564eabc76aecb3d35dc7b53505f1d1ab3f7aceb064f. Feb 13 19:50:09.359517 containerd[2031]: time="2025-02-13T19:50:09.359462376Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-fzmgc,Uid:5c26cb20-5ab9-4510-8c96-41e66a4cb07e,Namespace:kube-system,Attempt:0,} returns sandbox id \"a10adfec433f5a7112ae0564eabc76aecb3d35dc7b53505f1d1ab3f7aceb064f\"" Feb 13 19:50:09.367073 containerd[2031]: time="2025-02-13T19:50:09.366965292Z" level=info msg="CreateContainer within sandbox \"a10adfec433f5a7112ae0564eabc76aecb3d35dc7b53505f1d1ab3f7aceb064f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:50:09.394649 containerd[2031]: time="2025-02-13T19:50:09.393711588Z" level=info msg="CreateContainer within sandbox \"a10adfec433f5a7112ae0564eabc76aecb3d35dc7b53505f1d1ab3f7aceb064f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"e2c3dc3a5c94180e925f0f9a8b9a79ee40e2904cef50ae0e478c3ab658b3994a\"" Feb 13 19:50:09.395197 containerd[2031]: time="2025-02-13T19:50:09.395145876Z" level=info msg="StartContainer for \"e2c3dc3a5c94180e925f0f9a8b9a79ee40e2904cef50ae0e478c3ab658b3994a\"" Feb 13 19:50:09.442976 systemd[1]: Started cri-containerd-e2c3dc3a5c94180e925f0f9a8b9a79ee40e2904cef50ae0e478c3ab658b3994a.scope - libcontainer container e2c3dc3a5c94180e925f0f9a8b9a79ee40e2904cef50ae0e478c3ab658b3994a. Feb 13 19:50:09.499528 containerd[2031]: time="2025-02-13T19:50:09.499371684Z" level=info msg="StartContainer for \"e2c3dc3a5c94180e925f0f9a8b9a79ee40e2904cef50ae0e478c3ab658b3994a\" returns successfully" Feb 13 19:50:10.407224 kubelet[3472]: I0213 19:50:10.407121 3472 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-fzmgc" podStartSLOduration=24.407097469 podStartE2EDuration="24.407097469s" podCreationTimestamp="2025-02-13 19:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:50:10.405151081 +0000 UTC m=+29.459114595" watchObservedRunningTime="2025-02-13 19:50:10.407097469 +0000 UTC m=+29.461060959" Feb 13 19:50:10.730947 systemd-networkd[1926]: vetha03cbff9: Gained IPv6LL Feb 13 19:50:12.862069 ntpd[1998]: Listen normally on 9 cni0 192.168.0.1:123 Feb 13 19:50:12.862216 ntpd[1998]: Listen normally on 10 cni0 [fe80::c458:9fff:fe6e:6d18%5]:123 Feb 13 19:50:12.862708 ntpd[1998]: 13 Feb 19:50:12 ntpd[1998]: Listen normally on 9 cni0 192.168.0.1:123 Feb 13 19:50:12.862708 ntpd[1998]: 13 Feb 19:50:12 ntpd[1998]: Listen normally on 10 cni0 [fe80::c458:9fff:fe6e:6d18%5]:123 Feb 13 19:50:12.862708 ntpd[1998]: 13 Feb 19:50:12 ntpd[1998]: Listen normally on 11 veth837aea4e [fe80::b8f7:a7ff:fe81:5925%6]:123 Feb 13 19:50:12.862708 ntpd[1998]: 13 Feb 19:50:12 ntpd[1998]: Listen normally on 12 vetha03cbff9 [fe80::9028:d3ff:fe33:6b05%7]:123 Feb 13 19:50:12.862299 ntpd[1998]: Listen normally on 11 veth837aea4e [fe80::b8f7:a7ff:fe81:5925%6]:123 Feb 13 19:50:12.862406 ntpd[1998]: Listen normally on 12 vetha03cbff9 [fe80::9028:d3ff:fe33:6b05%7]:123 Feb 13 19:50:25.098210 systemd[1]: Started sshd@7-172.31.25.134:22-139.178.89.65:51492.service - OpenSSH per-connection server daemon (139.178.89.65:51492). Feb 13 19:50:25.279802 sshd[4446]: Accepted publickey for core from 139.178.89.65 port 51492 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:25.282280 sshd[4446]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:25.289904 systemd-logind[2004]: New session 8 of user core. Feb 13 19:50:25.299990 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 19:50:25.562036 sshd[4446]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:25.569938 systemd[1]: sshd@7-172.31.25.134:22-139.178.89.65:51492.service: Deactivated successfully. Feb 13 19:50:25.574433 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 19:50:25.576772 systemd-logind[2004]: Session 8 logged out. Waiting for processes to exit. Feb 13 19:50:25.579329 systemd-logind[2004]: Removed session 8. Feb 13 19:50:30.600246 systemd[1]: Started sshd@8-172.31.25.134:22-139.178.89.65:51498.service - OpenSSH per-connection server daemon (139.178.89.65:51498). Feb 13 19:50:30.779752 sshd[4482]: Accepted publickey for core from 139.178.89.65 port 51498 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:30.782528 sshd[4482]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:30.791791 systemd-logind[2004]: New session 9 of user core. Feb 13 19:50:30.798932 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 19:50:31.038291 sshd[4482]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:31.045160 systemd[1]: sshd@8-172.31.25.134:22-139.178.89.65:51498.service: Deactivated successfully. Feb 13 19:50:31.049442 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 19:50:31.051571 systemd-logind[2004]: Session 9 logged out. Waiting for processes to exit. Feb 13 19:50:31.054000 systemd-logind[2004]: Removed session 9. Feb 13 19:50:36.078235 systemd[1]: Started sshd@9-172.31.25.134:22-139.178.89.65:54218.service - OpenSSH per-connection server daemon (139.178.89.65:54218). Feb 13 19:50:36.259712 sshd[4517]: Accepted publickey for core from 139.178.89.65 port 54218 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:36.262720 sshd[4517]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:36.272193 systemd-logind[2004]: New session 10 of user core. Feb 13 19:50:36.280015 systemd[1]: Started session-10.scope - Session 10 of User core. Feb 13 19:50:36.521715 sshd[4517]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:36.527252 systemd-logind[2004]: Session 10 logged out. Waiting for processes to exit. Feb 13 19:50:36.528316 systemd[1]: sshd@9-172.31.25.134:22-139.178.89.65:54218.service: Deactivated successfully. Feb 13 19:50:36.533219 systemd[1]: session-10.scope: Deactivated successfully. Feb 13 19:50:36.539222 systemd-logind[2004]: Removed session 10. Feb 13 19:50:36.560310 systemd[1]: Started sshd@10-172.31.25.134:22-139.178.89.65:54234.service - OpenSSH per-connection server daemon (139.178.89.65:54234). Feb 13 19:50:36.734328 sshd[4531]: Accepted publickey for core from 139.178.89.65 port 54234 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:36.736974 sshd[4531]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:36.745466 systemd-logind[2004]: New session 11 of user core. Feb 13 19:50:36.749940 systemd[1]: Started session-11.scope - Session 11 of User core. Feb 13 19:50:37.066674 sshd[4531]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:37.077364 systemd[1]: session-11.scope: Deactivated successfully. Feb 13 19:50:37.078994 systemd[1]: sshd@10-172.31.25.134:22-139.178.89.65:54234.service: Deactivated successfully. Feb 13 19:50:37.091843 systemd-logind[2004]: Session 11 logged out. Waiting for processes to exit. Feb 13 19:50:37.116452 systemd[1]: Started sshd@11-172.31.25.134:22-139.178.89.65:54242.service - OpenSSH per-connection server daemon (139.178.89.65:54242). Feb 13 19:50:37.118710 systemd-logind[2004]: Removed session 11. Feb 13 19:50:37.300123 sshd[4563]: Accepted publickey for core from 139.178.89.65 port 54242 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:37.303280 sshd[4563]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:37.313031 systemd-logind[2004]: New session 12 of user core. Feb 13 19:50:37.319265 systemd[1]: Started session-12.scope - Session 12 of User core. Feb 13 19:50:37.589843 sshd[4563]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:37.594800 systemd-logind[2004]: Session 12 logged out. Waiting for processes to exit. Feb 13 19:50:37.596267 systemd[1]: sshd@11-172.31.25.134:22-139.178.89.65:54242.service: Deactivated successfully. Feb 13 19:50:37.602963 systemd[1]: session-12.scope: Deactivated successfully. Feb 13 19:50:37.608635 systemd-logind[2004]: Removed session 12. Feb 13 19:50:42.633198 systemd[1]: Started sshd@12-172.31.25.134:22-139.178.89.65:54258.service - OpenSSH per-connection server daemon (139.178.89.65:54258). Feb 13 19:50:42.813147 sshd[4598]: Accepted publickey for core from 139.178.89.65 port 54258 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:42.815882 sshd[4598]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:42.824593 systemd-logind[2004]: New session 13 of user core. Feb 13 19:50:42.830957 systemd[1]: Started session-13.scope - Session 13 of User core. Feb 13 19:50:43.070526 sshd[4598]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:43.076142 systemd[1]: sshd@12-172.31.25.134:22-139.178.89.65:54258.service: Deactivated successfully. Feb 13 19:50:43.079956 systemd[1]: session-13.scope: Deactivated successfully. Feb 13 19:50:43.085927 systemd-logind[2004]: Session 13 logged out. Waiting for processes to exit. Feb 13 19:50:43.087928 systemd-logind[2004]: Removed session 13. Feb 13 19:50:48.113231 systemd[1]: Started sshd@13-172.31.25.134:22-139.178.89.65:51874.service - OpenSSH per-connection server daemon (139.178.89.65:51874). Feb 13 19:50:48.287207 sshd[4634]: Accepted publickey for core from 139.178.89.65 port 51874 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:48.289846 sshd[4634]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:48.298229 systemd-logind[2004]: New session 14 of user core. Feb 13 19:50:48.304948 systemd[1]: Started session-14.scope - Session 14 of User core. Feb 13 19:50:48.542816 sshd[4634]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:48.550260 systemd-logind[2004]: Session 14 logged out. Waiting for processes to exit. Feb 13 19:50:48.550563 systemd[1]: sshd@13-172.31.25.134:22-139.178.89.65:51874.service: Deactivated successfully. Feb 13 19:50:48.554386 systemd[1]: session-14.scope: Deactivated successfully. Feb 13 19:50:48.558266 systemd-logind[2004]: Removed session 14. Feb 13 19:50:53.582230 systemd[1]: Started sshd@14-172.31.25.134:22-139.178.89.65:51888.service - OpenSSH per-connection server daemon (139.178.89.65:51888). Feb 13 19:50:53.754210 sshd[4668]: Accepted publickey for core from 139.178.89.65 port 51888 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:53.756919 sshd[4668]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:53.764488 systemd-logind[2004]: New session 15 of user core. Feb 13 19:50:53.776011 systemd[1]: Started session-15.scope - Session 15 of User core. Feb 13 19:50:54.023313 sshd[4668]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:54.029929 systemd[1]: sshd@14-172.31.25.134:22-139.178.89.65:51888.service: Deactivated successfully. Feb 13 19:50:54.034161 systemd[1]: session-15.scope: Deactivated successfully. Feb 13 19:50:54.035970 systemd-logind[2004]: Session 15 logged out. Waiting for processes to exit. Feb 13 19:50:54.037644 systemd-logind[2004]: Removed session 15. Feb 13 19:50:59.062251 systemd[1]: Started sshd@15-172.31.25.134:22-139.178.89.65:44548.service - OpenSSH per-connection server daemon (139.178.89.65:44548). Feb 13 19:50:59.238317 sshd[4704]: Accepted publickey for core from 139.178.89.65 port 44548 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:59.240944 sshd[4704]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:59.248654 systemd-logind[2004]: New session 16 of user core. Feb 13 19:50:59.256928 systemd[1]: Started session-16.scope - Session 16 of User core. Feb 13 19:50:59.494518 sshd[4704]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:59.500817 systemd-logind[2004]: Session 16 logged out. Waiting for processes to exit. Feb 13 19:50:59.503037 systemd[1]: sshd@15-172.31.25.134:22-139.178.89.65:44548.service: Deactivated successfully. Feb 13 19:50:59.507560 systemd[1]: session-16.scope: Deactivated successfully. Feb 13 19:50:59.511128 systemd-logind[2004]: Removed session 16. Feb 13 19:50:59.534210 systemd[1]: Started sshd@16-172.31.25.134:22-139.178.89.65:44554.service - OpenSSH per-connection server daemon (139.178.89.65:44554). Feb 13 19:50:59.706164 sshd[4717]: Accepted publickey for core from 139.178.89.65 port 44554 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:50:59.708824 sshd[4717]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:59.716186 systemd-logind[2004]: New session 17 of user core. Feb 13 19:50:59.726972 systemd[1]: Started session-17.scope - Session 17 of User core. Feb 13 19:51:00.025022 sshd[4717]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:00.030746 systemd[1]: sshd@16-172.31.25.134:22-139.178.89.65:44554.service: Deactivated successfully. Feb 13 19:51:00.036189 systemd[1]: session-17.scope: Deactivated successfully. Feb 13 19:51:00.039222 systemd-logind[2004]: Session 17 logged out. Waiting for processes to exit. Feb 13 19:51:00.041736 systemd-logind[2004]: Removed session 17. Feb 13 19:51:00.067194 systemd[1]: Started sshd@17-172.31.25.134:22-139.178.89.65:44560.service - OpenSSH per-connection server daemon (139.178.89.65:44560). Feb 13 19:51:00.236193 sshd[4728]: Accepted publickey for core from 139.178.89.65 port 44560 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:51:00.238920 sshd[4728]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:00.245921 systemd-logind[2004]: New session 18 of user core. Feb 13 19:51:00.254103 systemd[1]: Started session-18.scope - Session 18 of User core. Feb 13 19:51:01.408941 sshd[4728]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:01.418494 systemd[1]: sshd@17-172.31.25.134:22-139.178.89.65:44560.service: Deactivated successfully. Feb 13 19:51:01.427593 systemd[1]: session-18.scope: Deactivated successfully. Feb 13 19:51:01.437984 systemd-logind[2004]: Session 18 logged out. Waiting for processes to exit. Feb 13 19:51:01.458412 systemd[1]: Started sshd@18-172.31.25.134:22-139.178.89.65:44562.service - OpenSSH per-connection server daemon (139.178.89.65:44562). Feb 13 19:51:01.460265 systemd-logind[2004]: Removed session 18. Feb 13 19:51:01.636083 sshd[4746]: Accepted publickey for core from 139.178.89.65 port 44562 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:51:01.638824 sshd[4746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:01.646217 systemd-logind[2004]: New session 19 of user core. Feb 13 19:51:01.662995 systemd[1]: Started session-19.scope - Session 19 of User core. Feb 13 19:51:02.147627 sshd[4746]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:02.155269 systemd[1]: sshd@18-172.31.25.134:22-139.178.89.65:44562.service: Deactivated successfully. Feb 13 19:51:02.160343 systemd[1]: session-19.scope: Deactivated successfully. Feb 13 19:51:02.164257 systemd-logind[2004]: Session 19 logged out. Waiting for processes to exit. Feb 13 19:51:02.166890 systemd-logind[2004]: Removed session 19. Feb 13 19:51:02.190403 systemd[1]: Started sshd@19-172.31.25.134:22-139.178.89.65:44574.service - OpenSSH per-connection server daemon (139.178.89.65:44574). Feb 13 19:51:02.366573 sshd[4772]: Accepted publickey for core from 139.178.89.65 port 44574 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:51:02.369233 sshd[4772]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:02.378863 systemd-logind[2004]: New session 20 of user core. Feb 13 19:51:02.385953 systemd[1]: Started session-20.scope - Session 20 of User core. Feb 13 19:51:02.626026 sshd[4772]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:02.632512 systemd-logind[2004]: Session 20 logged out. Waiting for processes to exit. Feb 13 19:51:02.634009 systemd[1]: sshd@19-172.31.25.134:22-139.178.89.65:44574.service: Deactivated successfully. Feb 13 19:51:02.637180 systemd[1]: session-20.scope: Deactivated successfully. Feb 13 19:51:02.639660 systemd-logind[2004]: Removed session 20. Feb 13 19:51:07.665240 systemd[1]: Started sshd@20-172.31.25.134:22-139.178.89.65:53896.service - OpenSSH per-connection server daemon (139.178.89.65:53896). Feb 13 19:51:07.844786 sshd[4812]: Accepted publickey for core from 139.178.89.65 port 53896 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:51:07.846383 sshd[4812]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:07.857528 systemd-logind[2004]: New session 21 of user core. Feb 13 19:51:07.863966 systemd[1]: Started session-21.scope - Session 21 of User core. Feb 13 19:51:08.105112 sshd[4812]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:08.110094 systemd[1]: sshd@20-172.31.25.134:22-139.178.89.65:53896.service: Deactivated successfully. Feb 13 19:51:08.115451 systemd[1]: session-21.scope: Deactivated successfully. Feb 13 19:51:08.119492 systemd-logind[2004]: Session 21 logged out. Waiting for processes to exit. Feb 13 19:51:08.122304 systemd-logind[2004]: Removed session 21. Feb 13 19:51:13.147208 systemd[1]: Started sshd@21-172.31.25.134:22-139.178.89.65:53904.service - OpenSSH per-connection server daemon (139.178.89.65:53904). Feb 13 19:51:13.324046 sshd[4848]: Accepted publickey for core from 139.178.89.65 port 53904 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:51:13.327343 sshd[4848]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:13.337576 systemd-logind[2004]: New session 22 of user core. Feb 13 19:51:13.345966 systemd[1]: Started session-22.scope - Session 22 of User core. Feb 13 19:51:13.585988 sshd[4848]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:13.592343 systemd[1]: sshd@21-172.31.25.134:22-139.178.89.65:53904.service: Deactivated successfully. Feb 13 19:51:13.596103 systemd[1]: session-22.scope: Deactivated successfully. Feb 13 19:51:13.597956 systemd-logind[2004]: Session 22 logged out. Waiting for processes to exit. Feb 13 19:51:13.600141 systemd-logind[2004]: Removed session 22. Feb 13 19:51:18.628248 systemd[1]: Started sshd@22-172.31.25.134:22-139.178.89.65:52542.service - OpenSSH per-connection server daemon (139.178.89.65:52542). Feb 13 19:51:18.796327 sshd[4884]: Accepted publickey for core from 139.178.89.65 port 52542 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:51:18.798920 sshd[4884]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:18.807471 systemd-logind[2004]: New session 23 of user core. Feb 13 19:51:18.816962 systemd[1]: Started session-23.scope - Session 23 of User core. Feb 13 19:51:19.050370 sshd[4884]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:19.057456 systemd[1]: sshd@22-172.31.25.134:22-139.178.89.65:52542.service: Deactivated successfully. Feb 13 19:51:19.062158 systemd[1]: session-23.scope: Deactivated successfully. Feb 13 19:51:19.063490 systemd-logind[2004]: Session 23 logged out. Waiting for processes to exit. Feb 13 19:51:19.067192 systemd-logind[2004]: Removed session 23. Feb 13 19:51:24.092253 systemd[1]: Started sshd@23-172.31.25.134:22-139.178.89.65:52548.service - OpenSSH per-connection server daemon (139.178.89.65:52548). Feb 13 19:51:24.262606 sshd[4918]: Accepted publickey for core from 139.178.89.65 port 52548 ssh2: RSA SHA256:H27J0U/EpkvOcUDI+hexgwVcKe7FsK9V5j851fkSvZ4 Feb 13 19:51:24.266781 sshd[4918]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:24.278439 systemd-logind[2004]: New session 24 of user core. Feb 13 19:51:24.283985 systemd[1]: Started session-24.scope - Session 24 of User core. Feb 13 19:51:24.518028 sshd[4918]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:24.525252 systemd[1]: sshd@23-172.31.25.134:22-139.178.89.65:52548.service: Deactivated successfully. Feb 13 19:51:24.529161 systemd[1]: session-24.scope: Deactivated successfully. Feb 13 19:51:24.530916 systemd-logind[2004]: Session 24 logged out. Waiting for processes to exit. Feb 13 19:51:24.533002 systemd-logind[2004]: Removed session 24. Feb 13 19:51:39.537913 systemd[1]: cri-containerd-8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f.scope: Deactivated successfully. Feb 13 19:51:39.540296 systemd[1]: cri-containerd-8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f.scope: Consumed 4.635s CPU time, 19.8M memory peak, 0B memory swap peak. Feb 13 19:51:39.579780 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f-rootfs.mount: Deactivated successfully. Feb 13 19:51:39.597949 containerd[2031]: time="2025-02-13T19:51:39.597821056Z" level=info msg="shim disconnected" id=8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f namespace=k8s.io Feb 13 19:51:39.598792 containerd[2031]: time="2025-02-13T19:51:39.598054144Z" level=warning msg="cleaning up after shim disconnected" id=8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f namespace=k8s.io Feb 13 19:51:39.598792 containerd[2031]: time="2025-02-13T19:51:39.598079572Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:51:40.590185 kubelet[3472]: I0213 19:51:40.590138 3472 scope.go:117] "RemoveContainer" containerID="8ab72854853c3f93f46e07faced49ced503ae0b0e16af46da3c86258c46d177f" Feb 13 19:51:40.593140 containerd[2031]: time="2025-02-13T19:51:40.593086277Z" level=info msg="CreateContainer within sandbox \"c6475971ecc09d0ddb241f7a4290493070613d48637dd00a9864d72ae8bbf546\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Feb 13 19:51:40.625248 containerd[2031]: time="2025-02-13T19:51:40.624715409Z" level=info msg="CreateContainer within sandbox \"c6475971ecc09d0ddb241f7a4290493070613d48637dd00a9864d72ae8bbf546\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"e1a7755a1ffdb5f1e9e81983321286d1c01849436f375a54e3b670c20be7ee77\"" Feb 13 19:51:40.626133 containerd[2031]: time="2025-02-13T19:51:40.626040941Z" level=info msg="StartContainer for \"e1a7755a1ffdb5f1e9e81983321286d1c01849436f375a54e3b670c20be7ee77\"" Feb 13 19:51:40.686971 systemd[1]: Started cri-containerd-e1a7755a1ffdb5f1e9e81983321286d1c01849436f375a54e3b670c20be7ee77.scope - libcontainer container e1a7755a1ffdb5f1e9e81983321286d1c01849436f375a54e3b670c20be7ee77. Feb 13 19:51:40.756596 containerd[2031]: time="2025-02-13T19:51:40.756429414Z" level=info msg="StartContainer for \"e1a7755a1ffdb5f1e9e81983321286d1c01849436f375a54e3b670c20be7ee77\" returns successfully" Feb 13 19:51:42.798906 kubelet[3472]: E0213 19:51:42.798702 3472 controller.go:195] "Failed to update lease" err="Put \"https://172.31.25.134:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-134?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 13 19:51:43.343228 systemd[1]: cri-containerd-4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da.scope: Deactivated successfully. Feb 13 19:51:43.344062 systemd[1]: cri-containerd-4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da.scope: Consumed 2.926s CPU time, 16.0M memory peak, 0B memory swap peak. Feb 13 19:51:43.394449 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da-rootfs.mount: Deactivated successfully. Feb 13 19:51:43.409082 containerd[2031]: time="2025-02-13T19:51:43.408966487Z" level=info msg="shim disconnected" id=4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da namespace=k8s.io Feb 13 19:51:43.409082 containerd[2031]: time="2025-02-13T19:51:43.409071307Z" level=warning msg="cleaning up after shim disconnected" id=4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da namespace=k8s.io Feb 13 19:51:43.409806 containerd[2031]: time="2025-02-13T19:51:43.409094107Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:51:43.607166 kubelet[3472]: I0213 19:51:43.606150 3472 scope.go:117] "RemoveContainer" containerID="4bb23eea6a07d1da3c2088a583d5571251d43cbd0fccba6fec4f1be1c36898da" Feb 13 19:51:43.609124 containerd[2031]: time="2025-02-13T19:51:43.609050420Z" level=info msg="CreateContainer within sandbox \"450ec195605da19dd4f9f3b322a18b31602f3c90ac971390a678b12f204aac82\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Feb 13 19:51:43.638879 containerd[2031]: time="2025-02-13T19:51:43.638813732Z" level=info msg="CreateContainer within sandbox \"450ec195605da19dd4f9f3b322a18b31602f3c90ac971390a678b12f204aac82\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"7f93f142fad2d75e70f3e7c0a81ff4587dabb5c57031c6b8baceb9cbcf58804c\"" Feb 13 19:51:43.639856 containerd[2031]: time="2025-02-13T19:51:43.639813752Z" level=info msg="StartContainer for \"7f93f142fad2d75e70f3e7c0a81ff4587dabb5c57031c6b8baceb9cbcf58804c\"" Feb 13 19:51:43.700036 systemd[1]: Started cri-containerd-7f93f142fad2d75e70f3e7c0a81ff4587dabb5c57031c6b8baceb9cbcf58804c.scope - libcontainer container 7f93f142fad2d75e70f3e7c0a81ff4587dabb5c57031c6b8baceb9cbcf58804c. Feb 13 19:51:43.765361 containerd[2031]: time="2025-02-13T19:51:43.765148461Z" level=info msg="StartContainer for \"7f93f142fad2d75e70f3e7c0a81ff4587dabb5c57031c6b8baceb9cbcf58804c\" returns successfully"