Feb 13 20:02:44.132411 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Feb 13 18:03:41 -00 2025 Feb 13 20:02:44.132454 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=a8740cbac5121ade856b040634ad9badacd879298c24f899668a59d96c178b13 Feb 13 20:02:44.132470 kernel: BIOS-provided physical RAM map: Feb 13 20:02:44.132482 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Feb 13 20:02:44.132494 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Feb 13 20:02:44.132506 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Feb 13 20:02:44.132524 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007d9e9fff] usable Feb 13 20:02:44.132537 kernel: BIOS-e820: [mem 0x000000007d9ea000-0x000000007fffffff] reserved Feb 13 20:02:44.132550 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000e03fffff] reserved Feb 13 20:02:44.132563 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Feb 13 20:02:44.132575 kernel: NX (Execute Disable) protection: active Feb 13 20:02:44.132588 kernel: APIC: Static calls initialized Feb 13 20:02:44.132601 kernel: SMBIOS 2.7 present. Feb 13 20:02:44.132611 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Feb 13 20:02:44.132628 kernel: Hypervisor detected: KVM Feb 13 20:02:44.132641 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Feb 13 20:02:44.132652 kernel: kvm-clock: using sched offset of 6407830338 cycles Feb 13 20:02:44.132665 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Feb 13 20:02:44.132678 kernel: tsc: Detected 2499.996 MHz processor Feb 13 20:02:44.132690 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 13 20:02:44.132705 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 13 20:02:44.132727 kernel: last_pfn = 0x7d9ea max_arch_pfn = 0x400000000 Feb 13 20:02:44.132744 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Feb 13 20:02:44.132761 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 13 20:02:44.132777 kernel: Using GB pages for direct mapping Feb 13 20:02:44.132792 kernel: ACPI: Early table checksum verification disabled Feb 13 20:02:44.132804 kernel: ACPI: RSDP 0x00000000000F8F40 000014 (v00 AMAZON) Feb 13 20:02:44.132831 kernel: ACPI: RSDT 0x000000007D9EE350 000044 (v01 AMAZON AMZNRSDT 00000001 AMZN 00000001) Feb 13 20:02:44.132844 kernel: ACPI: FACP 0x000000007D9EFF80 000074 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Feb 13 20:02:44.132858 kernel: ACPI: DSDT 0x000000007D9EE3A0 0010E9 (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Feb 13 20:02:44.132875 kernel: ACPI: FACS 0x000000007D9EFF40 000040 Feb 13 20:02:44.132888 kernel: ACPI: SSDT 0x000000007D9EF6C0 00087A (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Feb 13 20:02:44.132902 kernel: ACPI: APIC 0x000000007D9EF5D0 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Feb 13 20:02:44.132916 kernel: ACPI: SRAT 0x000000007D9EF530 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Feb 13 20:02:44.132930 kernel: ACPI: SLIT 0x000000007D9EF4C0 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Feb 13 20:02:44.132944 kernel: ACPI: WAET 0x000000007D9EF490 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Feb 13 20:02:44.132957 kernel: ACPI: HPET 0x00000000000C9000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Feb 13 20:02:44.132971 kernel: ACPI: SSDT 0x00000000000C9040 00007B (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Feb 13 20:02:44.132985 kernel: ACPI: Reserving FACP table memory at [mem 0x7d9eff80-0x7d9efff3] Feb 13 20:02:44.133002 kernel: ACPI: Reserving DSDT table memory at [mem 0x7d9ee3a0-0x7d9ef488] Feb 13 20:02:44.133022 kernel: ACPI: Reserving FACS table memory at [mem 0x7d9eff40-0x7d9eff7f] Feb 13 20:02:44.133036 kernel: ACPI: Reserving SSDT table memory at [mem 0x7d9ef6c0-0x7d9eff39] Feb 13 20:02:44.133050 kernel: ACPI: Reserving APIC table memory at [mem 0x7d9ef5d0-0x7d9ef645] Feb 13 20:02:44.133065 kernel: ACPI: Reserving SRAT table memory at [mem 0x7d9ef530-0x7d9ef5cf] Feb 13 20:02:44.133082 kernel: ACPI: Reserving SLIT table memory at [mem 0x7d9ef4c0-0x7d9ef52b] Feb 13 20:02:44.133097 kernel: ACPI: Reserving WAET table memory at [mem 0x7d9ef490-0x7d9ef4b7] Feb 13 20:02:44.133111 kernel: ACPI: Reserving HPET table memory at [mem 0xc9000-0xc9037] Feb 13 20:02:44.133126 kernel: ACPI: Reserving SSDT table memory at [mem 0xc9040-0xc90ba] Feb 13 20:02:44.133140 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Feb 13 20:02:44.133155 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Feb 13 20:02:44.133169 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Feb 13 20:02:44.133184 kernel: NUMA: Initialized distance table, cnt=1 Feb 13 20:02:44.133198 kernel: NODE_DATA(0) allocated [mem 0x7d9e3000-0x7d9e8fff] Feb 13 20:02:44.133215 kernel: Zone ranges: Feb 13 20:02:44.133230 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 13 20:02:44.133245 kernel: DMA32 [mem 0x0000000001000000-0x000000007d9e9fff] Feb 13 20:02:44.133259 kernel: Normal empty Feb 13 20:02:44.133274 kernel: Movable zone start for each node Feb 13 20:02:44.133288 kernel: Early memory node ranges Feb 13 20:02:44.133303 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Feb 13 20:02:44.133317 kernel: node 0: [mem 0x0000000000100000-0x000000007d9e9fff] Feb 13 20:02:44.133332 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007d9e9fff] Feb 13 20:02:44.133347 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 20:02:44.133439 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Feb 13 20:02:44.133454 kernel: On node 0, zone DMA32: 9750 pages in unavailable ranges Feb 13 20:02:44.133469 kernel: ACPI: PM-Timer IO Port: 0xb008 Feb 13 20:02:44.133484 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Feb 13 20:02:44.133499 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Feb 13 20:02:44.133514 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Feb 13 20:02:44.133528 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 13 20:02:44.133543 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Feb 13 20:02:44.133558 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Feb 13 20:02:44.133596 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 13 20:02:44.133611 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Feb 13 20:02:44.133627 kernel: TSC deadline timer available Feb 13 20:02:44.133642 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Feb 13 20:02:44.133658 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Feb 13 20:02:44.133673 kernel: [mem 0x80000000-0xdfffffff] available for PCI devices Feb 13 20:02:44.133688 kernel: Booting paravirtualized kernel on KVM Feb 13 20:02:44.133704 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 13 20:02:44.133719 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Feb 13 20:02:44.133737 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u1048576 Feb 13 20:02:44.133753 kernel: pcpu-alloc: s197032 r8192 d32344 u1048576 alloc=1*2097152 Feb 13 20:02:44.133768 kernel: pcpu-alloc: [0] 0 1 Feb 13 20:02:44.133782 kernel: kvm-guest: PV spinlocks enabled Feb 13 20:02:44.133797 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Feb 13 20:02:44.133814 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=a8740cbac5121ade856b040634ad9badacd879298c24f899668a59d96c178b13 Feb 13 20:02:44.133830 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 20:02:44.133845 kernel: random: crng init done Feb 13 20:02:44.133863 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 20:02:44.133879 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Feb 13 20:02:44.133894 kernel: Fallback order for Node 0: 0 Feb 13 20:02:44.133909 kernel: Built 1 zonelists, mobility grouping on. Total pages: 506242 Feb 13 20:02:44.133924 kernel: Policy zone: DMA32 Feb 13 20:02:44.133940 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 20:02:44.133956 kernel: Memory: 1932348K/2057760K available (12288K kernel code, 2301K rwdata, 22728K rodata, 42840K init, 2352K bss, 125152K reserved, 0K cma-reserved) Feb 13 20:02:44.133971 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 13 20:02:44.133990 kernel: Kernel/User page tables isolation: enabled Feb 13 20:02:44.134005 kernel: ftrace: allocating 37921 entries in 149 pages Feb 13 20:02:44.134020 kernel: ftrace: allocated 149 pages with 4 groups Feb 13 20:02:44.134036 kernel: Dynamic Preempt: voluntary Feb 13 20:02:44.134051 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 20:02:44.134067 kernel: rcu: RCU event tracing is enabled. Feb 13 20:02:44.134083 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 13 20:02:44.134098 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 20:02:44.134114 kernel: Rude variant of Tasks RCU enabled. Feb 13 20:02:44.134129 kernel: Tracing variant of Tasks RCU enabled. Feb 13 20:02:44.134147 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 20:02:44.134162 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 13 20:02:44.134178 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Feb 13 20:02:44.134193 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 20:02:44.134209 kernel: Console: colour VGA+ 80x25 Feb 13 20:02:44.134224 kernel: printk: console [ttyS0] enabled Feb 13 20:02:44.134239 kernel: ACPI: Core revision 20230628 Feb 13 20:02:44.134255 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Feb 13 20:02:44.134270 kernel: APIC: Switch to symmetric I/O mode setup Feb 13 20:02:44.134288 kernel: x2apic enabled Feb 13 20:02:44.134304 kernel: APIC: Switched APIC routing to: physical x2apic Feb 13 20:02:44.134331 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Feb 13 20:02:44.134350 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499996) Feb 13 20:02:44.134382 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Feb 13 20:02:44.134399 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Feb 13 20:02:44.134415 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 13 20:02:44.134430 kernel: Spectre V2 : Mitigation: Retpolines Feb 13 20:02:44.134446 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 13 20:02:44.134462 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 13 20:02:44.134478 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Feb 13 20:02:44.134494 kernel: RETBleed: Vulnerable Feb 13 20:02:44.134514 kernel: Speculative Store Bypass: Vulnerable Feb 13 20:02:44.134530 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 20:02:44.134546 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Feb 13 20:02:44.134562 kernel: GDS: Unknown: Dependent on hypervisor status Feb 13 20:02:44.134675 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 13 20:02:44.134692 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 13 20:02:44.134708 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 13 20:02:44.134728 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Feb 13 20:02:44.134744 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Feb 13 20:02:44.134760 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Feb 13 20:02:44.134776 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Feb 13 20:02:44.134792 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Feb 13 20:02:44.134808 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Feb 13 20:02:44.134825 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 13 20:02:44.134841 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Feb 13 20:02:44.134857 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Feb 13 20:02:44.134872 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Feb 13 20:02:44.134888 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Feb 13 20:02:44.134907 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Feb 13 20:02:44.134923 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Feb 13 20:02:44.134939 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Feb 13 20:02:44.134956 kernel: Freeing SMP alternatives memory: 32K Feb 13 20:02:44.134971 kernel: pid_max: default: 32768 minimum: 301 Feb 13 20:02:44.134987 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 20:02:44.135003 kernel: landlock: Up and running. Feb 13 20:02:44.135019 kernel: SELinux: Initializing. Feb 13 20:02:44.135035 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 13 20:02:44.135051 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 13 20:02:44.135068 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) Feb 13 20:02:44.135087 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:02:44.135104 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:02:44.135120 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 20:02:44.135136 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Feb 13 20:02:44.135153 kernel: signal: max sigframe size: 3632 Feb 13 20:02:44.135169 kernel: rcu: Hierarchical SRCU implementation. Feb 13 20:02:44.135185 kernel: rcu: Max phase no-delay instances is 400. Feb 13 20:02:44.135201 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Feb 13 20:02:44.135217 kernel: smp: Bringing up secondary CPUs ... Feb 13 20:02:44.135237 kernel: smpboot: x86: Booting SMP configuration: Feb 13 20:02:44.135253 kernel: .... node #0, CPUs: #1 Feb 13 20:02:44.135270 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Feb 13 20:02:44.135287 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Feb 13 20:02:44.135302 kernel: smp: Brought up 1 node, 2 CPUs Feb 13 20:02:44.135318 kernel: smpboot: Max logical packages: 1 Feb 13 20:02:44.135334 kernel: smpboot: Total of 2 processors activated (9999.98 BogoMIPS) Feb 13 20:02:44.135351 kernel: devtmpfs: initialized Feb 13 20:02:44.135388 kernel: x86/mm: Memory block size: 128MB Feb 13 20:02:44.135405 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 20:02:44.135421 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 13 20:02:44.135437 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 20:02:44.135453 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 20:02:44.135469 kernel: audit: initializing netlink subsys (disabled) Feb 13 20:02:44.135485 kernel: audit: type=2000 audit(1739476963.180:1): state=initialized audit_enabled=0 res=1 Feb 13 20:02:44.135501 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 20:02:44.135518 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 13 20:02:44.135537 kernel: cpuidle: using governor menu Feb 13 20:02:44.135553 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 20:02:44.135569 kernel: dca service started, version 1.12.1 Feb 13 20:02:44.135585 kernel: PCI: Using configuration type 1 for base access Feb 13 20:02:44.135601 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 13 20:02:44.135617 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 20:02:44.135633 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 20:02:44.135650 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 20:02:44.135665 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 20:02:44.135832 kernel: ACPI: Added _OSI(Module Device) Feb 13 20:02:44.135850 kernel: ACPI: Added _OSI(Processor Device) Feb 13 20:02:44.135866 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 20:02:44.135883 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 20:02:44.135899 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Feb 13 20:02:44.135916 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Feb 13 20:02:44.135932 kernel: ACPI: Interpreter enabled Feb 13 20:02:44.135948 kernel: ACPI: PM: (supports S0 S5) Feb 13 20:02:44.135964 kernel: ACPI: Using IOAPIC for interrupt routing Feb 13 20:02:44.135980 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 13 20:02:44.136001 kernel: PCI: Using E820 reservations for host bridge windows Feb 13 20:02:44.136017 kernel: ACPI: Enabled 16 GPEs in block 00 to 0F Feb 13 20:02:44.136033 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 13 20:02:44.136344 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Feb 13 20:02:44.136519 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Feb 13 20:02:44.136654 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Feb 13 20:02:44.136674 kernel: acpiphp: Slot [3] registered Feb 13 20:02:44.136695 kernel: acpiphp: Slot [4] registered Feb 13 20:02:44.136711 kernel: acpiphp: Slot [5] registered Feb 13 20:02:44.136727 kernel: acpiphp: Slot [6] registered Feb 13 20:02:44.136743 kernel: acpiphp: Slot [7] registered Feb 13 20:02:44.136758 kernel: acpiphp: Slot [8] registered Feb 13 20:02:44.136775 kernel: acpiphp: Slot [9] registered Feb 13 20:02:44.136791 kernel: acpiphp: Slot [10] registered Feb 13 20:02:44.136807 kernel: acpiphp: Slot [11] registered Feb 13 20:02:44.136823 kernel: acpiphp: Slot [12] registered Feb 13 20:02:44.136842 kernel: acpiphp: Slot [13] registered Feb 13 20:02:44.136858 kernel: acpiphp: Slot [14] registered Feb 13 20:02:44.136874 kernel: acpiphp: Slot [15] registered Feb 13 20:02:44.136890 kernel: acpiphp: Slot [16] registered Feb 13 20:02:44.136906 kernel: acpiphp: Slot [17] registered Feb 13 20:02:44.136922 kernel: acpiphp: Slot [18] registered Feb 13 20:02:44.136949 kernel: acpiphp: Slot [19] registered Feb 13 20:02:44.136965 kernel: acpiphp: Slot [20] registered Feb 13 20:02:44.136980 kernel: acpiphp: Slot [21] registered Feb 13 20:02:44.136996 kernel: acpiphp: Slot [22] registered Feb 13 20:02:44.137014 kernel: acpiphp: Slot [23] registered Feb 13 20:02:44.137030 kernel: acpiphp: Slot [24] registered Feb 13 20:02:44.137045 kernel: acpiphp: Slot [25] registered Feb 13 20:02:44.137060 kernel: acpiphp: Slot [26] registered Feb 13 20:02:44.137074 kernel: acpiphp: Slot [27] registered Feb 13 20:02:44.137086 kernel: acpiphp: Slot [28] registered Feb 13 20:02:44.137100 kernel: acpiphp: Slot [29] registered Feb 13 20:02:44.137115 kernel: acpiphp: Slot [30] registered Feb 13 20:02:44.137130 kernel: acpiphp: Slot [31] registered Feb 13 20:02:44.137148 kernel: PCI host bridge to bus 0000:00 Feb 13 20:02:44.137272 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Feb 13 20:02:44.137525 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Feb 13 20:02:44.137644 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Feb 13 20:02:44.137756 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Feb 13 20:02:44.137866 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 13 20:02:44.138012 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Feb 13 20:02:44.138168 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Feb 13 20:02:44.138303 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 Feb 13 20:02:44.138479 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Feb 13 20:02:44.138632 kernel: pci 0000:00:01.3: quirk: [io 0xb100-0xb10f] claimed by PIIX4 SMB Feb 13 20:02:44.138761 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Feb 13 20:02:44.138892 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Feb 13 20:02:44.139014 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Feb 13 20:02:44.139143 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Feb 13 20:02:44.139267 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Feb 13 20:02:44.139409 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Feb 13 20:02:44.139531 kernel: pci 0000:00:01.3: quirk_piix4_acpi+0x0/0x180 took 10742 usecs Feb 13 20:02:44.139662 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 Feb 13 20:02:44.139785 kernel: pci 0000:00:03.0: reg 0x10: [mem 0xfe400000-0xfe7fffff pref] Feb 13 20:02:44.139912 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Feb 13 20:02:44.140033 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Feb 13 20:02:44.140163 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Feb 13 20:02:44.140288 kernel: pci 0000:00:04.0: reg 0x10: [mem 0xfebf0000-0xfebf3fff] Feb 13 20:02:44.141461 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Feb 13 20:02:44.141612 kernel: pci 0000:00:05.0: reg 0x10: [mem 0xfebf4000-0xfebf7fff] Feb 13 20:02:44.141634 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Feb 13 20:02:44.141657 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Feb 13 20:02:44.141674 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Feb 13 20:02:44.141690 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Feb 13 20:02:44.141706 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Feb 13 20:02:44.141722 kernel: iommu: Default domain type: Translated Feb 13 20:02:44.141738 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 13 20:02:44.141755 kernel: PCI: Using ACPI for IRQ routing Feb 13 20:02:44.141771 kernel: PCI: pci_cache_line_size set to 64 bytes Feb 13 20:02:44.141787 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Feb 13 20:02:44.141806 kernel: e820: reserve RAM buffer [mem 0x7d9ea000-0x7fffffff] Feb 13 20:02:44.141938 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Feb 13 20:02:44.142066 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Feb 13 20:02:44.142195 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Feb 13 20:02:44.142215 kernel: vgaarb: loaded Feb 13 20:02:44.142231 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Feb 13 20:02:44.142247 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Feb 13 20:02:44.142264 kernel: clocksource: Switched to clocksource kvm-clock Feb 13 20:02:44.142280 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 20:02:44.142300 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 20:02:44.142316 kernel: pnp: PnP ACPI init Feb 13 20:02:44.142332 kernel: pnp: PnP ACPI: found 5 devices Feb 13 20:02:44.142349 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 13 20:02:44.142375 kernel: NET: Registered PF_INET protocol family Feb 13 20:02:44.142392 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 20:02:44.142408 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Feb 13 20:02:44.142424 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 20:02:44.142439 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 13 20:02:44.142458 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Feb 13 20:02:44.142481 kernel: TCP: Hash tables configured (established 16384 bind 16384) Feb 13 20:02:44.142512 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 13 20:02:44.142543 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 13 20:02:44.142556 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 20:02:44.142651 kernel: NET: Registered PF_XDP protocol family Feb 13 20:02:44.142869 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Feb 13 20:02:44.143006 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Feb 13 20:02:44.143144 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Feb 13 20:02:44.143279 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Feb 13 20:02:44.143476 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Feb 13 20:02:44.143498 kernel: PCI: CLS 0 bytes, default 64 Feb 13 20:02:44.143513 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Feb 13 20:02:44.144639 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Feb 13 20:02:44.144658 kernel: clocksource: Switched to clocksource tsc Feb 13 20:02:44.144671 kernel: Initialise system trusted keyrings Feb 13 20:02:44.144691 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Feb 13 20:02:44.144704 kernel: Key type asymmetric registered Feb 13 20:02:44.144717 kernel: Asymmetric key parser 'x509' registered Feb 13 20:02:44.144730 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Feb 13 20:02:44.144743 kernel: io scheduler mq-deadline registered Feb 13 20:02:44.144757 kernel: io scheduler kyber registered Feb 13 20:02:44.144770 kernel: io scheduler bfq registered Feb 13 20:02:44.144784 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 13 20:02:44.144798 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 20:02:44.144816 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 13 20:02:44.144830 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Feb 13 20:02:44.144843 kernel: i8042: Warning: Keylock active Feb 13 20:02:44.144856 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Feb 13 20:02:44.144870 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Feb 13 20:02:44.145045 kernel: rtc_cmos 00:00: RTC can wake from S4 Feb 13 20:02:44.145173 kernel: rtc_cmos 00:00: registered as rtc0 Feb 13 20:02:44.145298 kernel: rtc_cmos 00:00: setting system clock to 2025-02-13T20:02:43 UTC (1739476963) Feb 13 20:02:44.145464 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Feb 13 20:02:44.145487 kernel: intel_pstate: CPU model not supported Feb 13 20:02:44.145505 kernel: NET: Registered PF_INET6 protocol family Feb 13 20:02:44.145522 kernel: Segment Routing with IPv6 Feb 13 20:02:44.145540 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 20:02:44.145556 kernel: NET: Registered PF_PACKET protocol family Feb 13 20:02:44.145573 kernel: Key type dns_resolver registered Feb 13 20:02:44.145590 kernel: IPI shorthand broadcast: enabled Feb 13 20:02:44.145607 kernel: sched_clock: Marking stable (748002191, 215899104)->(1045973994, -82072699) Feb 13 20:02:44.145629 kernel: registered taskstats version 1 Feb 13 20:02:44.145645 kernel: Loading compiled-in X.509 certificates Feb 13 20:02:44.145662 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 6e17590ca2768b672aa48f3e0cedc4061febfe93' Feb 13 20:02:44.145679 kernel: Key type .fscrypt registered Feb 13 20:02:44.145696 kernel: Key type fscrypt-provisioning registered Feb 13 20:02:44.145712 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 20:02:44.145729 kernel: ima: Allocated hash algorithm: sha1 Feb 13 20:02:44.145745 kernel: ima: No architecture policies found Feb 13 20:02:44.145761 kernel: clk: Disabling unused clocks Feb 13 20:02:44.145782 kernel: Freeing unused kernel image (initmem) memory: 42840K Feb 13 20:02:44.145799 kernel: Write protecting the kernel read-only data: 36864k Feb 13 20:02:44.145817 kernel: Freeing unused kernel image (rodata/data gap) memory: 1848K Feb 13 20:02:44.145832 kernel: Run /init as init process Feb 13 20:02:44.145849 kernel: with arguments: Feb 13 20:02:44.145866 kernel: /init Feb 13 20:02:44.145882 kernel: with environment: Feb 13 20:02:44.145899 kernel: HOME=/ Feb 13 20:02:44.145915 kernel: TERM=linux Feb 13 20:02:44.145935 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 20:02:44.145984 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 20:02:44.146006 systemd[1]: Detected virtualization amazon. Feb 13 20:02:44.146024 systemd[1]: Detected architecture x86-64. Feb 13 20:02:44.146042 systemd[1]: Running in initrd. Feb 13 20:02:44.146059 systemd[1]: No hostname configured, using default hostname. Feb 13 20:02:44.146076 systemd[1]: Hostname set to . Feb 13 20:02:44.146099 systemd[1]: Initializing machine ID from VM UUID. Feb 13 20:02:44.146116 systemd[1]: Queued start job for default target initrd.target. Feb 13 20:02:44.146134 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Feb 13 20:02:44.146153 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:02:44.146171 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:02:44.146191 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 20:02:44.146209 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 20:02:44.146228 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 20:02:44.146250 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 20:02:44.146272 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 20:02:44.146291 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 20:02:44.146309 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:02:44.146327 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:02:44.146346 systemd[1]: Reached target paths.target - Path Units. Feb 13 20:02:44.146398 systemd[1]: Reached target slices.target - Slice Units. Feb 13 20:02:44.146421 systemd[1]: Reached target swap.target - Swaps. Feb 13 20:02:44.146441 systemd[1]: Reached target timers.target - Timer Units. Feb 13 20:02:44.146458 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 20:02:44.146477 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 20:02:44.146496 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 20:02:44.146514 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Feb 13 20:02:44.146533 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:02:44.146550 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 20:02:44.146649 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:02:44.146670 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 20:02:44.146688 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 20:02:44.146706 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 20:02:44.146724 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 20:02:44.146749 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 20:02:44.146764 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 20:02:44.146779 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 20:02:44.146795 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:02:44.146813 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 20:02:44.146829 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:02:44.146845 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 20:02:44.146898 systemd-journald[178]: Collecting audit messages is disabled. Feb 13 20:02:44.146935 systemd-journald[178]: Journal started Feb 13 20:02:44.146973 systemd-journald[178]: Runtime Journal (/run/log/journal/ec230fb81b7ce66434108879c969aa20) is 4.8M, max 38.6M, 33.7M free. Feb 13 20:02:44.156424 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 20:02:44.147227 systemd-modules-load[179]: Inserted module 'overlay' Feb 13 20:02:44.162397 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 20:02:44.188414 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 20:02:44.190271 systemd-modules-load[179]: Inserted module 'br_netfilter' Feb 13 20:02:44.302556 kernel: Bridge firewalling registered Feb 13 20:02:44.305099 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 20:02:44.307807 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:02:44.311393 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 20:02:44.320688 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:02:44.327535 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 20:02:44.328427 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 20:02:44.349314 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 20:02:44.371751 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:02:44.376720 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:02:44.388623 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 20:02:44.391406 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:02:44.392914 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:02:44.401910 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 20:02:44.424743 dracut-cmdline[210]: dracut-dracut-053 Feb 13 20:02:44.436474 dracut-cmdline[210]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=a8740cbac5121ade856b040634ad9badacd879298c24f899668a59d96c178b13 Feb 13 20:02:44.473902 systemd-resolved[213]: Positive Trust Anchors: Feb 13 20:02:44.473923 systemd-resolved[213]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 20:02:44.473986 systemd-resolved[213]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 20:02:44.489548 systemd-resolved[213]: Defaulting to hostname 'linux'. Feb 13 20:02:44.492992 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 20:02:44.507205 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:02:44.573411 kernel: SCSI subsystem initialized Feb 13 20:02:44.583382 kernel: Loading iSCSI transport class v2.0-870. Feb 13 20:02:44.596395 kernel: iscsi: registered transport (tcp) Feb 13 20:02:44.632561 kernel: iscsi: registered transport (qla4xxx) Feb 13 20:02:44.632644 kernel: QLogic iSCSI HBA Driver Feb 13 20:02:44.676950 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 20:02:44.685074 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 20:02:44.747448 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 20:02:44.747644 kernel: device-mapper: uevent: version 1.0.3 Feb 13 20:02:44.747669 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 20:02:44.813502 kernel: raid6: avx512x4 gen() 12741 MB/s Feb 13 20:02:44.830470 kernel: raid6: avx512x2 gen() 10891 MB/s Feb 13 20:02:44.848417 kernel: raid6: avx512x1 gen() 8651 MB/s Feb 13 20:02:44.865412 kernel: raid6: avx2x4 gen() 12877 MB/s Feb 13 20:02:44.882408 kernel: raid6: avx2x2 gen() 7067 MB/s Feb 13 20:02:44.900104 kernel: raid6: avx2x1 gen() 8228 MB/s Feb 13 20:02:44.900178 kernel: raid6: using algorithm avx2x4 gen() 12877 MB/s Feb 13 20:02:44.917546 kernel: raid6: .... xor() 5192 MB/s, rmw enabled Feb 13 20:02:44.917626 kernel: raid6: using avx512x2 recovery algorithm Feb 13 20:02:44.944433 kernel: xor: automatically using best checksumming function avx Feb 13 20:02:45.132389 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 20:02:45.142606 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 20:02:45.150708 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:02:45.168216 systemd-udevd[397]: Using default interface naming scheme 'v255'. Feb 13 20:02:45.173633 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:02:45.181587 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 20:02:45.203218 dracut-pre-trigger[403]: rd.md=0: removing MD RAID activation Feb 13 20:02:45.236604 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 20:02:45.241925 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 20:02:45.323221 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:02:45.333615 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 20:02:45.376350 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 20:02:45.383744 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 20:02:45.385906 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:02:45.392666 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 20:02:45.409895 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 20:02:45.460090 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 20:02:45.477399 kernel: cryptd: max_cpu_qlen set to 1000 Feb 13 20:02:45.499046 kernel: ena 0000:00:05.0: ENA device version: 0.10 Feb 13 20:02:45.530166 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Feb 13 20:02:45.530392 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Feb 13 20:02:45.530589 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem febf4000, mac addr 06:15:e6:ab:44:07 Feb 13 20:02:45.503424 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 20:02:45.503586 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:02:45.506270 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:02:45.510445 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:02:45.510684 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:02:45.514503 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:02:45.526773 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:02:45.533912 (udev-worker)[456]: Network interface NamePolicy= disabled on kernel command line. Feb 13 20:02:45.549021 kernel: AVX2 version of gcm_enc/dec engaged. Feb 13 20:02:45.549083 kernel: AES CTR mode by8 optimization enabled Feb 13 20:02:45.575380 kernel: nvme nvme0: pci function 0000:00:04.0 Feb 13 20:02:45.575643 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Feb 13 20:02:45.588411 kernel: nvme nvme0: 2/0/0 default/read/poll queues Feb 13 20:02:45.593380 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 20:02:45.593449 kernel: GPT:9289727 != 16777215 Feb 13 20:02:45.593472 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 20:02:45.593494 kernel: GPT:9289727 != 16777215 Feb 13 20:02:45.593514 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 20:02:45.593535 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 20:02:45.709266 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:02:45.715691 kernel: BTRFS: device fsid 892c7470-7713-4b0f-880a-4c5f7bf5b72d devid 1 transid 37 /dev/nvme0n1p3 scanned by (udev-worker) (456) Feb 13 20:02:45.716623 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 20:02:45.720405 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 scanned by (udev-worker) (455) Feb 13 20:02:45.760742 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:02:45.817250 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Feb 13 20:02:45.836513 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Feb 13 20:02:45.844457 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Feb 13 20:02:45.844595 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Feb 13 20:02:45.859464 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 20:02:45.867518 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 20:02:45.882202 disk-uuid[626]: Primary Header is updated. Feb 13 20:02:45.882202 disk-uuid[626]: Secondary Entries is updated. Feb 13 20:02:45.882202 disk-uuid[626]: Secondary Header is updated. Feb 13 20:02:45.893529 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 20:02:45.898498 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 20:02:46.908431 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Feb 13 20:02:46.910198 disk-uuid[627]: The operation has completed successfully. Feb 13 20:02:47.078794 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 20:02:47.079019 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 20:02:47.102940 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 20:02:47.107084 sh[885]: Success Feb 13 20:02:47.121405 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Feb 13 20:02:47.240072 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 20:02:47.252500 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 20:02:47.254837 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 20:02:47.299790 kernel: BTRFS info (device dm-0): first mount of filesystem 892c7470-7713-4b0f-880a-4c5f7bf5b72d Feb 13 20:02:47.299856 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:02:47.299884 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 20:02:47.301551 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 20:02:47.301579 kernel: BTRFS info (device dm-0): using free space tree Feb 13 20:02:47.398454 kernel: BTRFS info (device dm-0): enabling ssd optimizations Feb 13 20:02:47.412290 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 20:02:47.413141 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 20:02:47.419554 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 20:02:47.433216 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 20:02:47.466044 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem b405b664-b121-4411-9ed3-1128bc9da790 Feb 13 20:02:47.466117 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:02:47.466246 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 20:02:47.473391 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 20:02:47.510751 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 20:02:47.513452 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem b405b664-b121-4411-9ed3-1128bc9da790 Feb 13 20:02:47.523984 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 20:02:47.534200 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 20:02:47.594344 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 20:02:47.600758 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 20:02:47.639483 systemd-networkd[1077]: lo: Link UP Feb 13 20:02:47.639493 systemd-networkd[1077]: lo: Gained carrier Feb 13 20:02:47.646660 systemd-networkd[1077]: Enumeration completed Feb 13 20:02:47.647825 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 20:02:47.649587 systemd[1]: Reached target network.target - Network. Feb 13 20:02:47.652567 systemd-networkd[1077]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:02:47.652572 systemd-networkd[1077]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 20:02:47.663234 systemd-networkd[1077]: eth0: Link UP Feb 13 20:02:47.663244 systemd-networkd[1077]: eth0: Gained carrier Feb 13 20:02:47.663258 systemd-networkd[1077]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:02:47.683598 systemd-networkd[1077]: eth0: DHCPv4 address 172.31.25.222/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 20:02:48.004768 ignition[1018]: Ignition 2.19.0 Feb 13 20:02:48.004781 ignition[1018]: Stage: fetch-offline Feb 13 20:02:48.004993 ignition[1018]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:02:48.005002 ignition[1018]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 20:02:48.006167 ignition[1018]: Ignition finished successfully Feb 13 20:02:48.011708 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 20:02:48.020953 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Feb 13 20:02:48.044559 ignition[1085]: Ignition 2.19.0 Feb 13 20:02:48.044684 ignition[1085]: Stage: fetch Feb 13 20:02:48.045847 ignition[1085]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:02:48.045880 ignition[1085]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 20:02:48.046007 ignition[1085]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 20:02:48.113113 ignition[1085]: PUT result: OK Feb 13 20:02:48.120586 ignition[1085]: parsed url from cmdline: "" Feb 13 20:02:48.120598 ignition[1085]: no config URL provided Feb 13 20:02:48.120610 ignition[1085]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 20:02:48.120625 ignition[1085]: no config at "/usr/lib/ignition/user.ign" Feb 13 20:02:48.120651 ignition[1085]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 20:02:48.125892 ignition[1085]: PUT result: OK Feb 13 20:02:48.125985 ignition[1085]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Feb 13 20:02:48.131231 ignition[1085]: GET result: OK Feb 13 20:02:48.131332 ignition[1085]: parsing config with SHA512: d2b01b3ab0603b79b3a2de2507c9440d618d161c7f2a4e9cb4220df6b7079fd624888ffa96f9311aa82da53fb501ec364f82c94d4bf01d1a6493de43f9a0288e Feb 13 20:02:48.136694 unknown[1085]: fetched base config from "system" Feb 13 20:02:48.136710 unknown[1085]: fetched base config from "system" Feb 13 20:02:48.138395 ignition[1085]: fetch: fetch complete Feb 13 20:02:48.136720 unknown[1085]: fetched user config from "aws" Feb 13 20:02:48.138404 ignition[1085]: fetch: fetch passed Feb 13 20:02:48.138474 ignition[1085]: Ignition finished successfully Feb 13 20:02:48.143871 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Feb 13 20:02:48.151988 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 20:02:48.205225 ignition[1092]: Ignition 2.19.0 Feb 13 20:02:48.205240 ignition[1092]: Stage: kargs Feb 13 20:02:48.206283 ignition[1092]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:02:48.206298 ignition[1092]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 20:02:48.207057 ignition[1092]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 20:02:48.210674 ignition[1092]: PUT result: OK Feb 13 20:02:48.213420 ignition[1092]: kargs: kargs passed Feb 13 20:02:48.213494 ignition[1092]: Ignition finished successfully Feb 13 20:02:48.216255 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 20:02:48.230004 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 20:02:48.280840 ignition[1098]: Ignition 2.19.0 Feb 13 20:02:48.280855 ignition[1098]: Stage: disks Feb 13 20:02:48.281374 ignition[1098]: no configs at "/usr/lib/ignition/base.d" Feb 13 20:02:48.281390 ignition[1098]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 20:02:48.281498 ignition[1098]: PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 20:02:48.283759 ignition[1098]: PUT result: OK Feb 13 20:02:48.289042 ignition[1098]: disks: disks passed Feb 13 20:02:48.289106 ignition[1098]: Ignition finished successfully Feb 13 20:02:48.290086 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 20:02:48.293035 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 20:02:48.294953 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 20:02:48.297627 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 20:02:48.300559 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 20:02:48.301817 systemd[1]: Reached target basic.target - Basic System. Feb 13 20:02:48.309527 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 20:02:48.362566 systemd-fsck[1106]: ROOT: clean, 14/553520 files, 52654/553472 blocks Feb 13 20:02:48.365484 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 20:02:48.371498 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 20:02:48.567630 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 85215ce4-0be3-4782-863e-8dde129924f0 r/w with ordered data mode. Quota mode: none. Feb 13 20:02:48.568244 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 20:02:48.568940 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 20:02:48.589518 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 20:02:48.597491 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 20:02:48.599964 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Feb 13 20:02:48.602121 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 20:02:48.602161 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 20:02:48.615932 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 20:02:48.622606 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 20:02:48.633405 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/nvme0n1p6 scanned by mount (1125) Feb 13 20:02:48.638719 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem b405b664-b121-4411-9ed3-1128bc9da790 Feb 13 20:02:48.638791 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:02:48.638819 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 20:02:48.651741 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 20:02:48.653215 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 20:02:48.912536 systemd-networkd[1077]: eth0: Gained IPv6LL Feb 13 20:02:49.083502 initrd-setup-root[1149]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 20:02:49.130493 initrd-setup-root[1156]: cut: /sysroot/etc/group: No such file or directory Feb 13 20:02:49.142343 initrd-setup-root[1163]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 20:02:49.164738 initrd-setup-root[1170]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 20:02:49.545176 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 20:02:49.558565 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 20:02:49.563291 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 20:02:49.607515 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 20:02:49.608689 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem b405b664-b121-4411-9ed3-1128bc9da790 Feb 13 20:02:49.654871 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 20:02:49.658378 ignition[1237]: INFO : Ignition 2.19.0 Feb 13 20:02:49.659968 ignition[1237]: INFO : Stage: mount Feb 13 20:02:49.661395 ignition[1237]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:02:49.661395 ignition[1237]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 20:02:49.665993 ignition[1237]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 20:02:49.668050 ignition[1237]: INFO : PUT result: OK Feb 13 20:02:49.672791 ignition[1237]: INFO : mount: mount passed Feb 13 20:02:49.674428 ignition[1237]: INFO : Ignition finished successfully Feb 13 20:02:49.679684 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 20:02:49.690560 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 20:02:49.708145 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 20:02:49.739818 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by mount (1249) Feb 13 20:02:49.739882 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem b405b664-b121-4411-9ed3-1128bc9da790 Feb 13 20:02:49.739901 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Feb 13 20:02:49.741383 kernel: BTRFS info (device nvme0n1p6): using free space tree Feb 13 20:02:49.747273 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Feb 13 20:02:49.750288 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 20:02:49.791029 ignition[1266]: INFO : Ignition 2.19.0 Feb 13 20:02:49.791029 ignition[1266]: INFO : Stage: files Feb 13 20:02:49.794055 ignition[1266]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:02:49.794055 ignition[1266]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 20:02:49.794055 ignition[1266]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 20:02:49.798407 ignition[1266]: INFO : PUT result: OK Feb 13 20:02:49.800871 ignition[1266]: DEBUG : files: compiled without relabeling support, skipping Feb 13 20:02:49.803575 ignition[1266]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 20:02:49.803575 ignition[1266]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 20:02:49.820778 ignition[1266]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 20:02:49.822675 ignition[1266]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 20:02:49.824526 unknown[1266]: wrote ssh authorized keys file for user: core Feb 13 20:02:49.826333 ignition[1266]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 20:02:49.829190 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 20:02:49.831839 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Feb 13 20:02:49.953716 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 20:02:50.167344 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 20:02:50.167344 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:02:50.172108 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 Feb 13 20:02:50.666105 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 13 20:02:51.091658 ignition[1266]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 20:02:51.091658 ignition[1266]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Feb 13 20:02:51.098063 ignition[1266]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 20:02:51.100781 ignition[1266]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 20:02:51.100781 ignition[1266]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Feb 13 20:02:51.107907 ignition[1266]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Feb 13 20:02:51.107907 ignition[1266]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 20:02:51.111274 ignition[1266]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 20:02:51.113445 ignition[1266]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 20:02:51.113445 ignition[1266]: INFO : files: files passed Feb 13 20:02:51.113445 ignition[1266]: INFO : Ignition finished successfully Feb 13 20:02:51.114814 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 20:02:51.122802 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 20:02:51.126615 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 20:02:51.160963 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 20:02:51.163151 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 20:02:51.181840 initrd-setup-root-after-ignition[1294]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:02:51.181840 initrd-setup-root-after-ignition[1294]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:02:51.191064 initrd-setup-root-after-ignition[1298]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 20:02:51.194944 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 20:02:51.195284 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 20:02:51.202636 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 20:02:51.239952 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 20:02:51.240061 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 20:02:51.243336 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 20:02:51.246694 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 20:02:51.249592 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 20:02:51.255764 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 20:02:51.287888 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 20:02:51.310197 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 20:02:51.349698 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:02:51.350126 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:02:51.356487 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 20:02:51.362450 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 20:02:51.364745 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 20:02:51.369761 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 20:02:51.372066 systemd[1]: Stopped target basic.target - Basic System. Feb 13 20:02:51.374712 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 20:02:51.378022 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 20:02:51.382386 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 20:02:51.385525 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 20:02:51.385720 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 20:02:51.390544 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 20:02:51.391893 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 20:02:51.393912 systemd[1]: Stopped target swap.target - Swaps. Feb 13 20:02:51.396920 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 20:02:51.398054 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 20:02:51.400562 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:02:51.402824 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:02:51.406201 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 20:02:51.406316 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:02:51.410378 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 20:02:51.411552 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 20:02:51.414139 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 20:02:51.418024 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 20:02:51.420700 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 20:02:51.421832 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 20:02:51.433648 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 20:02:51.435387 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 20:02:51.435535 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:02:51.442285 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 20:02:51.446767 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 20:02:51.447016 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:02:51.448821 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 20:02:51.488696 ignition[1318]: INFO : Ignition 2.19.0 Feb 13 20:02:51.488696 ignition[1318]: INFO : Stage: umount Feb 13 20:02:51.488696 ignition[1318]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 20:02:51.488696 ignition[1318]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Feb 13 20:02:51.488696 ignition[1318]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Feb 13 20:02:51.448990 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 20:02:51.497183 ignition[1318]: INFO : PUT result: OK Feb 13 20:02:51.498953 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 20:02:51.499194 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 20:02:51.509744 ignition[1318]: INFO : umount: umount passed Feb 13 20:02:51.509744 ignition[1318]: INFO : Ignition finished successfully Feb 13 20:02:51.511985 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 20:02:51.512151 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 20:02:51.518082 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 20:02:51.518226 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 20:02:51.520257 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 20:02:51.520327 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 20:02:51.522244 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 13 20:02:51.522294 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Feb 13 20:02:51.523694 systemd[1]: Stopped target network.target - Network. Feb 13 20:02:51.525599 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 20:02:51.525652 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 20:02:51.527210 systemd[1]: Stopped target paths.target - Path Units. Feb 13 20:02:51.528178 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 20:02:51.530753 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:02:51.534786 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 20:02:51.538999 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 20:02:51.542114 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 20:02:51.545986 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 20:02:51.551977 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 20:02:51.552041 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 20:02:51.557686 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 20:02:51.557799 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 20:02:51.562449 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 20:02:51.562536 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 20:02:51.571315 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 20:02:51.573931 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 20:02:51.579168 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 20:02:51.585415 systemd-networkd[1077]: eth0: DHCPv6 lease lost Feb 13 20:02:51.588525 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 20:02:51.588676 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 20:02:51.606809 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 20:02:51.614102 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 20:02:51.639149 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 20:02:51.639227 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:02:51.652600 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 20:02:51.653736 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 20:02:51.653814 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 20:02:51.658673 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 20:02:51.658751 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:02:51.658885 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 20:02:51.658934 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 20:02:51.666003 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 20:02:51.666077 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:02:51.672654 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:02:51.705455 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 20:02:51.707183 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:02:51.712472 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 20:02:51.712923 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 20:02:51.718326 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 20:02:51.718499 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 20:02:51.722972 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 20:02:51.723089 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:02:51.727097 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 20:02:51.727226 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 20:02:51.731298 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 20:02:51.731373 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 20:02:51.732753 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 20:02:51.732811 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 20:02:51.751039 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 20:02:51.753086 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 20:02:51.753177 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:02:51.754548 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 20:02:51.754742 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:02:51.764847 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 20:02:51.765122 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 20:02:51.768045 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 20:02:51.768173 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 20:02:51.776274 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 20:02:51.776984 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 20:02:51.781378 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 20:02:51.789249 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 20:02:51.800537 systemd[1]: Switching root. Feb 13 20:02:51.852100 systemd-journald[178]: Journal stopped Feb 13 20:02:54.123454 systemd-journald[178]: Received SIGTERM from PID 1 (systemd). Feb 13 20:02:54.123538 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 20:02:54.123562 kernel: SELinux: policy capability open_perms=1 Feb 13 20:02:54.123581 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 20:02:54.123598 kernel: SELinux: policy capability always_check_network=0 Feb 13 20:02:54.123621 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 20:02:54.123639 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 20:02:54.123656 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 20:02:54.123673 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 20:02:54.123695 kernel: audit: type=1403 audit(1739476972.385:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 20:02:54.123716 systemd[1]: Successfully loaded SELinux policy in 85.488ms. Feb 13 20:02:54.123747 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 18.880ms. Feb 13 20:02:54.123769 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 20:02:54.123790 systemd[1]: Detected virtualization amazon. Feb 13 20:02:54.123809 systemd[1]: Detected architecture x86-64. Feb 13 20:02:54.123827 systemd[1]: Detected first boot. Feb 13 20:02:54.123846 systemd[1]: Initializing machine ID from VM UUID. Feb 13 20:02:54.123866 zram_generator::config[1361]: No configuration found. Feb 13 20:02:54.123889 systemd[1]: Populated /etc with preset unit settings. Feb 13 20:02:54.123907 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 20:02:54.123927 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 20:02:54.123947 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 20:02:54.123971 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 20:02:54.123990 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 20:02:54.124008 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 20:02:54.124028 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 20:02:54.124047 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 20:02:54.124070 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 20:02:54.124094 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 20:02:54.124116 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 20:02:54.124137 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 20:02:54.124158 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 20:02:54.124180 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 20:02:54.124201 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 20:02:54.124221 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 20:02:54.124241 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 20:02:54.124267 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Feb 13 20:02:54.124290 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 20:02:54.124387 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 20:02:54.124410 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 20:02:54.124432 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 20:02:54.124450 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 20:02:54.124471 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 20:02:54.124499 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 20:02:54.124522 systemd[1]: Reached target slices.target - Slice Units. Feb 13 20:02:54.124545 systemd[1]: Reached target swap.target - Swaps. Feb 13 20:02:54.124567 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 20:02:54.124589 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 20:02:54.124612 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 20:02:54.124636 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 20:02:54.124659 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 20:02:54.124683 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 20:02:54.124705 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 20:02:54.124734 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 20:02:54.124758 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 20:02:54.124782 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:54.124804 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 20:02:54.124828 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 20:02:54.124849 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 20:02:54.124871 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 20:02:54.124893 systemd[1]: Reached target machines.target - Containers. Feb 13 20:02:54.124919 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 20:02:54.124941 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:02:54.124964 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 20:02:54.124983 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 20:02:54.125003 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 20:02:54.125023 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 20:02:54.125043 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 20:02:54.125064 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 20:02:54.125084 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 20:02:54.125108 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 20:02:54.125129 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 20:02:54.125150 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 20:02:54.125170 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 20:02:54.125192 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 20:02:54.125212 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 20:02:54.125235 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 20:02:54.125259 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 20:02:54.125286 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 20:02:54.125310 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 20:02:54.125335 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 20:02:54.127409 systemd[1]: Stopped verity-setup.service. Feb 13 20:02:54.127454 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:54.127483 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 20:02:54.127503 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 20:02:54.127522 kernel: fuse: init (API version 7.39) Feb 13 20:02:54.127541 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 20:02:54.127566 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 20:02:54.127585 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 20:02:54.127605 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 20:02:54.127624 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 20:02:54.127643 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 20:02:54.127666 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 20:02:54.127685 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 20:02:54.127704 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 20:02:54.127723 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 20:02:54.127742 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 20:02:54.127762 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 20:02:54.127781 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 20:02:54.127799 kernel: loop: module loaded Feb 13 20:02:54.127820 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 20:02:54.127842 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 20:02:54.127861 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 20:02:54.127914 systemd-journald[1433]: Collecting audit messages is disabled. Feb 13 20:02:54.127950 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 20:02:54.127974 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 20:02:54.127995 systemd-journald[1433]: Journal started Feb 13 20:02:54.128030 systemd-journald[1433]: Runtime Journal (/run/log/journal/ec230fb81b7ce66434108879c969aa20) is 4.8M, max 38.6M, 33.7M free. Feb 13 20:02:53.594293 systemd[1]: Queued start job for default target multi-user.target. Feb 13 20:02:53.689993 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Feb 13 20:02:53.690440 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 20:02:54.142110 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 20:02:54.160820 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 20:02:54.160921 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 20:02:54.171440 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 20:02:54.176381 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 20:02:54.178605 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 20:02:54.180527 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 20:02:54.182002 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 20:02:54.245055 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 20:02:54.245249 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 20:02:54.250104 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Feb 13 20:02:54.309044 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 20:02:54.312326 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 20:02:54.315796 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:02:54.318974 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 20:02:54.321559 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 20:02:54.323666 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 20:02:54.327112 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 20:02:54.331092 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 20:02:54.334937 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 20:02:54.358872 kernel: ACPI: bus type drm_connector registered Feb 13 20:02:54.371844 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 20:02:54.372070 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 20:02:54.374187 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 20:02:54.392759 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 20:02:54.398679 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 20:02:54.405878 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 20:02:54.407623 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 20:02:54.414279 systemd-journald[1433]: Time spent on flushing to /var/log/journal/ec230fb81b7ce66434108879c969aa20 is 51.923ms for 962 entries. Feb 13 20:02:54.414279 systemd-journald[1433]: System Journal (/var/log/journal/ec230fb81b7ce66434108879c969aa20) is 8.0M, max 195.6M, 187.6M free. Feb 13 20:02:54.478752 systemd-journald[1433]: Received client request to flush runtime journal. Feb 13 20:02:54.478824 kernel: loop0: detected capacity change from 0 to 61336 Feb 13 20:02:54.419245 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 20:02:54.425558 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Feb 13 20:02:54.435666 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 20:02:54.452133 udevadm[1493]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Feb 13 20:02:54.481714 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 20:02:54.506068 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 20:02:54.509267 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Feb 13 20:02:54.535726 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 20:02:54.562878 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 20:02:54.574519 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 20:02:54.577566 kernel: loop1: detected capacity change from 0 to 140768 Feb 13 20:02:54.642706 systemd-tmpfiles[1507]: ACLs are not supported, ignoring. Feb 13 20:02:54.642734 systemd-tmpfiles[1507]: ACLs are not supported, ignoring. Feb 13 20:02:54.650728 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 20:02:54.711389 kernel: loop2: detected capacity change from 0 to 218376 Feb 13 20:02:54.863993 kernel: loop3: detected capacity change from 0 to 142488 Feb 13 20:02:55.025564 kernel: loop4: detected capacity change from 0 to 61336 Feb 13 20:02:55.056427 kernel: loop5: detected capacity change from 0 to 140768 Feb 13 20:02:55.097331 kernel: loop6: detected capacity change from 0 to 218376 Feb 13 20:02:55.163587 kernel: loop7: detected capacity change from 0 to 142488 Feb 13 20:02:55.198338 (sd-merge)[1513]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Feb 13 20:02:55.202404 (sd-merge)[1513]: Merged extensions into '/usr'. Feb 13 20:02:55.219282 systemd[1]: Reloading requested from client PID 1485 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 20:02:55.219657 systemd[1]: Reloading... Feb 13 20:02:55.372422 zram_generator::config[1542]: No configuration found. Feb 13 20:02:55.597227 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:02:55.725047 systemd[1]: Reloading finished in 504 ms. Feb 13 20:02:55.757914 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 20:02:55.766717 systemd[1]: Starting ensure-sysext.service... Feb 13 20:02:55.778271 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 20:02:55.800339 systemd[1]: Reloading requested from client PID 1587 ('systemctl') (unit ensure-sysext.service)... Feb 13 20:02:55.800375 systemd[1]: Reloading... Feb 13 20:02:55.834730 systemd-tmpfiles[1588]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 20:02:55.835550 systemd-tmpfiles[1588]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 20:02:55.836875 systemd-tmpfiles[1588]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 20:02:55.843736 systemd-tmpfiles[1588]: ACLs are not supported, ignoring. Feb 13 20:02:55.846544 systemd-tmpfiles[1588]: ACLs are not supported, ignoring. Feb 13 20:02:55.858185 systemd-tmpfiles[1588]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 20:02:55.858201 systemd-tmpfiles[1588]: Skipping /boot Feb 13 20:02:55.893635 systemd-tmpfiles[1588]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 20:02:55.893655 systemd-tmpfiles[1588]: Skipping /boot Feb 13 20:02:55.941952 zram_generator::config[1616]: No configuration found. Feb 13 20:02:56.119389 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:02:56.190225 systemd[1]: Reloading finished in 389 ms. Feb 13 20:02:56.209307 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 20:02:56.214979 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 20:02:56.230645 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Feb 13 20:02:56.236612 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 20:02:56.241700 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 20:02:56.246525 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 20:02:56.254311 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 20:02:56.259660 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 20:02:56.272453 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:56.272986 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:02:56.280536 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 20:02:56.286407 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 20:02:56.300260 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 20:02:56.303815 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:02:56.304096 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:56.312048 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:56.312461 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:02:56.312727 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:02:56.341472 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 20:02:56.343202 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:56.344340 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 20:02:56.344583 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 20:02:56.354245 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 20:02:56.354458 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 20:02:56.367821 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 20:02:56.402724 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:56.403081 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 20:02:56.416928 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 20:02:56.426807 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 20:02:56.439964 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 20:02:56.447563 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 20:02:56.447909 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 20:02:56.451110 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 20:02:56.452723 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 20:02:56.458044 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 20:02:56.458855 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 20:02:56.484856 systemd[1]: Finished ensure-sysext.service. Feb 13 20:02:56.511462 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 20:02:56.515761 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 20:02:56.516770 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 20:02:56.517951 systemd-udevd[1673]: Using default interface naming scheme 'v255'. Feb 13 20:02:56.522009 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 20:02:56.530208 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 20:02:56.530896 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 20:02:56.549058 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 20:02:56.550474 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 20:02:56.552752 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 20:02:56.570765 augenrules[1704]: No rules Feb 13 20:02:56.572534 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Feb 13 20:02:56.590297 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 20:02:56.622600 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 20:02:56.634743 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 20:02:56.660656 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 20:02:56.663956 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 20:02:56.761468 systemd-resolved[1672]: Positive Trust Anchors: Feb 13 20:02:56.761964 systemd-resolved[1672]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 20:02:56.762350 systemd-resolved[1672]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 20:02:56.795933 systemd-resolved[1672]: Defaulting to hostname 'linux'. Feb 13 20:02:56.800469 systemd-networkd[1718]: lo: Link UP Feb 13 20:02:56.800934 systemd-networkd[1718]: lo: Gained carrier Feb 13 20:02:56.802805 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 20:02:56.804590 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 20:02:56.805233 systemd-networkd[1718]: Enumeration completed Feb 13 20:02:56.806309 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 20:02:56.808026 systemd[1]: Reached target network.target - Network. Feb 13 20:02:56.819696 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 20:02:56.836758 ldconfig[1481]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 20:02:56.850064 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 20:02:56.862859 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 20:02:56.864721 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Feb 13 20:02:56.890987 (udev-worker)[1719]: Network interface NamePolicy= disabled on kernel command line. Feb 13 20:02:56.910674 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 20:02:56.919153 systemd-networkd[1718]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:02:56.919170 systemd-networkd[1718]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 20:02:56.923152 systemd-networkd[1718]: eth0: Link UP Feb 13 20:02:56.923422 systemd-networkd[1718]: eth0: Gained carrier Feb 13 20:02:56.923448 systemd-networkd[1718]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 20:02:56.933461 systemd-networkd[1718]: eth0: DHCPv4 address 172.31.25.222/20, gateway 172.31.16.1 acquired from 172.31.16.1 Feb 13 20:02:56.962469 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (1723) Feb 13 20:02:57.035393 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Feb 13 20:02:57.050462 kernel: ACPI: button: Power Button [PWRF] Feb 13 20:02:57.054581 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input4 Feb 13 20:02:57.059421 kernel: ACPI: button: Sleep Button [SLPF] Feb 13 20:02:57.089412 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input5 Feb 13 20:02:57.091382 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0xb100, revision 255 Feb 13 20:02:57.178997 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 20:02:57.188394 kernel: mousedev: PS/2 mouse device common for all mice Feb 13 20:02:57.211932 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Feb 13 20:02:57.214026 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 20:02:57.224541 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 20:02:57.226805 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 20:02:57.266749 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 20:02:57.270787 lvm[1834]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 20:02:57.305992 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 20:02:57.308154 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 20:02:57.314644 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 20:02:57.330400 lvm[1841]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 20:02:57.366515 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 20:02:57.484908 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 20:02:57.488915 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 20:02:57.490491 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 20:02:57.492174 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 20:02:57.494093 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 20:02:57.495955 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 20:02:57.499124 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 20:02:57.500716 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 20:02:57.500766 systemd[1]: Reached target paths.target - Path Units. Feb 13 20:02:57.501753 systemd[1]: Reached target timers.target - Timer Units. Feb 13 20:02:57.504220 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 20:02:57.507633 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 20:02:57.513926 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 20:02:57.516388 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 20:02:57.518872 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 20:02:57.521593 systemd[1]: Reached target basic.target - Basic System. Feb 13 20:02:57.522941 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 20:02:57.522974 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 20:02:57.531560 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 20:02:57.535079 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Feb 13 20:02:57.541697 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 20:02:57.552801 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 20:02:57.560070 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 20:02:57.562205 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 20:02:57.564095 jq[1850]: false Feb 13 20:02:57.583859 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 20:02:57.596388 systemd[1]: Started ntpd.service - Network Time Service. Feb 13 20:02:57.615285 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 20:02:57.655173 systemd[1]: Starting setup-oem.service - Setup OEM... Feb 13 20:02:57.660079 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 20:02:57.663595 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 20:02:57.698993 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 20:02:57.701093 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 20:02:57.701906 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 20:02:57.710186 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 20:02:57.734249 extend-filesystems[1851]: Found loop4 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found loop5 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found loop6 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found loop7 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found nvme0n1 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found nvme0n1p1 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found nvme0n1p2 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found nvme0n1p3 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found usr Feb 13 20:02:57.734249 extend-filesystems[1851]: Found nvme0n1p4 Feb 13 20:02:57.734249 extend-filesystems[1851]: Found nvme0n1p6 Feb 13 20:02:57.758797 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 20:02:57.767015 extend-filesystems[1851]: Found nvme0n1p7 Feb 13 20:02:57.767015 extend-filesystems[1851]: Found nvme0n1p9 Feb 13 20:02:57.767015 extend-filesystems[1851]: Checking size of /dev/nvme0n1p9 Feb 13 20:02:57.784846 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 20:02:57.785105 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 20:02:57.786057 ntpd[1853]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 17:30:53 UTC 2025 (1): Starting Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: ntpd 4.2.8p17@1.4004-o Thu Feb 13 17:30:53 UTC 2025 (1): Starting Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: ---------------------------------------------------- Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: ntp-4 is maintained by Network Time Foundation, Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: corporation. Support and training for ntp-4 are Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: available at https://www.nwtime.org/support Feb 13 20:02:57.788480 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: ---------------------------------------------------- Feb 13 20:02:57.786089 ntpd[1853]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Feb 13 20:02:57.786100 ntpd[1853]: ---------------------------------------------------- Feb 13 20:02:57.786110 ntpd[1853]: ntp-4 is maintained by Network Time Foundation, Feb 13 20:02:57.786119 ntpd[1853]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Feb 13 20:02:57.786192 ntpd[1853]: corporation. Support and training for ntp-4 are Feb 13 20:02:57.804601 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 20:02:57.810209 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: proto: precision = 0.075 usec (-24) Feb 13 20:02:57.810209 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: basedate set to 2025-02-01 Feb 13 20:02:57.810209 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: gps base set to 2025-02-02 (week 2352) Feb 13 20:02:57.786202 ntpd[1853]: available at https://www.nwtime.org/support Feb 13 20:02:57.804862 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 20:02:57.786213 ntpd[1853]: ---------------------------------------------------- Feb 13 20:02:57.808342 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 20:02:57.794354 ntpd[1853]: proto: precision = 0.075 usec (-24) Feb 13 20:02:57.794733 ntpd[1853]: basedate set to 2025-02-01 Feb 13 20:02:57.794750 ntpd[1853]: gps base set to 2025-02-02 (week 2352) Feb 13 20:02:57.806522 dbus-daemon[1849]: [system] SELinux support is enabled Feb 13 20:02:57.823336 dbus-daemon[1849]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1718 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Feb 13 20:02:57.846689 update_engine[1865]: I20250213 20:02:57.844439 1865 main.cc:92] Flatcar Update Engine starting Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Listen normally on 3 eth0 172.31.25.222:123 Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Listen normally on 4 lo [::1]:123 Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: bind(21) AF_INET6 fe80::415:e6ff:feab:4407%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: unable to create socket on eth0 (5) for fe80::415:e6ff:feab:4407%2#123 Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: failed to init interface for address fe80::415:e6ff:feab:4407%2 Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: Listening on routing socket on fd #21 for interface updates Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 20:02:57.853079 ntpd[1853]: 13 Feb 20:02:57 ntpd[1853]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 20:02:57.838275 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 20:02:57.824663 ntpd[1853]: Listen and drop on 0 v6wildcard [::]:123 Feb 13 20:02:57.865729 update_engine[1865]: I20250213 20:02:57.847419 1865 update_check_scheduler.cc:74] Next update check in 3m43s Feb 13 20:02:57.838982 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 20:02:57.824718 ntpd[1853]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Feb 13 20:02:57.827567 ntpd[1853]: Listen normally on 2 lo 127.0.0.1:123 Feb 13 20:02:57.885897 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 20:02:57.827612 ntpd[1853]: Listen normally on 3 eth0 172.31.25.222:123 Feb 13 20:02:57.887959 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 20:02:57.827654 ntpd[1853]: Listen normally on 4 lo [::1]:123 Feb 13 20:02:57.887992 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 20:02:57.827704 ntpd[1853]: bind(21) AF_INET6 fe80::415:e6ff:feab:4407%2#123 flags 0x11 failed: Cannot assign requested address Feb 13 20:02:57.891621 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 20:02:57.827727 ntpd[1853]: unable to create socket on eth0 (5) for fe80::415:e6ff:feab:4407%2#123 Feb 13 20:02:57.891702 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 20:02:57.827742 ntpd[1853]: failed to init interface for address fe80::415:e6ff:feab:4407%2 Feb 13 20:02:57.827877 ntpd[1853]: Listening on routing socket on fd #21 for interface updates Feb 13 20:02:57.847764 ntpd[1853]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 20:02:57.847799 ntpd[1853]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Feb 13 20:02:57.910387 jq[1870]: true Feb 13 20:02:57.904819 (ntainerd)[1887]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 20:02:57.902469 dbus-daemon[1849]: [system] Successfully activated service 'org.freedesktop.systemd1' Feb 13 20:02:57.911751 systemd[1]: Started update-engine.service - Update Engine. Feb 13 20:02:57.923628 extend-filesystems[1851]: Resized partition /dev/nvme0n1p9 Feb 13 20:02:57.940519 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Feb 13 20:02:57.943389 extend-filesystems[1897]: resize2fs 1.47.1 (20-May-2024) Feb 13 20:02:57.960664 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 20:02:57.974960 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Feb 13 20:02:57.983335 jq[1894]: true Feb 13 20:02:58.020189 tar[1875]: linux-amd64/LICENSE Feb 13 20:02:58.020529 tar[1875]: linux-amd64/helm Feb 13 20:02:58.060558 systemd[1]: Finished setup-oem.service - Setup OEM. Feb 13 20:02:58.068235 systemd-logind[1863]: Watching system buttons on /dev/input/event1 (Power Button) Feb 13 20:02:58.072538 systemd-logind[1863]: Watching system buttons on /dev/input/event2 (Sleep Button) Feb 13 20:02:58.072576 systemd-logind[1863]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Feb 13 20:02:58.072863 systemd-logind[1863]: New seat seat0. Feb 13 20:02:58.074224 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 20:02:58.128718 systemd-networkd[1718]: eth0: Gained IPv6LL Feb 13 20:02:58.142530 coreos-metadata[1848]: Feb 13 20:02:58.138 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 20:02:58.142530 coreos-metadata[1848]: Feb 13 20:02:58.138 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Feb 13 20:02:58.142530 coreos-metadata[1848]: Feb 13 20:02:58.138 INFO Fetch successful Feb 13 20:02:58.142530 coreos-metadata[1848]: Feb 13 20:02:58.138 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Feb 13 20:02:58.146643 coreos-metadata[1848]: Feb 13 20:02:58.143 INFO Fetch successful Feb 13 20:02:58.146643 coreos-metadata[1848]: Feb 13 20:02:58.143 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Feb 13 20:02:58.149403 coreos-metadata[1848]: Feb 13 20:02:58.147 INFO Fetch successful Feb 13 20:02:58.149403 coreos-metadata[1848]: Feb 13 20:02:58.147 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Feb 13 20:02:58.151638 coreos-metadata[1848]: Feb 13 20:02:58.151 INFO Fetch successful Feb 13 20:02:58.151638 coreos-metadata[1848]: Feb 13 20:02:58.151 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Feb 13 20:02:58.152779 coreos-metadata[1848]: Feb 13 20:02:58.152 INFO Fetch failed with 404: resource not found Feb 13 20:02:58.152779 coreos-metadata[1848]: Feb 13 20:02:58.152 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Feb 13 20:02:58.154140 coreos-metadata[1848]: Feb 13 20:02:58.153 INFO Fetch successful Feb 13 20:02:58.154140 coreos-metadata[1848]: Feb 13 20:02:58.154 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Feb 13 20:02:58.154991 coreos-metadata[1848]: Feb 13 20:02:58.154 INFO Fetch successful Feb 13 20:02:58.154991 coreos-metadata[1848]: Feb 13 20:02:58.154 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Feb 13 20:02:58.162474 coreos-metadata[1848]: Feb 13 20:02:58.160 INFO Fetch successful Feb 13 20:02:58.162474 coreos-metadata[1848]: Feb 13 20:02:58.160 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Feb 13 20:02:58.171041 coreos-metadata[1848]: Feb 13 20:02:58.170 INFO Fetch successful Feb 13 20:02:58.171041 coreos-metadata[1848]: Feb 13 20:02:58.170 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Feb 13 20:02:58.185321 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Feb 13 20:02:58.187947 coreos-metadata[1848]: Feb 13 20:02:58.172 INFO Fetch successful Feb 13 20:02:58.206460 extend-filesystems[1897]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Feb 13 20:02:58.206460 extend-filesystems[1897]: old_desc_blocks = 1, new_desc_blocks = 1 Feb 13 20:02:58.206460 extend-filesystems[1897]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Feb 13 20:02:58.213106 extend-filesystems[1851]: Resized filesystem in /dev/nvme0n1p9 Feb 13 20:02:58.221471 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 20:02:58.256335 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 20:02:58.257735 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 20:02:58.270189 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 20:02:58.284964 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Feb 13 20:02:58.300860 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:02:58.311795 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 20:02:58.410315 bash[1931]: Updated "/home/core/.ssh/authorized_keys" Feb 13 20:02:58.415673 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 20:02:58.431452 systemd[1]: Starting sshkeys.service... Feb 13 20:02:58.440515 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Feb 13 20:02:58.443549 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 20:02:58.540862 dbus-daemon[1849]: [system] Successfully activated service 'org.freedesktop.hostname1' Feb 13 20:02:58.542581 dbus-daemon[1849]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=1896 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Feb 13 20:02:58.542793 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Feb 13 20:02:58.572098 systemd[1]: Starting polkit.service - Authorization Manager... Feb 13 20:02:58.617442 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (1726) Feb 13 20:02:58.589236 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Feb 13 20:02:58.615986 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Feb 13 20:02:58.664598 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 20:02:58.697732 polkitd[1952]: Started polkitd version 121 Feb 13 20:02:58.804712 polkitd[1952]: Loading rules from directory /etc/polkit-1/rules.d Feb 13 20:02:58.805536 polkitd[1952]: Loading rules from directory /usr/share/polkit-1/rules.d Feb 13 20:02:58.817718 amazon-ssm-agent[1930]: Initializing new seelog logger Feb 13 20:02:58.818083 amazon-ssm-agent[1930]: New Seelog Logger Creation Complete Feb 13 20:02:58.818083 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.818083 amazon-ssm-agent[1930]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.818418 polkitd[1952]: Finished loading, compiling and executing 2 rules Feb 13 20:02:58.826912 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 processing appconfig overrides Feb 13 20:02:58.828873 dbus-daemon[1849]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Feb 13 20:02:58.844383 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.844383 amazon-ssm-agent[1930]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.844383 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 processing appconfig overrides Feb 13 20:02:58.844383 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.844383 amazon-ssm-agent[1930]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.844383 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 processing appconfig overrides Feb 13 20:02:58.843138 systemd[1]: Started polkit.service - Authorization Manager. Feb 13 20:02:58.853667 polkitd[1952]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Feb 13 20:02:58.861385 amazon-ssm-agent[1930]: 2025-02-13 20:02:58 INFO Proxy environment variables: Feb 13 20:02:58.865948 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.865948 amazon-ssm-agent[1930]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Feb 13 20:02:58.866396 amazon-ssm-agent[1930]: 2025/02/13 20:02:58 processing appconfig overrides Feb 13 20:02:58.918207 systemd-hostnamed[1896]: Hostname set to (transient) Feb 13 20:02:58.919435 systemd-resolved[1672]: System hostname changed to 'ip-172-31-25-222'. Feb 13 20:02:58.969804 amazon-ssm-agent[1930]: 2025-02-13 20:02:58 INFO https_proxy: Feb 13 20:02:59.015023 coreos-metadata[1954]: Feb 13 20:02:59.014 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Feb 13 20:02:59.018693 coreos-metadata[1954]: Feb 13 20:02:59.018 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Feb 13 20:02:59.040617 coreos-metadata[1954]: Feb 13 20:02:59.040 INFO Fetch successful Feb 13 20:02:59.041028 coreos-metadata[1954]: Feb 13 20:02:59.040 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Feb 13 20:02:59.043924 sshd_keygen[1879]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 20:02:59.044809 coreos-metadata[1954]: Feb 13 20:02:59.044 INFO Fetch successful Feb 13 20:02:59.055991 unknown[1954]: wrote ssh authorized keys file for user: core Feb 13 20:02:59.063402 amazon-ssm-agent[1930]: 2025-02-13 20:02:58 INFO http_proxy: Feb 13 20:02:59.098981 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 20:02:59.109722 locksmithd[1900]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 20:02:59.118271 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 20:02:59.127504 systemd[1]: Started sshd@0-172.31.25.222:22-139.178.89.65:46194.service - OpenSSH per-connection server daemon (139.178.89.65:46194). Feb 13 20:02:59.164759 amazon-ssm-agent[1930]: 2025-02-13 20:02:58 INFO no_proxy: Feb 13 20:02:59.190454 update-ssh-keys[2057]: Updated "/home/core/.ssh/authorized_keys" Feb 13 20:02:59.193772 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Feb 13 20:02:59.206524 systemd[1]: Finished sshkeys.service. Feb 13 20:02:59.233146 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 20:02:59.233423 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 20:02:59.241897 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 20:02:59.269873 amazon-ssm-agent[1930]: 2025-02-13 20:02:58 INFO Checking if agent identity type OnPrem can be assumed Feb 13 20:02:59.270234 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 20:02:59.271706 containerd[1887]: time="2025-02-13T20:02:59.270471845Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Feb 13 20:02:59.279990 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 20:02:59.289887 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Feb 13 20:02:59.291798 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 20:02:59.365624 amazon-ssm-agent[1930]: 2025-02-13 20:02:58 INFO Checking if agent identity type EC2 can be assumed Feb 13 20:02:59.476873 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO Agent will take identity from EC2 Feb 13 20:02:59.520667 containerd[1887]: time="2025-02-13T20:02:59.519422201Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:02:59.527332 containerd[1887]: time="2025-02-13T20:02:59.527262384Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:02:59.527332 containerd[1887]: time="2025-02-13T20:02:59.527328631Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 20:02:59.527501 containerd[1887]: time="2025-02-13T20:02:59.527354395Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 20:02:59.527621 containerd[1887]: time="2025-02-13T20:02:59.527597445Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 20:02:59.527667 containerd[1887]: time="2025-02-13T20:02:59.527629297Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 20:02:59.528211 containerd[1887]: time="2025-02-13T20:02:59.527708653Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:02:59.528211 containerd[1887]: time="2025-02-13T20:02:59.527737449Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:02:59.528211 containerd[1887]: time="2025-02-13T20:02:59.528159387Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:02:59.528211 containerd[1887]: time="2025-02-13T20:02:59.528193193Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 20:02:59.528773 containerd[1887]: time="2025-02-13T20:02:59.528215115Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:02:59.528773 containerd[1887]: time="2025-02-13T20:02:59.528230214Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 20:02:59.528773 containerd[1887]: time="2025-02-13T20:02:59.528740335Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:02:59.529063 containerd[1887]: time="2025-02-13T20:02:59.529035951Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 20:02:59.530117 containerd[1887]: time="2025-02-13T20:02:59.529253620Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 20:02:59.530117 containerd[1887]: time="2025-02-13T20:02:59.529281012Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 20:02:59.530117 containerd[1887]: time="2025-02-13T20:02:59.529434932Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 20:02:59.530117 containerd[1887]: time="2025-02-13T20:02:59.529490348Z" level=info msg="metadata content store policy set" policy=shared Feb 13 20:02:59.538283 containerd[1887]: time="2025-02-13T20:02:59.538230575Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 20:02:59.538784 containerd[1887]: time="2025-02-13T20:02:59.538560972Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 20:02:59.538943 containerd[1887]: time="2025-02-13T20:02:59.538923749Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 20:02:59.539720 containerd[1887]: time="2025-02-13T20:02:59.539002305Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 20:02:59.539720 containerd[1887]: time="2025-02-13T20:02:59.539027397Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 20:02:59.540424 containerd[1887]: time="2025-02-13T20:02:59.540174782Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.540907345Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541076200Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541099211Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541182943Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541207608Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541227907Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541247548Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541270147Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541292934Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541313812Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.541374 containerd[1887]: time="2025-02-13T20:02:59.541334871Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.541353389Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544154726Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544185262Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544205721Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544227913Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544248115Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544268890Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544287854Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544324228Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544348500Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544401749Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544420856Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544440113Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544461358Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.544641 containerd[1887]: time="2025-02-13T20:02:59.544484517Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 20:02:59.545888 containerd[1887]: time="2025-02-13T20:02:59.544517470Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.545888 containerd[1887]: time="2025-02-13T20:02:59.544535961Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.545888 containerd[1887]: time="2025-02-13T20:02:59.544555154Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546189045Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546236110Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546255814Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546303004Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546320725Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546340691Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546355104Z" level=info msg="NRI interface is disabled by configuration." Feb 13 20:02:59.547379 containerd[1887]: time="2025-02-13T20:02:59.546380822Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 20:02:59.547725 containerd[1887]: time="2025-02-13T20:02:59.546812009Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 20:02:59.547725 containerd[1887]: time="2025-02-13T20:02:59.546898927Z" level=info msg="Connect containerd service" Feb 13 20:02:59.547725 containerd[1887]: time="2025-02-13T20:02:59.546959190Z" level=info msg="using legacy CRI server" Feb 13 20:02:59.547725 containerd[1887]: time="2025-02-13T20:02:59.546969144Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 20:02:59.547725 containerd[1887]: time="2025-02-13T20:02:59.547173505Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 20:02:59.551934 containerd[1887]: time="2025-02-13T20:02:59.550676114Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 20:02:59.551934 containerd[1887]: time="2025-02-13T20:02:59.550844388Z" level=info msg="Start subscribing containerd event" Feb 13 20:02:59.551934 containerd[1887]: time="2025-02-13T20:02:59.550913495Z" level=info msg="Start recovering state" Feb 13 20:02:59.551934 containerd[1887]: time="2025-02-13T20:02:59.550997296Z" level=info msg="Start event monitor" Feb 13 20:02:59.551934 containerd[1887]: time="2025-02-13T20:02:59.551020848Z" level=info msg="Start snapshots syncer" Feb 13 20:02:59.551934 containerd[1887]: time="2025-02-13T20:02:59.551033480Z" level=info msg="Start cni network conf syncer for default" Feb 13 20:02:59.551934 containerd[1887]: time="2025-02-13T20:02:59.551043996Z" level=info msg="Start streaming server" Feb 13 20:02:59.553621 containerd[1887]: time="2025-02-13T20:02:59.552725894Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 20:02:59.553621 containerd[1887]: time="2025-02-13T20:02:59.552789760Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 20:02:59.553621 containerd[1887]: time="2025-02-13T20:02:59.552852264Z" level=info msg="containerd successfully booted in 0.296175s" Feb 13 20:02:59.553514 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 20:02:59.577093 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 20:02:59.666112 sshd[2064]: Accepted publickey for core from 139.178.89.65 port 46194 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:02:59.669730 sshd[2064]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:02:59.676676 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 20:02:59.695715 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 20:02:59.705164 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 20:02:59.734046 systemd-logind[1863]: New session 1 of user core. Feb 13 20:02:59.775514 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [amazon-ssm-agent] using named pipe channel for IPC Feb 13 20:02:59.781182 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 20:02:59.797886 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 20:02:59.843995 (systemd)[2094]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 20:02:59.878466 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Feb 13 20:02:59.975181 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Feb 13 20:03:00.077535 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [amazon-ssm-agent] Starting Core Agent Feb 13 20:03:00.178439 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [amazon-ssm-agent] registrar detected. Attempting registration Feb 13 20:03:00.192293 systemd[2094]: Queued start job for default target default.target. Feb 13 20:03:00.198794 systemd[2094]: Created slice app.slice - User Application Slice. Feb 13 20:03:00.198835 systemd[2094]: Reached target paths.target - Paths. Feb 13 20:03:00.198855 systemd[2094]: Reached target timers.target - Timers. Feb 13 20:03:00.208515 systemd[2094]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 20:03:00.226864 systemd[2094]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 20:03:00.227037 systemd[2094]: Reached target sockets.target - Sockets. Feb 13 20:03:00.227060 systemd[2094]: Reached target basic.target - Basic System. Feb 13 20:03:00.227198 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 20:03:00.229987 systemd[2094]: Reached target default.target - Main User Target. Feb 13 20:03:00.230067 systemd[2094]: Startup finished in 358ms. Feb 13 20:03:00.232616 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 20:03:00.278855 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [Registrar] Starting registrar module Feb 13 20:03:00.379864 amazon-ssm-agent[1930]: 2025-02-13 20:02:59 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Feb 13 20:03:00.396816 systemd[1]: Started sshd@1-172.31.25.222:22-139.178.89.65:46204.service - OpenSSH per-connection server daemon (139.178.89.65:46204). Feb 13 20:03:00.543228 tar[1875]: linux-amd64/README.md Feb 13 20:03:00.547651 amazon-ssm-agent[1930]: 2025-02-13 20:03:00 INFO [EC2Identity] EC2 registration was successful. Feb 13 20:03:00.558996 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 20:03:00.581533 amazon-ssm-agent[1930]: 2025-02-13 20:03:00 INFO [CredentialRefresher] credentialRefresher has started Feb 13 20:03:00.581533 amazon-ssm-agent[1930]: 2025-02-13 20:03:00 INFO [CredentialRefresher] Starting credentials refresher loop Feb 13 20:03:00.581533 amazon-ssm-agent[1930]: 2025-02-13 20:03:00 INFO EC2RoleProvider Successfully connected with instance profile role credentials Feb 13 20:03:00.611228 sshd[2105]: Accepted publickey for core from 139.178.89.65 port 46204 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:00.615007 sshd[2105]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:00.623437 systemd-logind[1863]: New session 2 of user core. Feb 13 20:03:00.634592 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 20:03:00.648978 amazon-ssm-agent[1930]: 2025-02-13 20:03:00 INFO [CredentialRefresher] Next credential rotation will be in 30.258325405683333 minutes Feb 13 20:03:00.769104 sshd[2105]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:00.774367 systemd[1]: sshd@1-172.31.25.222:22-139.178.89.65:46204.service: Deactivated successfully. Feb 13 20:03:00.778050 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 20:03:00.781724 systemd-logind[1863]: Session 2 logged out. Waiting for processes to exit. Feb 13 20:03:00.783858 systemd-logind[1863]: Removed session 2. Feb 13 20:03:00.786730 ntpd[1853]: Listen normally on 6 eth0 [fe80::415:e6ff:feab:4407%2]:123 Feb 13 20:03:00.787152 ntpd[1853]: 13 Feb 20:03:00 ntpd[1853]: Listen normally on 6 eth0 [fe80::415:e6ff:feab:4407%2]:123 Feb 13 20:03:00.812404 systemd[1]: Started sshd@2-172.31.25.222:22-139.178.89.65:46214.service - OpenSSH per-connection server daemon (139.178.89.65:46214). Feb 13 20:03:01.093469 sshd[2115]: Accepted publickey for core from 139.178.89.65 port 46214 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:01.097111 sshd[2115]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:01.117185 systemd-logind[1863]: New session 3 of user core. Feb 13 20:03:01.121790 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 20:03:01.260171 sshd[2115]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:01.273885 systemd[1]: sshd@2-172.31.25.222:22-139.178.89.65:46214.service: Deactivated successfully. Feb 13 20:03:01.281101 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 20:03:01.297443 systemd-logind[1863]: Session 3 logged out. Waiting for processes to exit. Feb 13 20:03:01.310687 systemd-logind[1863]: Removed session 3. Feb 13 20:03:01.624279 amazon-ssm-agent[1930]: 2025-02-13 20:03:01 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Feb 13 20:03:01.728811 amazon-ssm-agent[1930]: 2025-02-13 20:03:01 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2122) started Feb 13 20:03:01.829498 amazon-ssm-agent[1930]: 2025-02-13 20:03:01 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Feb 13 20:03:02.492230 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:02.495340 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 20:03:02.498200 systemd[1]: Startup finished in 942ms (kernel) + 8.531s (initrd) + 10.195s (userspace) = 19.669s. Feb 13 20:03:02.512047 (kubelet)[2138]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:03:04.025820 kubelet[2138]: E0213 20:03:04.025762 2138 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:03:04.029630 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:03:04.029852 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:03:04.030959 systemd[1]: kubelet.service: Consumed 1.095s CPU time. Feb 13 20:03:05.379534 systemd-resolved[1672]: Clock change detected. Flushing caches. Feb 13 20:03:11.881827 systemd[1]: Started sshd@3-172.31.25.222:22-139.178.89.65:35978.service - OpenSSH per-connection server daemon (139.178.89.65:35978). Feb 13 20:03:12.054466 sshd[2150]: Accepted publickey for core from 139.178.89.65 port 35978 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:12.056574 sshd[2150]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:12.071983 systemd-logind[1863]: New session 4 of user core. Feb 13 20:03:12.082058 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 20:03:12.217255 sshd[2150]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:12.224406 systemd[1]: sshd@3-172.31.25.222:22-139.178.89.65:35978.service: Deactivated successfully. Feb 13 20:03:12.226972 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 20:03:12.227743 systemd-logind[1863]: Session 4 logged out. Waiting for processes to exit. Feb 13 20:03:12.228996 systemd-logind[1863]: Removed session 4. Feb 13 20:03:12.256003 systemd[1]: Started sshd@4-172.31.25.222:22-139.178.89.65:35988.service - OpenSSH per-connection server daemon (139.178.89.65:35988). Feb 13 20:03:12.416638 sshd[2157]: Accepted publickey for core from 139.178.89.65 port 35988 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:12.418981 sshd[2157]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:12.426996 systemd-logind[1863]: New session 5 of user core. Feb 13 20:03:12.438176 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 20:03:12.559564 sshd[2157]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:12.576660 systemd[1]: sshd@4-172.31.25.222:22-139.178.89.65:35988.service: Deactivated successfully. Feb 13 20:03:12.579770 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 20:03:12.580810 systemd-logind[1863]: Session 5 logged out. Waiting for processes to exit. Feb 13 20:03:12.600786 systemd[1]: Started sshd@5-172.31.25.222:22-139.178.89.65:35998.service - OpenSSH per-connection server daemon (139.178.89.65:35998). Feb 13 20:03:12.602759 systemd-logind[1863]: Removed session 5. Feb 13 20:03:12.771953 sshd[2164]: Accepted publickey for core from 139.178.89.65 port 35998 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:12.773428 sshd[2164]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:12.783176 systemd-logind[1863]: New session 6 of user core. Feb 13 20:03:12.790354 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 20:03:12.923466 sshd[2164]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:12.928203 systemd[1]: sshd@5-172.31.25.222:22-139.178.89.65:35998.service: Deactivated successfully. Feb 13 20:03:12.930213 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 20:03:12.931089 systemd-logind[1863]: Session 6 logged out. Waiting for processes to exit. Feb 13 20:03:12.932243 systemd-logind[1863]: Removed session 6. Feb 13 20:03:12.961586 systemd[1]: Started sshd@6-172.31.25.222:22-139.178.89.65:36000.service - OpenSSH per-connection server daemon (139.178.89.65:36000). Feb 13 20:03:13.124432 sshd[2171]: Accepted publickey for core from 139.178.89.65 port 36000 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:13.126215 sshd[2171]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:13.132738 systemd-logind[1863]: New session 7 of user core. Feb 13 20:03:13.142378 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 20:03:13.260040 sudo[2174]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Feb 13 20:03:13.260468 sudo[2174]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:03:13.279867 sudo[2174]: pam_unix(sudo:session): session closed for user root Feb 13 20:03:13.303790 sshd[2171]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:13.310232 systemd-logind[1863]: Session 7 logged out. Waiting for processes to exit. Feb 13 20:03:13.311288 systemd[1]: sshd@6-172.31.25.222:22-139.178.89.65:36000.service: Deactivated successfully. Feb 13 20:03:13.313336 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 20:03:13.314478 systemd-logind[1863]: Removed session 7. Feb 13 20:03:13.343575 systemd[1]: Started sshd@7-172.31.25.222:22-139.178.89.65:36006.service - OpenSSH per-connection server daemon (139.178.89.65:36006). Feb 13 20:03:13.524291 sshd[2179]: Accepted publickey for core from 139.178.89.65 port 36006 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:13.525896 sshd[2179]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:13.544515 systemd-logind[1863]: New session 8 of user core. Feb 13 20:03:13.551373 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 20:03:13.658049 sudo[2183]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Feb 13 20:03:13.658490 sudo[2183]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:03:13.665490 sudo[2183]: pam_unix(sudo:session): session closed for user root Feb 13 20:03:13.676813 sudo[2182]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Feb 13 20:03:13.677234 sudo[2182]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:03:13.694501 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Feb 13 20:03:13.698262 auditctl[2186]: No rules Feb 13 20:03:13.698977 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 20:03:13.699242 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Feb 13 20:03:13.707618 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Feb 13 20:03:13.733409 augenrules[2204]: No rules Feb 13 20:03:13.734849 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Feb 13 20:03:13.736309 sudo[2182]: pam_unix(sudo:session): session closed for user root Feb 13 20:03:13.759529 sshd[2179]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:13.762900 systemd[1]: sshd@7-172.31.25.222:22-139.178.89.65:36006.service: Deactivated successfully. Feb 13 20:03:13.764994 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 20:03:13.766683 systemd-logind[1863]: Session 8 logged out. Waiting for processes to exit. Feb 13 20:03:13.767853 systemd-logind[1863]: Removed session 8. Feb 13 20:03:13.809588 systemd[1]: Started sshd@8-172.31.25.222:22-139.178.89.65:36010.service - OpenSSH per-connection server daemon (139.178.89.65:36010). Feb 13 20:03:13.966981 sshd[2212]: Accepted publickey for core from 139.178.89.65 port 36010 ssh2: RSA SHA256:7nv7xaFFWmIAvPewvKjLuTxkMrDcPy3WtQ5BDo3Wg0I Feb 13 20:03:13.968527 sshd[2212]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 20:03:13.973773 systemd-logind[1863]: New session 9 of user core. Feb 13 20:03:13.980362 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 20:03:14.080357 sudo[2215]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 20:03:14.080754 sudo[2215]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 20:03:14.721650 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 20:03:14.727490 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 20:03:14.729356 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:03:14.731189 (dockerd)[2233]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 20:03:15.105348 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:15.110800 (kubelet)[2241]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:03:15.192188 kubelet[2241]: E0213 20:03:15.192063 2241 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:03:15.206423 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:03:15.206622 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:03:15.598406 dockerd[2233]: time="2025-02-13T20:03:15.598250221Z" level=info msg="Starting up" Feb 13 20:03:15.814347 dockerd[2233]: time="2025-02-13T20:03:15.814091130Z" level=info msg="Loading containers: start." Feb 13 20:03:15.965140 kernel: Initializing XFRM netlink socket Feb 13 20:03:16.002162 (udev-worker)[2269]: Network interface NamePolicy= disabled on kernel command line. Feb 13 20:03:16.077162 systemd-networkd[1718]: docker0: Link UP Feb 13 20:03:16.125311 dockerd[2233]: time="2025-02-13T20:03:16.125260314Z" level=info msg="Loading containers: done." Feb 13 20:03:16.157049 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck687671859-merged.mount: Deactivated successfully. Feb 13 20:03:16.165731 dockerd[2233]: time="2025-02-13T20:03:16.165672560Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 20:03:16.165928 dockerd[2233]: time="2025-02-13T20:03:16.165809959Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Feb 13 20:03:16.165979 dockerd[2233]: time="2025-02-13T20:03:16.165950127Z" level=info msg="Daemon has completed initialization" Feb 13 20:03:16.256370 dockerd[2233]: time="2025-02-13T20:03:16.255633296Z" level=info msg="API listen on /run/docker.sock" Feb 13 20:03:16.255871 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 20:03:17.770787 containerd[1887]: time="2025-02-13T20:03:17.770563829Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\"" Feb 13 20:03:18.642985 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2640204168.mount: Deactivated successfully. Feb 13 20:03:21.273158 containerd[1887]: time="2025-02-13T20:03:21.273064709Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:21.274697 containerd[1887]: time="2025-02-13T20:03:21.274653051Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.2: active requests=0, bytes read=28673931" Feb 13 20:03:21.276062 containerd[1887]: time="2025-02-13T20:03:21.275556421Z" level=info msg="ImageCreate event name:\"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:21.279698 containerd[1887]: time="2025-02-13T20:03:21.279650684Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:21.281093 containerd[1887]: time="2025-02-13T20:03:21.281049542Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.2\" with image id \"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\", size \"28670731\" in 3.510253755s" Feb 13 20:03:21.281333 containerd[1887]: time="2025-02-13T20:03:21.281309848Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\" returns image reference \"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\"" Feb 13 20:03:21.282178 containerd[1887]: time="2025-02-13T20:03:21.282145142Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\"" Feb 13 20:03:23.648738 containerd[1887]: time="2025-02-13T20:03:23.648686972Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:23.650064 containerd[1887]: time="2025-02-13T20:03:23.650014933Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.2: active requests=0, bytes read=24771784" Feb 13 20:03:23.652141 containerd[1887]: time="2025-02-13T20:03:23.652021146Z" level=info msg="ImageCreate event name:\"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:23.655450 containerd[1887]: time="2025-02-13T20:03:23.655392440Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:23.656700 containerd[1887]: time="2025-02-13T20:03:23.656662881Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.2\" with image id \"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\", size \"26259392\" in 2.374481602s" Feb 13 20:03:23.656928 containerd[1887]: time="2025-02-13T20:03:23.656820867Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\" returns image reference \"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\"" Feb 13 20:03:23.657972 containerd[1887]: time="2025-02-13T20:03:23.657946400Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\"" Feb 13 20:03:25.239475 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 13 20:03:25.251514 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:03:25.547472 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:25.556052 (kubelet)[2456]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:03:25.685205 kubelet[2456]: E0213 20:03:25.684859 2456 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:03:25.691237 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:03:25.691476 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:03:25.974711 containerd[1887]: time="2025-02-13T20:03:25.974149346Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:25.975890 containerd[1887]: time="2025-02-13T20:03:25.975711952Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.2: active requests=0, bytes read=19170276" Feb 13 20:03:25.977325 containerd[1887]: time="2025-02-13T20:03:25.976930192Z" level=info msg="ImageCreate event name:\"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:25.979687 containerd[1887]: time="2025-02-13T20:03:25.979650896Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:25.981192 containerd[1887]: time="2025-02-13T20:03:25.981150368Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.2\" with image id \"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\", size \"20657902\" in 2.323170909s" Feb 13 20:03:25.981328 containerd[1887]: time="2025-02-13T20:03:25.981307094Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\" returns image reference \"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\"" Feb 13 20:03:25.982093 containerd[1887]: time="2025-02-13T20:03:25.982071442Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\"" Feb 13 20:03:27.909662 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2399460509.mount: Deactivated successfully. Feb 13 20:03:28.777597 containerd[1887]: time="2025-02-13T20:03:28.777545219Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:28.778910 containerd[1887]: time="2025-02-13T20:03:28.778767270Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.2: active requests=0, bytes read=30908839" Feb 13 20:03:28.781068 containerd[1887]: time="2025-02-13T20:03:28.779866357Z" level=info msg="ImageCreate event name:\"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:28.783129 containerd[1887]: time="2025-02-13T20:03:28.782035359Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:28.783129 containerd[1887]: time="2025-02-13T20:03:28.782923713Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.2\" with image id \"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\", repo tag \"registry.k8s.io/kube-proxy:v1.32.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\", size \"30907858\" in 2.800511089s" Feb 13 20:03:28.783129 containerd[1887]: time="2025-02-13T20:03:28.782964220Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\" returns image reference \"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\"" Feb 13 20:03:28.784006 containerd[1887]: time="2025-02-13T20:03:28.783980622Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Feb 13 20:03:29.349704 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3580137727.mount: Deactivated successfully. Feb 13 20:03:29.546574 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Feb 13 20:03:30.784237 containerd[1887]: time="2025-02-13T20:03:30.784149572Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:30.787287 containerd[1887]: time="2025-02-13T20:03:30.787192720Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Feb 13 20:03:30.788807 containerd[1887]: time="2025-02-13T20:03:30.788767974Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:30.793064 containerd[1887]: time="2025-02-13T20:03:30.792999831Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:30.794626 containerd[1887]: time="2025-02-13T20:03:30.794464890Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 2.010450504s" Feb 13 20:03:30.794626 containerd[1887]: time="2025-02-13T20:03:30.794511748Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Feb 13 20:03:30.795208 containerd[1887]: time="2025-02-13T20:03:30.795178907Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Feb 13 20:03:31.369296 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2953771897.mount: Deactivated successfully. Feb 13 20:03:31.387521 containerd[1887]: time="2025-02-13T20:03:31.387464259Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:31.390534 containerd[1887]: time="2025-02-13T20:03:31.390470715Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Feb 13 20:03:31.391603 containerd[1887]: time="2025-02-13T20:03:31.391540547Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:31.404287 containerd[1887]: time="2025-02-13T20:03:31.401364986Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:31.404287 containerd[1887]: time="2025-02-13T20:03:31.404097502Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 608.882789ms" Feb 13 20:03:31.404287 containerd[1887]: time="2025-02-13T20:03:31.404159785Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Feb 13 20:03:31.405961 containerd[1887]: time="2025-02-13T20:03:31.405924156Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Feb 13 20:03:32.388052 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3868690159.mount: Deactivated successfully. Feb 13 20:03:35.739610 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Feb 13 20:03:35.746435 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:03:35.927568 containerd[1887]: time="2025-02-13T20:03:35.927502265Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:35.929766 containerd[1887]: time="2025-02-13T20:03:35.929708829Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551320" Feb 13 20:03:35.937127 containerd[1887]: time="2025-02-13T20:03:35.935313847Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:35.946268 containerd[1887]: time="2025-02-13T20:03:35.946219149Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:35.949587 containerd[1887]: time="2025-02-13T20:03:35.949537960Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 4.54357226s" Feb 13 20:03:35.949697 containerd[1887]: time="2025-02-13T20:03:35.949592074Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Feb 13 20:03:36.173576 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:36.185551 (kubelet)[2598]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 20:03:36.256905 kubelet[2598]: E0213 20:03:36.256849 2598 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 20:03:36.259580 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 20:03:36.259786 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 20:03:38.805941 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:38.813487 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:03:38.864326 systemd[1]: Reloading requested from client PID 2627 ('systemctl') (unit session-9.scope)... Feb 13 20:03:38.864350 systemd[1]: Reloading... Feb 13 20:03:39.023178 zram_generator::config[2668]: No configuration found. Feb 13 20:03:39.194271 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:03:39.329172 systemd[1]: Reloading finished in 464 ms. Feb 13 20:03:39.397652 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Feb 13 20:03:39.397782 systemd[1]: kubelet.service: Failed with result 'signal'. Feb 13 20:03:39.398168 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:39.406535 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:03:39.670835 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:39.674589 (kubelet)[2728]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 20:03:39.772301 kubelet[2728]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:03:39.773155 kubelet[2728]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 20:03:39.773155 kubelet[2728]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:03:39.773155 kubelet[2728]: I0213 20:03:39.772771 2728 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 20:03:40.091534 kubelet[2728]: I0213 20:03:40.091483 2728 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 20:03:40.091534 kubelet[2728]: I0213 20:03:40.091516 2728 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 20:03:40.091885 kubelet[2728]: I0213 20:03:40.091861 2728 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 20:03:40.170540 kubelet[2728]: I0213 20:03:40.170435 2728 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 20:03:40.172344 kubelet[2728]: E0213 20:03:40.172298 2728 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.25.222:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:40.217096 kubelet[2728]: E0213 20:03:40.217043 2728 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 20:03:40.217096 kubelet[2728]: I0213 20:03:40.217087 2728 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 20:03:40.230805 kubelet[2728]: I0213 20:03:40.230528 2728 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 20:03:40.239677 kubelet[2728]: I0213 20:03:40.239535 2728 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 20:03:40.239994 kubelet[2728]: I0213 20:03:40.239670 2728 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-25-222","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 20:03:40.244095 kubelet[2728]: I0213 20:03:40.243278 2728 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 20:03:40.244095 kubelet[2728]: I0213 20:03:40.244098 2728 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 20:03:40.251851 kubelet[2728]: I0213 20:03:40.251813 2728 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:03:40.268100 kubelet[2728]: I0213 20:03:40.268053 2728 kubelet.go:446] "Attempting to sync node with API server" Feb 13 20:03:40.268100 kubelet[2728]: I0213 20:03:40.268098 2728 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 20:03:40.268305 kubelet[2728]: I0213 20:03:40.268151 2728 kubelet.go:352] "Adding apiserver pod source" Feb 13 20:03:40.268305 kubelet[2728]: I0213 20:03:40.268167 2728 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 20:03:40.282520 kubelet[2728]: W0213 20:03:40.282256 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.25.222:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:40.282520 kubelet[2728]: E0213 20:03:40.282342 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.25.222:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:40.282520 kubelet[2728]: W0213 20:03:40.282439 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.25.222:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-222&limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:40.282520 kubelet[2728]: E0213 20:03:40.282478 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.25.222:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-222&limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:40.283563 kubelet[2728]: I0213 20:03:40.283384 2728 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Feb 13 20:03:40.288309 kubelet[2728]: I0213 20:03:40.288273 2728 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 20:03:40.289065 kubelet[2728]: W0213 20:03:40.288518 2728 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 20:03:40.292061 kubelet[2728]: I0213 20:03:40.291838 2728 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 20:03:40.292061 kubelet[2728]: I0213 20:03:40.291878 2728 server.go:1287] "Started kubelet" Feb 13 20:03:40.292497 kubelet[2728]: I0213 20:03:40.292460 2728 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 20:03:40.301700 kubelet[2728]: I0213 20:03:40.297373 2728 server.go:490] "Adding debug handlers to kubelet server" Feb 13 20:03:40.301700 kubelet[2728]: I0213 20:03:40.298170 2728 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 20:03:40.301700 kubelet[2728]: I0213 20:03:40.299323 2728 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 20:03:40.302887 kubelet[2728]: E0213 20:03:40.298569 2728 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.25.222:6443/api/v1/namespaces/default/events\": dial tcp 172.31.25.222:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-25-222.1823dd2241737610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-25-222,UID:ip-172-31-25-222,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-25-222,},FirstTimestamp:2025-02-13 20:03:40.291855888 +0000 UTC m=+0.611889045,LastTimestamp:2025-02-13 20:03:40.291855888 +0000 UTC m=+0.611889045,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-25-222,}" Feb 13 20:03:40.307438 kubelet[2728]: I0213 20:03:40.307410 2728 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 20:03:40.309480 kubelet[2728]: I0213 20:03:40.307558 2728 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 20:03:40.309694 kubelet[2728]: I0213 20:03:40.309680 2728 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 20:03:40.310217 kubelet[2728]: E0213 20:03:40.310184 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:40.322139 kubelet[2728]: I0213 20:03:40.321543 2728 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 20:03:40.326419 kubelet[2728]: I0213 20:03:40.326383 2728 reconciler.go:26] "Reconciler: start to sync state" Feb 13 20:03:40.326999 kubelet[2728]: E0213 20:03:40.326955 2728 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.222:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-222?timeout=10s\": dial tcp 172.31.25.222:6443: connect: connection refused" interval="200ms" Feb 13 20:03:40.327435 kubelet[2728]: W0213 20:03:40.327089 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.25.222:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:40.327435 kubelet[2728]: E0213 20:03:40.327189 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.25.222:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:40.333637 kubelet[2728]: I0213 20:03:40.333599 2728 factory.go:221] Registration of the containerd container factory successfully Feb 13 20:03:40.333637 kubelet[2728]: I0213 20:03:40.333623 2728 factory.go:221] Registration of the systemd container factory successfully Feb 13 20:03:40.333812 kubelet[2728]: I0213 20:03:40.333744 2728 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 20:03:40.358458 kubelet[2728]: E0213 20:03:40.358276 2728 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 20:03:40.364681 kubelet[2728]: I0213 20:03:40.364456 2728 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 20:03:40.366660 kubelet[2728]: I0213 20:03:40.366588 2728 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 20:03:40.366660 kubelet[2728]: I0213 20:03:40.366607 2728 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 20:03:40.366660 kubelet[2728]: I0213 20:03:40.366637 2728 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:03:40.367564 kubelet[2728]: I0213 20:03:40.367446 2728 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 20:03:40.367564 kubelet[2728]: I0213 20:03:40.367544 2728 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 20:03:40.367870 kubelet[2728]: I0213 20:03:40.367744 2728 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 20:03:40.367870 kubelet[2728]: I0213 20:03:40.367768 2728 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 20:03:40.368472 kubelet[2728]: E0213 20:03:40.368409 2728 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 20:03:40.370163 kubelet[2728]: W0213 20:03:40.369647 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.25.222:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:40.370394 kubelet[2728]: E0213 20:03:40.370274 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.25.222:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:40.371145 kubelet[2728]: I0213 20:03:40.370665 2728 policy_none.go:49] "None policy: Start" Feb 13 20:03:40.371145 kubelet[2728]: I0213 20:03:40.370731 2728 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 20:03:40.371145 kubelet[2728]: I0213 20:03:40.370749 2728 state_mem.go:35] "Initializing new in-memory state store" Feb 13 20:03:40.379749 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 20:03:40.390232 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 20:03:40.394440 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 20:03:40.412584 kubelet[2728]: E0213 20:03:40.412544 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:40.414302 kubelet[2728]: I0213 20:03:40.413592 2728 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 20:03:40.414302 kubelet[2728]: I0213 20:03:40.413909 2728 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 20:03:40.414302 kubelet[2728]: I0213 20:03:40.413923 2728 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 20:03:40.414302 kubelet[2728]: I0213 20:03:40.414177 2728 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 20:03:40.425612 kubelet[2728]: E0213 20:03:40.421978 2728 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 20:03:40.425612 kubelet[2728]: E0213 20:03:40.422053 2728 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-25-222\" not found" Feb 13 20:03:40.492038 systemd[1]: Created slice kubepods-burstable-podfaf6713a9dad2a13975e98f5c591f88e.slice - libcontainer container kubepods-burstable-podfaf6713a9dad2a13975e98f5c591f88e.slice. Feb 13 20:03:40.506495 kubelet[2728]: E0213 20:03:40.506254 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:40.514769 systemd[1]: Created slice kubepods-burstable-pod73c480cf63f0b0ed72d37ea8cbc1df29.slice - libcontainer container kubepods-burstable-pod73c480cf63f0b0ed72d37ea8cbc1df29.slice. Feb 13 20:03:40.518701 kubelet[2728]: I0213 20:03:40.518670 2728 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-222" Feb 13 20:03:40.520324 kubelet[2728]: E0213 20:03:40.520254 2728 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.222:6443/api/v1/nodes\": dial tcp 172.31.25.222:6443: connect: connection refused" node="ip-172-31-25-222" Feb 13 20:03:40.521539 kubelet[2728]: E0213 20:03:40.521353 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:40.526964 systemd[1]: Created slice kubepods-burstable-podb204870b3e481c92b18dac4ca6ee6c9f.slice - libcontainer container kubepods-burstable-podb204870b3e481c92b18dac4ca6ee6c9f.slice. Feb 13 20:03:40.528029 kubelet[2728]: I0213 20:03:40.527686 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b204870b3e481c92b18dac4ca6ee6c9f-kubeconfig\") pod \"kube-scheduler-ip-172-31-25-222\" (UID: \"b204870b3e481c92b18dac4ca6ee6c9f\") " pod="kube-system/kube-scheduler-ip-172-31-25-222" Feb 13 20:03:40.528029 kubelet[2728]: I0213 20:03:40.527723 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/faf6713a9dad2a13975e98f5c591f88e-ca-certs\") pod \"kube-apiserver-ip-172-31-25-222\" (UID: \"faf6713a9dad2a13975e98f5c591f88e\") " pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:40.528029 kubelet[2728]: I0213 20:03:40.527751 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/faf6713a9dad2a13975e98f5c591f88e-k8s-certs\") pod \"kube-apiserver-ip-172-31-25-222\" (UID: \"faf6713a9dad2a13975e98f5c591f88e\") " pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:40.528029 kubelet[2728]: I0213 20:03:40.527781 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-k8s-certs\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:40.528029 kubelet[2728]: I0213 20:03:40.527805 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-kubeconfig\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:40.528301 kubelet[2728]: I0213 20:03:40.527829 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/faf6713a9dad2a13975e98f5c591f88e-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-25-222\" (UID: \"faf6713a9dad2a13975e98f5c591f88e\") " pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:40.528301 kubelet[2728]: I0213 20:03:40.527856 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-ca-certs\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:40.528301 kubelet[2728]: I0213 20:03:40.527883 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:40.528301 kubelet[2728]: I0213 20:03:40.527912 2728 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:40.529610 kubelet[2728]: E0213 20:03:40.529575 2728 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.222:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-222?timeout=10s\": dial tcp 172.31.25.222:6443: connect: connection refused" interval="400ms" Feb 13 20:03:40.531600 kubelet[2728]: E0213 20:03:40.531575 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:40.724671 kubelet[2728]: I0213 20:03:40.724308 2728 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-222" Feb 13 20:03:40.724671 kubelet[2728]: E0213 20:03:40.724633 2728 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.222:6443/api/v1/nodes\": dial tcp 172.31.25.222:6443: connect: connection refused" node="ip-172-31-25-222" Feb 13 20:03:40.810614 containerd[1887]: time="2025-02-13T20:03:40.810546212Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-25-222,Uid:faf6713a9dad2a13975e98f5c591f88e,Namespace:kube-system,Attempt:0,}" Feb 13 20:03:40.829604 containerd[1887]: time="2025-02-13T20:03:40.829554923Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-25-222,Uid:73c480cf63f0b0ed72d37ea8cbc1df29,Namespace:kube-system,Attempt:0,}" Feb 13 20:03:40.832766 containerd[1887]: time="2025-02-13T20:03:40.832718825Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-25-222,Uid:b204870b3e481c92b18dac4ca6ee6c9f,Namespace:kube-system,Attempt:0,}" Feb 13 20:03:40.930820 kubelet[2728]: E0213 20:03:40.930738 2728 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.222:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-222?timeout=10s\": dial tcp 172.31.25.222:6443: connect: connection refused" interval="800ms" Feb 13 20:03:41.107973 kubelet[2728]: W0213 20:03:41.105705 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.25.222:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-222&limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:41.107973 kubelet[2728]: E0213 20:03:41.105789 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.25.222:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-222&limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:41.126815 kubelet[2728]: I0213 20:03:41.126749 2728 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-222" Feb 13 20:03:41.127225 kubelet[2728]: E0213 20:03:41.127126 2728 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.222:6443/api/v1/nodes\": dial tcp 172.31.25.222:6443: connect: connection refused" node="ip-172-31-25-222" Feb 13 20:03:41.223744 kubelet[2728]: W0213 20:03:41.223641 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.25.222:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:41.223744 kubelet[2728]: E0213 20:03:41.223714 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.25.222:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:41.407478 kubelet[2728]: W0213 20:03:41.407261 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.25.222:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:41.407478 kubelet[2728]: E0213 20:03:41.407354 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.25.222:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:41.619218 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount233045456.mount: Deactivated successfully. Feb 13 20:03:41.643167 containerd[1887]: time="2025-02-13T20:03:41.643090867Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:03:41.648656 containerd[1887]: time="2025-02-13T20:03:41.646218098Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Feb 13 20:03:41.648787 kubelet[2728]: W0213 20:03:41.648570 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.25.222:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:41.648787 kubelet[2728]: E0213 20:03:41.648620 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.25.222:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:41.672978 containerd[1887]: time="2025-02-13T20:03:41.671812782Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 20:03:41.672978 containerd[1887]: time="2025-02-13T20:03:41.672097777Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 20:03:41.672978 containerd[1887]: time="2025-02-13T20:03:41.672515830Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:03:41.681547 containerd[1887]: time="2025-02-13T20:03:41.681496072Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:03:41.684599 containerd[1887]: time="2025-02-13T20:03:41.684080891Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 873.4299ms" Feb 13 20:03:41.686258 containerd[1887]: time="2025-02-13T20:03:41.686211931Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:03:41.690152 containerd[1887]: time="2025-02-13T20:03:41.689220972Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 859.568103ms" Feb 13 20:03:41.693049 containerd[1887]: time="2025-02-13T20:03:41.692998901Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 860.197453ms" Feb 13 20:03:41.694646 containerd[1887]: time="2025-02-13T20:03:41.694268588Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 20:03:41.740884 kubelet[2728]: E0213 20:03:41.740587 2728 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.222:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-222?timeout=10s\": dial tcp 172.31.25.222:6443: connect: connection refused" interval="1.6s" Feb 13 20:03:41.941577 kubelet[2728]: I0213 20:03:41.937717 2728 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-222" Feb 13 20:03:41.941577 kubelet[2728]: E0213 20:03:41.939087 2728 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.25.222:6443/api/v1/nodes\": dial tcp 172.31.25.222:6443: connect: connection refused" node="ip-172-31-25-222" Feb 13 20:03:41.998735 containerd[1887]: time="2025-02-13T20:03:41.997441759Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:03:41.998735 containerd[1887]: time="2025-02-13T20:03:41.997508294Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:03:41.998735 containerd[1887]: time="2025-02-13T20:03:41.997544266Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:41.998735 containerd[1887]: time="2025-02-13T20:03:41.997714303Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:42.046172 containerd[1887]: time="2025-02-13T20:03:42.041479168Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:03:42.046172 containerd[1887]: time="2025-02-13T20:03:42.041844653Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:03:42.046172 containerd[1887]: time="2025-02-13T20:03:42.041888863Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:42.048250 containerd[1887]: time="2025-02-13T20:03:42.043249894Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:42.061021 containerd[1887]: time="2025-02-13T20:03:42.060311210Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:03:42.061021 containerd[1887]: time="2025-02-13T20:03:42.060411245Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:03:42.061021 containerd[1887]: time="2025-02-13T20:03:42.060430371Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:42.061021 containerd[1887]: time="2025-02-13T20:03:42.060558660Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:42.081524 systemd[1]: Started cri-containerd-0183c4112df27dd005e00044356f7fe4536208e12aa7db7cdd0fb219001b88fc.scope - libcontainer container 0183c4112df27dd005e00044356f7fe4536208e12aa7db7cdd0fb219001b88fc. Feb 13 20:03:42.114423 systemd[1]: Started cri-containerd-971805c833cee288ab051dabb6e380c2a11c47aee76e29269adad77195285450.scope - libcontainer container 971805c833cee288ab051dabb6e380c2a11c47aee76e29269adad77195285450. Feb 13 20:03:42.121456 systemd[1]: Started cri-containerd-152eb051aef4bfc717576523d14f15a1db84c82c23ea09a6d2b544b2192b6d00.scope - libcontainer container 152eb051aef4bfc717576523d14f15a1db84c82c23ea09a6d2b544b2192b6d00. Feb 13 20:03:42.212051 containerd[1887]: time="2025-02-13T20:03:42.211282167Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-25-222,Uid:73c480cf63f0b0ed72d37ea8cbc1df29,Namespace:kube-system,Attempt:0,} returns sandbox id \"0183c4112df27dd005e00044356f7fe4536208e12aa7db7cdd0fb219001b88fc\"" Feb 13 20:03:42.229833 containerd[1887]: time="2025-02-13T20:03:42.229791886Z" level=info msg="CreateContainer within sandbox \"0183c4112df27dd005e00044356f7fe4536208e12aa7db7cdd0fb219001b88fc\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 20:03:42.231133 containerd[1887]: time="2025-02-13T20:03:42.230851592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-25-222,Uid:faf6713a9dad2a13975e98f5c591f88e,Namespace:kube-system,Attempt:0,} returns sandbox id \"152eb051aef4bfc717576523d14f15a1db84c82c23ea09a6d2b544b2192b6d00\"" Feb 13 20:03:42.235384 containerd[1887]: time="2025-02-13T20:03:42.235347845Z" level=info msg="CreateContainer within sandbox \"152eb051aef4bfc717576523d14f15a1db84c82c23ea09a6d2b544b2192b6d00\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 20:03:42.262852 containerd[1887]: time="2025-02-13T20:03:42.262805210Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-25-222,Uid:b204870b3e481c92b18dac4ca6ee6c9f,Namespace:kube-system,Attempt:0,} returns sandbox id \"971805c833cee288ab051dabb6e380c2a11c47aee76e29269adad77195285450\"" Feb 13 20:03:42.265902 containerd[1887]: time="2025-02-13T20:03:42.265856180Z" level=info msg="CreateContainer within sandbox \"971805c833cee288ab051dabb6e380c2a11c47aee76e29269adad77195285450\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 20:03:42.274774 containerd[1887]: time="2025-02-13T20:03:42.274723134Z" level=info msg="CreateContainer within sandbox \"0183c4112df27dd005e00044356f7fe4536208e12aa7db7cdd0fb219001b88fc\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c\"" Feb 13 20:03:42.275522 containerd[1887]: time="2025-02-13T20:03:42.275484075Z" level=info msg="StartContainer for \"f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c\"" Feb 13 20:03:42.297589 containerd[1887]: time="2025-02-13T20:03:42.297288745Z" level=info msg="CreateContainer within sandbox \"152eb051aef4bfc717576523d14f15a1db84c82c23ea09a6d2b544b2192b6d00\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"d1ca464a81ef826f7fe340eb56ac3f33f85f27cb71cc2999fcee9d026cbe621e\"" Feb 13 20:03:42.297835 containerd[1887]: time="2025-02-13T20:03:42.297804138Z" level=info msg="StartContainer for \"d1ca464a81ef826f7fe340eb56ac3f33f85f27cb71cc2999fcee9d026cbe621e\"" Feb 13 20:03:42.309730 containerd[1887]: time="2025-02-13T20:03:42.309681340Z" level=info msg="CreateContainer within sandbox \"971805c833cee288ab051dabb6e380c2a11c47aee76e29269adad77195285450\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0\"" Feb 13 20:03:42.310284 containerd[1887]: time="2025-02-13T20:03:42.310254508Z" level=info msg="StartContainer for \"0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0\"" Feb 13 20:03:42.325205 systemd[1]: Started cri-containerd-f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c.scope - libcontainer container f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c. Feb 13 20:03:42.374198 kubelet[2728]: E0213 20:03:42.374144 2728 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.25.222:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:42.376307 systemd[1]: Started cri-containerd-d1ca464a81ef826f7fe340eb56ac3f33f85f27cb71cc2999fcee9d026cbe621e.scope - libcontainer container d1ca464a81ef826f7fe340eb56ac3f33f85f27cb71cc2999fcee9d026cbe621e. Feb 13 20:03:42.408327 systemd[1]: Started cri-containerd-0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0.scope - libcontainer container 0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0. Feb 13 20:03:42.449312 containerd[1887]: time="2025-02-13T20:03:42.449158814Z" level=info msg="StartContainer for \"f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c\" returns successfully" Feb 13 20:03:42.512946 containerd[1887]: time="2025-02-13T20:03:42.512899821Z" level=info msg="StartContainer for \"d1ca464a81ef826f7fe340eb56ac3f33f85f27cb71cc2999fcee9d026cbe621e\" returns successfully" Feb 13 20:03:42.550788 containerd[1887]: time="2025-02-13T20:03:42.550741413Z" level=info msg="StartContainer for \"0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0\" returns successfully" Feb 13 20:03:43.342294 kubelet[2728]: E0213 20:03:43.342245 2728 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.25.222:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-222?timeout=10s\": dial tcp 172.31.25.222:6443: connect: connection refused" interval="3.2s" Feb 13 20:03:43.401768 kubelet[2728]: E0213 20:03:43.401586 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:43.406102 kubelet[2728]: E0213 20:03:43.406048 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:43.412170 kubelet[2728]: E0213 20:03:43.412138 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:43.423915 update_engine[1865]: I20250213 20:03:43.422147 1865 update_attempter.cc:509] Updating boot flags... Feb 13 20:03:43.437772 kubelet[2728]: W0213 20:03:43.437695 2728 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.25.222:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-222&limit=500&resourceVersion=0": dial tcp 172.31.25.222:6443: connect: connection refused Feb 13 20:03:43.438229 kubelet[2728]: E0213 20:03:43.437789 2728 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.25.222:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-25-222&limit=500&resourceVersion=0\": dial tcp 172.31.25.222:6443: connect: connection refused" logger="UnhandledError" Feb 13 20:03:43.548463 kubelet[2728]: I0213 20:03:43.546315 2728 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-222" Feb 13 20:03:43.587226 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (3013) Feb 13 20:03:43.987228 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (3015) Feb 13 20:03:44.411066 kubelet[2728]: E0213 20:03:44.411032 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:44.411664 kubelet[2728]: E0213 20:03:44.411567 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:44.411963 kubelet[2728]: E0213 20:03:44.411940 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:45.413838 kubelet[2728]: E0213 20:03:45.413801 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:45.416844 kubelet[2728]: E0213 20:03:45.416814 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:45.417074 kubelet[2728]: E0213 20:03:45.417054 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:46.340501 kubelet[2728]: I0213 20:03:46.340268 2728 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-25-222" Feb 13 20:03:46.340501 kubelet[2728]: E0213 20:03:46.340316 2728 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"ip-172-31-25-222\": node \"ip-172-31-25-222\" not found" Feb 13 20:03:46.356365 kubelet[2728]: E0213 20:03:46.355562 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:46.425784 kubelet[2728]: E0213 20:03:46.421075 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:46.425784 kubelet[2728]: E0213 20:03:46.421075 2728 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-25-222\" not found" node="ip-172-31-25-222" Feb 13 20:03:46.455799 kubelet[2728]: E0213 20:03:46.455746 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:46.556899 kubelet[2728]: E0213 20:03:46.556854 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:46.661034 kubelet[2728]: E0213 20:03:46.659074 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:46.759524 kubelet[2728]: E0213 20:03:46.759479 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:46.859746 kubelet[2728]: E0213 20:03:46.859702 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:46.960157 kubelet[2728]: E0213 20:03:46.959829 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:47.061764 kubelet[2728]: E0213 20:03:47.061724 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:47.162335 kubelet[2728]: E0213 20:03:47.162294 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:47.274076 kubelet[2728]: E0213 20:03:47.274032 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:47.374494 kubelet[2728]: E0213 20:03:47.374458 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:47.475786 kubelet[2728]: E0213 20:03:47.475736 2728 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-25-222\" not found" Feb 13 20:03:47.510514 kubelet[2728]: I0213 20:03:47.509831 2728 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:47.547921 kubelet[2728]: I0213 20:03:47.547366 2728 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:47.564661 kubelet[2728]: I0213 20:03:47.564407 2728 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-25-222" Feb 13 20:03:48.290862 kubelet[2728]: I0213 20:03:48.290818 2728 apiserver.go:52] "Watching apiserver" Feb 13 20:03:48.322368 kubelet[2728]: I0213 20:03:48.322324 2728 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 20:03:49.186828 systemd[1]: Reloading requested from client PID 3183 ('systemctl') (unit session-9.scope)... Feb 13 20:03:49.186846 systemd[1]: Reloading... Feb 13 20:03:49.364141 zram_generator::config[3223]: No configuration found. Feb 13 20:03:49.600711 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 20:03:49.771509 systemd[1]: Reloading finished in 583 ms. Feb 13 20:03:49.844480 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:03:49.865374 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 20:03:49.865686 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:49.876358 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 20:03:50.215038 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 20:03:50.230783 (kubelet)[3280]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 20:03:50.386936 kubelet[3280]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:03:50.386936 kubelet[3280]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 20:03:50.386936 kubelet[3280]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 20:03:50.387926 kubelet[3280]: I0213 20:03:50.387047 3280 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 20:03:50.400231 kubelet[3280]: I0213 20:03:50.399690 3280 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 20:03:50.400231 kubelet[3280]: I0213 20:03:50.399720 3280 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 20:03:50.401299 kubelet[3280]: I0213 20:03:50.400823 3280 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 20:03:50.406163 kubelet[3280]: I0213 20:03:50.405686 3280 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 20:03:50.423122 kubelet[3280]: I0213 20:03:50.423076 3280 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 20:03:50.435142 kubelet[3280]: E0213 20:03:50.433793 3280 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 20:03:50.435142 kubelet[3280]: I0213 20:03:50.433821 3280 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 20:03:50.437579 kubelet[3280]: I0213 20:03:50.437558 3280 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 20:03:50.438560 kubelet[3280]: I0213 20:03:50.438522 3280 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 20:03:50.438991 kubelet[3280]: I0213 20:03:50.438671 3280 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-25-222","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 20:03:50.439260 kubelet[3280]: I0213 20:03:50.439245 3280 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 20:03:50.439357 kubelet[3280]: I0213 20:03:50.439347 3280 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 20:03:50.439472 kubelet[3280]: I0213 20:03:50.439462 3280 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:03:50.447796 kubelet[3280]: I0213 20:03:50.447765 3280 kubelet.go:446] "Attempting to sync node with API server" Feb 13 20:03:50.448177 kubelet[3280]: I0213 20:03:50.448159 3280 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 20:03:50.448323 kubelet[3280]: I0213 20:03:50.448312 3280 kubelet.go:352] "Adding apiserver pod source" Feb 13 20:03:50.452058 kubelet[3280]: I0213 20:03:50.452032 3280 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 20:03:50.472799 kubelet[3280]: I0213 20:03:50.470790 3280 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Feb 13 20:03:50.472799 kubelet[3280]: I0213 20:03:50.472047 3280 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 20:03:50.478680 kubelet[3280]: I0213 20:03:50.478536 3280 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 20:03:50.478680 kubelet[3280]: I0213 20:03:50.478588 3280 server.go:1287] "Started kubelet" Feb 13 20:03:50.482706 kubelet[3280]: I0213 20:03:50.482669 3280 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 20:03:50.484131 kubelet[3280]: I0213 20:03:50.483661 3280 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 20:03:50.484245 kubelet[3280]: I0213 20:03:50.484177 3280 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 20:03:50.508128 kubelet[3280]: I0213 20:03:50.507956 3280 server.go:490] "Adding debug handlers to kubelet server" Feb 13 20:03:50.541332 kubelet[3280]: I0213 20:03:50.534393 3280 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 20:03:50.543350 kubelet[3280]: I0213 20:03:50.542835 3280 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 20:03:50.561439 kubelet[3280]: I0213 20:03:50.561412 3280 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 20:03:50.566179 kubelet[3280]: I0213 20:03:50.565641 3280 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 20:03:50.566704 kubelet[3280]: I0213 20:03:50.566498 3280 reconciler.go:26] "Reconciler: start to sync state" Feb 13 20:03:50.576530 kubelet[3280]: E0213 20:03:50.576493 3280 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 20:03:50.591893 kubelet[3280]: I0213 20:03:50.590673 3280 factory.go:221] Registration of the systemd container factory successfully Feb 13 20:03:50.591893 kubelet[3280]: I0213 20:03:50.590977 3280 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 20:03:50.593651 kubelet[3280]: I0213 20:03:50.593597 3280 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 20:03:50.597915 kubelet[3280]: I0213 20:03:50.597217 3280 factory.go:221] Registration of the containerd container factory successfully Feb 13 20:03:50.615456 kubelet[3280]: I0213 20:03:50.615411 3280 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 20:03:50.615456 kubelet[3280]: I0213 20:03:50.615467 3280 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 20:03:50.615662 kubelet[3280]: I0213 20:03:50.615491 3280 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 20:03:50.615662 kubelet[3280]: I0213 20:03:50.615499 3280 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 20:03:50.617960 kubelet[3280]: E0213 20:03:50.617915 3280 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 20:03:50.701834 kubelet[3280]: I0213 20:03:50.701810 3280 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 20:03:50.702136 kubelet[3280]: I0213 20:03:50.701998 3280 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 20:03:50.702136 kubelet[3280]: I0213 20:03:50.702029 3280 state_mem.go:36] "Initialized new in-memory state store" Feb 13 20:03:50.702267 kubelet[3280]: I0213 20:03:50.702255 3280 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 20:03:50.702311 kubelet[3280]: I0213 20:03:50.702269 3280 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 20:03:50.702311 kubelet[3280]: I0213 20:03:50.702295 3280 policy_none.go:49] "None policy: Start" Feb 13 20:03:50.702311 kubelet[3280]: I0213 20:03:50.702309 3280 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 20:03:50.702444 kubelet[3280]: I0213 20:03:50.702323 3280 state_mem.go:35] "Initializing new in-memory state store" Feb 13 20:03:50.702487 kubelet[3280]: I0213 20:03:50.702463 3280 state_mem.go:75] "Updated machine memory state" Feb 13 20:03:50.710556 kubelet[3280]: I0213 20:03:50.710493 3280 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 20:03:50.711776 kubelet[3280]: I0213 20:03:50.710734 3280 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 20:03:50.711776 kubelet[3280]: I0213 20:03:50.710751 3280 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 20:03:50.711776 kubelet[3280]: I0213 20:03:50.711568 3280 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 20:03:50.722389 kubelet[3280]: I0213 20:03:50.722356 3280 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-25-222" Feb 13 20:03:50.722906 kubelet[3280]: I0213 20:03:50.722820 3280 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:50.728550 kubelet[3280]: I0213 20:03:50.727229 3280 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:50.728550 kubelet[3280]: E0213 20:03:50.727323 3280 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 20:03:50.750941 kubelet[3280]: E0213 20:03:50.750657 3280 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-25-222\" already exists" pod="kube-system/kube-scheduler-ip-172-31-25-222" Feb 13 20:03:50.755602 kubelet[3280]: E0213 20:03:50.755563 3280 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-25-222\" already exists" pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:50.759786 kubelet[3280]: E0213 20:03:50.759361 3280 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-25-222\" already exists" pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:50.777864 kubelet[3280]: I0213 20:03:50.777355 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-kubeconfig\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:50.777864 kubelet[3280]: I0213 20:03:50.777409 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b204870b3e481c92b18dac4ca6ee6c9f-kubeconfig\") pod \"kube-scheduler-ip-172-31-25-222\" (UID: \"b204870b3e481c92b18dac4ca6ee6c9f\") " pod="kube-system/kube-scheduler-ip-172-31-25-222" Feb 13 20:03:50.777864 kubelet[3280]: I0213 20:03:50.777437 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/faf6713a9dad2a13975e98f5c591f88e-ca-certs\") pod \"kube-apiserver-ip-172-31-25-222\" (UID: \"faf6713a9dad2a13975e98f5c591f88e\") " pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:50.778382 kubelet[3280]: I0213 20:03:50.777462 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/faf6713a9dad2a13975e98f5c591f88e-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-25-222\" (UID: \"faf6713a9dad2a13975e98f5c591f88e\") " pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:50.780245 kubelet[3280]: I0213 20:03:50.778979 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:50.780245 kubelet[3280]: I0213 20:03:50.779058 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:50.780245 kubelet[3280]: I0213 20:03:50.779157 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/faf6713a9dad2a13975e98f5c591f88e-k8s-certs\") pod \"kube-apiserver-ip-172-31-25-222\" (UID: \"faf6713a9dad2a13975e98f5c591f88e\") " pod="kube-system/kube-apiserver-ip-172-31-25-222" Feb 13 20:03:50.780245 kubelet[3280]: I0213 20:03:50.779256 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-ca-certs\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:50.780245 kubelet[3280]: I0213 20:03:50.779325 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/73c480cf63f0b0ed72d37ea8cbc1df29-k8s-certs\") pod \"kube-controller-manager-ip-172-31-25-222\" (UID: \"73c480cf63f0b0ed72d37ea8cbc1df29\") " pod="kube-system/kube-controller-manager-ip-172-31-25-222" Feb 13 20:03:50.821096 kubelet[3280]: I0213 20:03:50.821068 3280 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-25-222" Feb 13 20:03:50.836655 kubelet[3280]: I0213 20:03:50.836621 3280 kubelet_node_status.go:125] "Node was previously registered" node="ip-172-31-25-222" Feb 13 20:03:50.836784 kubelet[3280]: I0213 20:03:50.836709 3280 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-25-222" Feb 13 20:03:51.455521 kubelet[3280]: I0213 20:03:51.454151 3280 apiserver.go:52] "Watching apiserver" Feb 13 20:03:51.466570 kubelet[3280]: I0213 20:03:51.466463 3280 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 20:03:51.676828 kubelet[3280]: I0213 20:03:51.676796 3280 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-25-222" Feb 13 20:03:51.687674 kubelet[3280]: E0213 20:03:51.686778 3280 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-25-222\" already exists" pod="kube-system/kube-scheduler-ip-172-31-25-222" Feb 13 20:03:51.728216 kubelet[3280]: I0213 20:03:51.726588 3280 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-25-222" podStartSLOduration=4.726561266 podStartE2EDuration="4.726561266s" podCreationTimestamp="2025-02-13 20:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:03:51.71150713 +0000 UTC m=+1.460731178" watchObservedRunningTime="2025-02-13 20:03:51.726561266 +0000 UTC m=+1.475785308" Feb 13 20:03:51.745402 kubelet[3280]: I0213 20:03:51.745325 3280 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-25-222" podStartSLOduration=4.745200111 podStartE2EDuration="4.745200111s" podCreationTimestamp="2025-02-13 20:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:03:51.727678394 +0000 UTC m=+1.476902443" watchObservedRunningTime="2025-02-13 20:03:51.745200111 +0000 UTC m=+1.494424150" Feb 13 20:03:51.745854 kubelet[3280]: I0213 20:03:51.745791 3280 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-25-222" podStartSLOduration=4.745775965 podStartE2EDuration="4.745775965s" podCreationTimestamp="2025-02-13 20:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:03:51.744803231 +0000 UTC m=+1.494027282" watchObservedRunningTime="2025-02-13 20:03:51.745775965 +0000 UTC m=+1.495000012" Feb 13 20:03:53.756139 kubelet[3280]: I0213 20:03:53.755694 3280 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 20:03:53.760838 containerd[1887]: time="2025-02-13T20:03:53.758920929Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 20:03:53.765074 kubelet[3280]: I0213 20:03:53.762722 3280 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 20:03:54.430953 sudo[2215]: pam_unix(sudo:session): session closed for user root Feb 13 20:03:54.455707 sshd[2212]: pam_unix(sshd:session): session closed for user core Feb 13 20:03:54.462605 systemd[1]: sshd@8-172.31.25.222:22-139.178.89.65:36010.service: Deactivated successfully. Feb 13 20:03:54.463095 systemd-logind[1863]: Session 9 logged out. Waiting for processes to exit. Feb 13 20:03:54.468086 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 20:03:54.468930 systemd[1]: session-9.scope: Consumed 4.415s CPU time, 143.4M memory peak, 0B memory swap peak. Feb 13 20:03:54.471600 systemd-logind[1863]: Removed session 9. Feb 13 20:03:54.736674 systemd[1]: Created slice kubepods-besteffort-pod0b029b31_7062_4da3_99d0_4dec4927241c.slice - libcontainer container kubepods-besteffort-pod0b029b31_7062_4da3_99d0_4dec4927241c.slice. Feb 13 20:03:54.820178 kubelet[3280]: I0213 20:03:54.820100 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0b029b31-7062-4da3-99d0-4dec4927241c-lib-modules\") pod \"kube-proxy-8jxqd\" (UID: \"0b029b31-7062-4da3-99d0-4dec4927241c\") " pod="kube-system/kube-proxy-8jxqd" Feb 13 20:03:54.820178 kubelet[3280]: I0213 20:03:54.820160 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/0b029b31-7062-4da3-99d0-4dec4927241c-kube-proxy\") pod \"kube-proxy-8jxqd\" (UID: \"0b029b31-7062-4da3-99d0-4dec4927241c\") " pod="kube-system/kube-proxy-8jxqd" Feb 13 20:03:54.820178 kubelet[3280]: I0213 20:03:54.820186 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0b029b31-7062-4da3-99d0-4dec4927241c-xtables-lock\") pod \"kube-proxy-8jxqd\" (UID: \"0b029b31-7062-4da3-99d0-4dec4927241c\") " pod="kube-system/kube-proxy-8jxqd" Feb 13 20:03:54.820765 kubelet[3280]: I0213 20:03:54.820209 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64r4x\" (UniqueName: \"kubernetes.io/projected/0b029b31-7062-4da3-99d0-4dec4927241c-kube-api-access-64r4x\") pod \"kube-proxy-8jxqd\" (UID: \"0b029b31-7062-4da3-99d0-4dec4927241c\") " pod="kube-system/kube-proxy-8jxqd" Feb 13 20:03:54.900038 systemd[1]: Created slice kubepods-besteffort-pod06a481f3_5452_4ea8_b0bf_7258f2a4e56c.slice - libcontainer container kubepods-besteffort-pod06a481f3_5452_4ea8_b0bf_7258f2a4e56c.slice. Feb 13 20:03:54.921319 kubelet[3280]: I0213 20:03:54.921268 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/06a481f3-5452-4ea8-b0bf-7258f2a4e56c-var-lib-calico\") pod \"tigera-operator-7d68577dc5-wzv5x\" (UID: \"06a481f3-5452-4ea8-b0bf-7258f2a4e56c\") " pod="tigera-operator/tigera-operator-7d68577dc5-wzv5x" Feb 13 20:03:54.921514 kubelet[3280]: I0213 20:03:54.921337 3280 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml2j8\" (UniqueName: \"kubernetes.io/projected/06a481f3-5452-4ea8-b0bf-7258f2a4e56c-kube-api-access-ml2j8\") pod \"tigera-operator-7d68577dc5-wzv5x\" (UID: \"06a481f3-5452-4ea8-b0bf-7258f2a4e56c\") " pod="tigera-operator/tigera-operator-7d68577dc5-wzv5x" Feb 13 20:03:55.050026 containerd[1887]: time="2025-02-13T20:03:55.049412299Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8jxqd,Uid:0b029b31-7062-4da3-99d0-4dec4927241c,Namespace:kube-system,Attempt:0,}" Feb 13 20:03:55.091743 containerd[1887]: time="2025-02-13T20:03:55.091177481Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:03:55.091743 containerd[1887]: time="2025-02-13T20:03:55.091253035Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:03:55.091743 containerd[1887]: time="2025-02-13T20:03:55.091304665Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:55.091743 containerd[1887]: time="2025-02-13T20:03:55.091461865Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:55.122751 systemd[1]: Started cri-containerd-b06bca60bc99cac496457c0f0e75ac0571f571c801c05b79249b5aedb012a2c0.scope - libcontainer container b06bca60bc99cac496457c0f0e75ac0571f571c801c05b79249b5aedb012a2c0. Feb 13 20:03:55.149487 containerd[1887]: time="2025-02-13T20:03:55.149422807Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8jxqd,Uid:0b029b31-7062-4da3-99d0-4dec4927241c,Namespace:kube-system,Attempt:0,} returns sandbox id \"b06bca60bc99cac496457c0f0e75ac0571f571c801c05b79249b5aedb012a2c0\"" Feb 13 20:03:55.153253 containerd[1887]: time="2025-02-13T20:03:55.153211002Z" level=info msg="CreateContainer within sandbox \"b06bca60bc99cac496457c0f0e75ac0571f571c801c05b79249b5aedb012a2c0\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 20:03:55.179188 containerd[1887]: time="2025-02-13T20:03:55.179123281Z" level=info msg="CreateContainer within sandbox \"b06bca60bc99cac496457c0f0e75ac0571f571c801c05b79249b5aedb012a2c0\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"fc31d5c181702a844f2710b133e2fa0f840f0934f0d588dec52c42cee7f65c8f\"" Feb 13 20:03:55.181215 containerd[1887]: time="2025-02-13T20:03:55.179914645Z" level=info msg="StartContainer for \"fc31d5c181702a844f2710b133e2fa0f840f0934f0d588dec52c42cee7f65c8f\"" Feb 13 20:03:55.206658 containerd[1887]: time="2025-02-13T20:03:55.206622479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7d68577dc5-wzv5x,Uid:06a481f3-5452-4ea8-b0bf-7258f2a4e56c,Namespace:tigera-operator,Attempt:0,}" Feb 13 20:03:55.219477 systemd[1]: Started cri-containerd-fc31d5c181702a844f2710b133e2fa0f840f0934f0d588dec52c42cee7f65c8f.scope - libcontainer container fc31d5c181702a844f2710b133e2fa0f840f0934f0d588dec52c42cee7f65c8f. Feb 13 20:03:55.275453 containerd[1887]: time="2025-02-13T20:03:55.275401589Z" level=info msg="StartContainer for \"fc31d5c181702a844f2710b133e2fa0f840f0934f0d588dec52c42cee7f65c8f\" returns successfully" Feb 13 20:03:55.279828 containerd[1887]: time="2025-02-13T20:03:55.279027480Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 20:03:55.279828 containerd[1887]: time="2025-02-13T20:03:55.279083162Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 20:03:55.279828 containerd[1887]: time="2025-02-13T20:03:55.279098403Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:55.279828 containerd[1887]: time="2025-02-13T20:03:55.279211318Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 20:03:55.329344 systemd[1]: Started cri-containerd-4b9148b6657c020355b0b6e28d7c62cb7608e0483f7b8366c78355f0dd09f14c.scope - libcontainer container 4b9148b6657c020355b0b6e28d7c62cb7608e0483f7b8366c78355f0dd09f14c. Feb 13 20:03:55.508357 containerd[1887]: time="2025-02-13T20:03:55.508309040Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7d68577dc5-wzv5x,Uid:06a481f3-5452-4ea8-b0bf-7258f2a4e56c,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"4b9148b6657c020355b0b6e28d7c62cb7608e0483f7b8366c78355f0dd09f14c\"" Feb 13 20:03:55.510900 containerd[1887]: time="2025-02-13T20:03:55.510843716Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\"" Feb 13 20:03:56.707957 kubelet[3280]: I0213 20:03:56.707845 3280 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-8jxqd" podStartSLOduration=2.707823679 podStartE2EDuration="2.707823679s" podCreationTimestamp="2025-02-13 20:03:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 20:03:55.721493789 +0000 UTC m=+5.470717848" watchObservedRunningTime="2025-02-13 20:03:56.707823679 +0000 UTC m=+6.457047738" Feb 13 20:03:57.217377 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2801879931.mount: Deactivated successfully. Feb 13 20:03:58.176547 containerd[1887]: time="2025-02-13T20:03:58.176491649Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:58.177789 containerd[1887]: time="2025-02-13T20:03:58.177735851Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.2: active requests=0, bytes read=21762497" Feb 13 20:03:58.179366 containerd[1887]: time="2025-02-13T20:03:58.178706095Z" level=info msg="ImageCreate event name:\"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:58.181776 containerd[1887]: time="2025-02-13T20:03:58.181723224Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 20:03:58.183205 containerd[1887]: time="2025-02-13T20:03:58.182641581Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.2\" with image id \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\", repo tag \"quay.io/tigera/operator:v1.36.2\", repo digest \"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\", size \"21758492\" in 2.671753884s" Feb 13 20:03:58.183205 containerd[1887]: time="2025-02-13T20:03:58.182687001Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\" returns image reference \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\"" Feb 13 20:03:58.187075 containerd[1887]: time="2025-02-13T20:03:58.187034665Z" level=info msg="CreateContainer within sandbox \"4b9148b6657c020355b0b6e28d7c62cb7608e0483f7b8366c78355f0dd09f14c\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Feb 13 20:03:58.202321 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3604957927.mount: Deactivated successfully. Feb 13 20:03:58.204416 containerd[1887]: time="2025-02-13T20:03:58.204235851Z" level=info msg="CreateContainer within sandbox \"4b9148b6657c020355b0b6e28d7c62cb7608e0483f7b8366c78355f0dd09f14c\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"e44663baa53654dbe1f872c6906965334e2af973729ac817a56d317bd9cab7ad\"" Feb 13 20:03:58.207192 containerd[1887]: time="2025-02-13T20:03:58.205743840Z" level=info msg="StartContainer for \"e44663baa53654dbe1f872c6906965334e2af973729ac817a56d317bd9cab7ad\"" Feb 13 20:03:58.244356 systemd[1]: Started cri-containerd-e44663baa53654dbe1f872c6906965334e2af973729ac817a56d317bd9cab7ad.scope - libcontainer container e44663baa53654dbe1f872c6906965334e2af973729ac817a56d317bd9cab7ad. Feb 13 20:03:58.280099 containerd[1887]: time="2025-02-13T20:03:58.279957238Z" level=info msg="StartContainer for \"e44663baa53654dbe1f872c6906965334e2af973729ac817a56d317bd9cab7ad\" returns successfully" Feb 13 20:04:10.990311 kubelet[3280]: E0213 20:04:10.990188 3280 controller.go:195] "Failed to update lease" err="Put \"https://172.31.25.222:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-222?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 13 20:04:11.015654 systemd[1]: cri-containerd-f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c.scope: Deactivated successfully. Feb 13 20:04:11.017034 systemd[1]: cri-containerd-f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c.scope: Consumed 2.003s CPU time, 37.4M memory peak, 0B memory swap peak. Feb 13 20:04:11.090076 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c-rootfs.mount: Deactivated successfully. Feb 13 20:04:11.103911 containerd[1887]: time="2025-02-13T20:04:11.103843268Z" level=info msg="shim disconnected" id=f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c namespace=k8s.io Feb 13 20:04:11.103911 containerd[1887]: time="2025-02-13T20:04:11.103899117Z" level=warning msg="cleaning up after shim disconnected" id=f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c namespace=k8s.io Feb 13 20:04:11.103911 containerd[1887]: time="2025-02-13T20:04:11.103911829Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:04:11.759586 kubelet[3280]: I0213 20:04:11.759537 3280 scope.go:117] "RemoveContainer" containerID="f7988ee9ed6924b6c20f17a2351410cb67a37c6ead6f41f2075ab0fbac80280c" Feb 13 20:04:11.771508 containerd[1887]: time="2025-02-13T20:04:11.770591842Z" level=info msg="CreateContainer within sandbox \"0183c4112df27dd005e00044356f7fe4536208e12aa7db7cdd0fb219001b88fc\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Feb 13 20:04:11.815511 containerd[1887]: time="2025-02-13T20:04:11.815454793Z" level=info msg="CreateContainer within sandbox \"0183c4112df27dd005e00044356f7fe4536208e12aa7db7cdd0fb219001b88fc\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"d5ccb57e815f33eb5e62272dea7a670204cb433d9988a400fd68eb81c3310303\"" Feb 13 20:04:11.816258 containerd[1887]: time="2025-02-13T20:04:11.816214296Z" level=info msg="StartContainer for \"d5ccb57e815f33eb5e62272dea7a670204cb433d9988a400fd68eb81c3310303\"" Feb 13 20:04:11.877686 systemd[1]: Started cri-containerd-d5ccb57e815f33eb5e62272dea7a670204cb433d9988a400fd68eb81c3310303.scope - libcontainer container d5ccb57e815f33eb5e62272dea7a670204cb433d9988a400fd68eb81c3310303. Feb 13 20:04:11.950476 containerd[1887]: time="2025-02-13T20:04:11.950375598Z" level=info msg="StartContainer for \"d5ccb57e815f33eb5e62272dea7a670204cb433d9988a400fd68eb81c3310303\" returns successfully" Feb 13 20:04:16.008034 systemd[1]: cri-containerd-0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0.scope: Deactivated successfully. Feb 13 20:04:16.009386 systemd[1]: cri-containerd-0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0.scope: Consumed 1.234s CPU time, 19.7M memory peak, 0B memory swap peak. Feb 13 20:04:16.082692 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0-rootfs.mount: Deactivated successfully. Feb 13 20:04:16.102681 containerd[1887]: time="2025-02-13T20:04:16.102522522Z" level=info msg="shim disconnected" id=0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0 namespace=k8s.io Feb 13 20:04:16.102681 containerd[1887]: time="2025-02-13T20:04:16.102682493Z" level=warning msg="cleaning up after shim disconnected" id=0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0 namespace=k8s.io Feb 13 20:04:16.103273 containerd[1887]: time="2025-02-13T20:04:16.102698484Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 20:04:16.762756 kubelet[3280]: I0213 20:04:16.762723 3280 scope.go:117] "RemoveContainer" containerID="0d9ceeab2e5d6da1bf12e9e5880ec50fa4c9074bb5bc5325732adf9fb83479e0" Feb 13 20:04:16.764310 containerd[1887]: time="2025-02-13T20:04:16.764264479Z" level=info msg="CreateContainer within sandbox \"971805c833cee288ab051dabb6e380c2a11c47aee76e29269adad77195285450\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Feb 13 20:04:16.788759 containerd[1887]: time="2025-02-13T20:04:16.788721216Z" level=info msg="CreateContainer within sandbox \"971805c833cee288ab051dabb6e380c2a11c47aee76e29269adad77195285450\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"694d6be75bd25384cc5334c71753f581ebf9a5e1483463e5614034199e7fd5b6\"" Feb 13 20:04:16.790289 containerd[1887]: time="2025-02-13T20:04:16.789188790Z" level=info msg="StartContainer for \"694d6be75bd25384cc5334c71753f581ebf9a5e1483463e5614034199e7fd5b6\"" Feb 13 20:04:16.830432 systemd[1]: Started cri-containerd-694d6be75bd25384cc5334c71753f581ebf9a5e1483463e5614034199e7fd5b6.scope - libcontainer container 694d6be75bd25384cc5334c71753f581ebf9a5e1483463e5614034199e7fd5b6. Feb 13 20:04:16.917870 containerd[1887]: time="2025-02-13T20:04:16.916588935Z" level=info msg="StartContainer for \"694d6be75bd25384cc5334c71753f581ebf9a5e1483463e5614034199e7fd5b6\" returns successfully" Feb 13 20:04:20.990845 kubelet[3280]: E0213 20:04:20.990796 3280 controller.go:195] "Failed to update lease" err="Put \"https://172.31.25.222:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-25-222?timeout=10s\": context deadline exceeded"