Sep 4 17:39:35.083068 kernel: Linux version 6.6.48-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Wed Sep 4 15:54:07 -00 2024 Sep 4 17:39:35.083112 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:39:35.083128 kernel: BIOS-provided physical RAM map: Sep 4 17:39:35.083140 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Sep 4 17:39:35.083152 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Sep 4 17:39:35.083164 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Sep 4 17:39:35.083183 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007d9e9fff] usable Sep 4 17:39:35.083297 kernel: BIOS-e820: [mem 0x000000007d9ea000-0x000000007fffffff] reserved Sep 4 17:39:35.083312 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000e03fffff] reserved Sep 4 17:39:35.083325 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Sep 4 17:39:35.083338 kernel: NX (Execute Disable) protection: active Sep 4 17:39:35.083351 kernel: APIC: Static calls initialized Sep 4 17:39:35.083363 kernel: SMBIOS 2.7 present. Sep 4 17:39:35.083376 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Sep 4 17:39:35.084625 kernel: Hypervisor detected: KVM Sep 4 17:39:35.084646 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 4 17:39:35.084661 kernel: kvm-clock: using sched offset of 6654322283 cycles Sep 4 17:39:35.084677 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 4 17:39:35.084698 kernel: tsc: Detected 2499.996 MHz processor Sep 4 17:39:35.084711 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 4 17:39:35.084727 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 4 17:39:35.084745 kernel: last_pfn = 0x7d9ea max_arch_pfn = 0x400000000 Sep 4 17:39:35.084760 kernel: MTRR map: 4 entries (3 fixed + 1 variable; max 19), built from 8 variable MTRRs Sep 4 17:39:35.084773 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 4 17:39:35.084785 kernel: Using GB pages for direct mapping Sep 4 17:39:35.084797 kernel: ACPI: Early table checksum verification disabled Sep 4 17:39:35.084809 kernel: ACPI: RSDP 0x00000000000F8F40 000014 (v00 AMAZON) Sep 4 17:39:35.084822 kernel: ACPI: RSDT 0x000000007D9EE350 000044 (v01 AMAZON AMZNRSDT 00000001 AMZN 00000001) Sep 4 17:39:35.084836 kernel: ACPI: FACP 0x000000007D9EFF80 000074 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Sep 4 17:39:35.084847 kernel: ACPI: DSDT 0x000000007D9EE3A0 0010E9 (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Sep 4 17:39:35.084861 kernel: ACPI: FACS 0x000000007D9EFF40 000040 Sep 4 17:39:35.084878 kernel: ACPI: SSDT 0x000000007D9EF6C0 00087A (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Sep 4 17:39:35.084893 kernel: ACPI: APIC 0x000000007D9EF5D0 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Sep 4 17:39:35.084911 kernel: ACPI: SRAT 0x000000007D9EF530 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Sep 4 17:39:35.084927 kernel: ACPI: SLIT 0x000000007D9EF4C0 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Sep 4 17:39:35.084944 kernel: ACPI: WAET 0x000000007D9EF490 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Sep 4 17:39:35.084959 kernel: ACPI: HPET 0x00000000000C9000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Sep 4 17:39:35.084972 kernel: ACPI: SSDT 0x00000000000C9040 00007B (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Sep 4 17:39:35.084986 kernel: ACPI: Reserving FACP table memory at [mem 0x7d9eff80-0x7d9efff3] Sep 4 17:39:35.085004 kernel: ACPI: Reserving DSDT table memory at [mem 0x7d9ee3a0-0x7d9ef488] Sep 4 17:39:35.085023 kernel: ACPI: Reserving FACS table memory at [mem 0x7d9eff40-0x7d9eff7f] Sep 4 17:39:35.085038 kernel: ACPI: Reserving SSDT table memory at [mem 0x7d9ef6c0-0x7d9eff39] Sep 4 17:39:35.085053 kernel: ACPI: Reserving APIC table memory at [mem 0x7d9ef5d0-0x7d9ef645] Sep 4 17:39:35.085068 kernel: ACPI: Reserving SRAT table memory at [mem 0x7d9ef530-0x7d9ef5cf] Sep 4 17:39:35.085152 kernel: ACPI: Reserving SLIT table memory at [mem 0x7d9ef4c0-0x7d9ef52b] Sep 4 17:39:35.085168 kernel: ACPI: Reserving WAET table memory at [mem 0x7d9ef490-0x7d9ef4b7] Sep 4 17:39:35.085184 kernel: ACPI: Reserving HPET table memory at [mem 0xc9000-0xc9037] Sep 4 17:39:35.085199 kernel: ACPI: Reserving SSDT table memory at [mem 0xc9040-0xc90ba] Sep 4 17:39:35.085215 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 4 17:39:35.085230 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 4 17:39:35.085346 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Sep 4 17:39:35.085362 kernel: NUMA: Initialized distance table, cnt=1 Sep 4 17:39:35.085377 kernel: NODE_DATA(0) allocated [mem 0x7d9e3000-0x7d9e8fff] Sep 4 17:39:35.087048 kernel: Zone ranges: Sep 4 17:39:35.087070 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 4 17:39:35.087083 kernel: DMA32 [mem 0x0000000001000000-0x000000007d9e9fff] Sep 4 17:39:35.087096 kernel: Normal empty Sep 4 17:39:35.087109 kernel: Movable zone start for each node Sep 4 17:39:35.087122 kernel: Early memory node ranges Sep 4 17:39:35.087135 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Sep 4 17:39:35.087148 kernel: node 0: [mem 0x0000000000100000-0x000000007d9e9fff] Sep 4 17:39:35.087161 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007d9e9fff] Sep 4 17:39:35.087174 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 4 17:39:35.087192 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Sep 4 17:39:35.087206 kernel: On node 0, zone DMA32: 9750 pages in unavailable ranges Sep 4 17:39:35.087219 kernel: ACPI: PM-Timer IO Port: 0xb008 Sep 4 17:39:35.087232 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 4 17:39:35.087244 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Sep 4 17:39:35.087257 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 4 17:39:35.087270 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 4 17:39:35.087282 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 4 17:39:35.087295 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 4 17:39:35.087312 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 4 17:39:35.087325 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 4 17:39:35.087337 kernel: TSC deadline timer available Sep 4 17:39:35.087350 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Sep 4 17:39:35.087363 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 4 17:39:35.087429 kernel: [mem 0x80000000-0xdfffffff] available for PCI devices Sep 4 17:39:35.087444 kernel: Booting paravirtualized kernel on KVM Sep 4 17:39:35.087457 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 4 17:39:35.087470 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Sep 4 17:39:35.087487 kernel: percpu: Embedded 58 pages/cpu s196904 r8192 d32472 u1048576 Sep 4 17:39:35.087500 kernel: pcpu-alloc: s196904 r8192 d32472 u1048576 alloc=1*2097152 Sep 4 17:39:35.087512 kernel: pcpu-alloc: [0] 0 1 Sep 4 17:39:35.087525 kernel: kvm-guest: PV spinlocks enabled Sep 4 17:39:35.087538 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 4 17:39:35.087552 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:39:35.087566 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 4 17:39:35.087578 kernel: random: crng init done Sep 4 17:39:35.087593 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 4 17:39:35.087606 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 4 17:39:35.087619 kernel: Fallback order for Node 0: 0 Sep 4 17:39:35.087632 kernel: Built 1 zonelists, mobility grouping on. Total pages: 506242 Sep 4 17:39:35.087644 kernel: Policy zone: DMA32 Sep 4 17:39:35.087657 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 4 17:39:35.087670 kernel: Memory: 1932348K/2057760K available (12288K kernel code, 2304K rwdata, 22708K rodata, 42704K init, 2488K bss, 125152K reserved, 0K cma-reserved) Sep 4 17:39:35.087683 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 4 17:39:35.087696 kernel: Kernel/User page tables isolation: enabled Sep 4 17:39:35.087712 kernel: ftrace: allocating 37748 entries in 148 pages Sep 4 17:39:35.087725 kernel: ftrace: allocated 148 pages with 3 groups Sep 4 17:39:35.087738 kernel: Dynamic Preempt: voluntary Sep 4 17:39:35.087751 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 4 17:39:35.087765 kernel: rcu: RCU event tracing is enabled. Sep 4 17:39:35.087828 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 4 17:39:35.087842 kernel: Trampoline variant of Tasks RCU enabled. Sep 4 17:39:35.087854 kernel: Rude variant of Tasks RCU enabled. Sep 4 17:39:35.087867 kernel: Tracing variant of Tasks RCU enabled. Sep 4 17:39:35.087884 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 4 17:39:35.087897 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 4 17:39:35.087909 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Sep 4 17:39:35.087922 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 4 17:39:35.087935 kernel: Console: colour VGA+ 80x25 Sep 4 17:39:35.087948 kernel: printk: console [ttyS0] enabled Sep 4 17:39:35.087961 kernel: ACPI: Core revision 20230628 Sep 4 17:39:35.087974 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Sep 4 17:39:35.087987 kernel: APIC: Switch to symmetric I/O mode setup Sep 4 17:39:35.088003 kernel: x2apic enabled Sep 4 17:39:35.088017 kernel: APIC: Switched APIC routing to: physical x2apic Sep 4 17:39:35.088042 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Sep 4 17:39:35.088059 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499996) Sep 4 17:39:35.088072 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Sep 4 17:39:35.088085 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Sep 4 17:39:35.088099 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 4 17:39:35.088112 kernel: Spectre V2 : Mitigation: Retpolines Sep 4 17:39:35.088126 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Sep 4 17:39:35.088139 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Sep 4 17:39:35.088153 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Sep 4 17:39:35.088167 kernel: RETBleed: Vulnerable Sep 4 17:39:35.088184 kernel: Speculative Store Bypass: Vulnerable Sep 4 17:39:35.088197 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Sep 4 17:39:35.088211 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 4 17:39:35.088224 kernel: GDS: Unknown: Dependent on hypervisor status Sep 4 17:39:35.088237 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 4 17:39:35.088251 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 4 17:39:35.089423 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 4 17:39:35.089452 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Sep 4 17:39:35.089468 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Sep 4 17:39:35.089484 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Sep 4 17:39:35.089500 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Sep 4 17:39:35.089517 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Sep 4 17:39:35.089533 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Sep 4 17:39:35.089550 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 4 17:39:35.089566 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Sep 4 17:39:35.089582 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Sep 4 17:39:35.089598 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Sep 4 17:39:35.089618 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Sep 4 17:39:35.089634 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Sep 4 17:39:35.089649 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Sep 4 17:39:35.089666 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Sep 4 17:39:35.089682 kernel: Freeing SMP alternatives memory: 32K Sep 4 17:39:35.089698 kernel: pid_max: default: 32768 minimum: 301 Sep 4 17:39:35.089713 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Sep 4 17:39:35.089730 kernel: landlock: Up and running. Sep 4 17:39:35.089745 kernel: SELinux: Initializing. Sep 4 17:39:35.089761 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 4 17:39:35.089778 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 4 17:39:35.089794 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) Sep 4 17:39:35.089813 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:39:35.089829 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:39:35.089846 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:39:35.089862 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Sep 4 17:39:35.089879 kernel: signal: max sigframe size: 3632 Sep 4 17:39:35.089895 kernel: rcu: Hierarchical SRCU implementation. Sep 4 17:39:35.089912 kernel: rcu: Max phase no-delay instances is 400. Sep 4 17:39:35.089928 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 4 17:39:35.089944 kernel: smp: Bringing up secondary CPUs ... Sep 4 17:39:35.089963 kernel: smpboot: x86: Booting SMP configuration: Sep 4 17:39:35.089980 kernel: .... node #0, CPUs: #1 Sep 4 17:39:35.091047 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Sep 4 17:39:35.091069 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 4 17:39:35.091085 kernel: smp: Brought up 1 node, 2 CPUs Sep 4 17:39:35.091101 kernel: smpboot: Max logical packages: 1 Sep 4 17:39:35.091116 kernel: smpboot: Total of 2 processors activated (9999.98 BogoMIPS) Sep 4 17:39:35.091132 kernel: devtmpfs: initialized Sep 4 17:39:35.091147 kernel: x86/mm: Memory block size: 128MB Sep 4 17:39:35.091167 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 4 17:39:35.091183 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 4 17:39:35.091199 kernel: pinctrl core: initialized pinctrl subsystem Sep 4 17:39:35.091214 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 4 17:39:35.091230 kernel: audit: initializing netlink subsys (disabled) Sep 4 17:39:35.091245 kernel: audit: type=2000 audit(1725471574.371:1): state=initialized audit_enabled=0 res=1 Sep 4 17:39:35.091260 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 4 17:39:35.091276 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 4 17:39:35.091294 kernel: cpuidle: using governor menu Sep 4 17:39:35.091310 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 4 17:39:35.091326 kernel: dca service started, version 1.12.1 Sep 4 17:39:35.091341 kernel: PCI: Using configuration type 1 for base access Sep 4 17:39:35.091357 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 4 17:39:35.091372 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 4 17:39:35.092471 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 4 17:39:35.092496 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 4 17:39:35.092512 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 4 17:39:35.092532 kernel: ACPI: Added _OSI(Module Device) Sep 4 17:39:35.092546 kernel: ACPI: Added _OSI(Processor Device) Sep 4 17:39:35.092559 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Sep 4 17:39:35.092573 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 4 17:39:35.092587 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Sep 4 17:39:35.092601 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Sep 4 17:39:35.092615 kernel: ACPI: Interpreter enabled Sep 4 17:39:35.092628 kernel: ACPI: PM: (supports S0 S5) Sep 4 17:39:35.092641 kernel: ACPI: Using IOAPIC for interrupt routing Sep 4 17:39:35.092655 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 4 17:39:35.092672 kernel: PCI: Using E820 reservations for host bridge windows Sep 4 17:39:35.092692 kernel: ACPI: Enabled 16 GPEs in block 00 to 0F Sep 4 17:39:35.092706 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 4 17:39:35.092921 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Sep 4 17:39:35.093059 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Sep 4 17:39:35.093182 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Sep 4 17:39:35.093199 kernel: acpiphp: Slot [3] registered Sep 4 17:39:35.093218 kernel: acpiphp: Slot [4] registered Sep 4 17:39:35.093232 kernel: acpiphp: Slot [5] registered Sep 4 17:39:35.093245 kernel: acpiphp: Slot [6] registered Sep 4 17:39:35.093259 kernel: acpiphp: Slot [7] registered Sep 4 17:39:35.093272 kernel: acpiphp: Slot [8] registered Sep 4 17:39:35.093287 kernel: acpiphp: Slot [9] registered Sep 4 17:39:35.093302 kernel: acpiphp: Slot [10] registered Sep 4 17:39:35.093317 kernel: acpiphp: Slot [11] registered Sep 4 17:39:35.093332 kernel: acpiphp: Slot [12] registered Sep 4 17:39:35.093349 kernel: acpiphp: Slot [13] registered Sep 4 17:39:35.093364 kernel: acpiphp: Slot [14] registered Sep 4 17:39:35.093378 kernel: acpiphp: Slot [15] registered Sep 4 17:39:35.093414 kernel: acpiphp: Slot [16] registered Sep 4 17:39:35.093429 kernel: acpiphp: Slot [17] registered Sep 4 17:39:35.093444 kernel: acpiphp: Slot [18] registered Sep 4 17:39:35.093459 kernel: acpiphp: Slot [19] registered Sep 4 17:39:35.093474 kernel: acpiphp: Slot [20] registered Sep 4 17:39:35.093488 kernel: acpiphp: Slot [21] registered Sep 4 17:39:35.093503 kernel: acpiphp: Slot [22] registered Sep 4 17:39:35.093522 kernel: acpiphp: Slot [23] registered Sep 4 17:39:35.093536 kernel: acpiphp: Slot [24] registered Sep 4 17:39:35.093551 kernel: acpiphp: Slot [25] registered Sep 4 17:39:35.093566 kernel: acpiphp: Slot [26] registered Sep 4 17:39:35.093580 kernel: acpiphp: Slot [27] registered Sep 4 17:39:35.093595 kernel: acpiphp: Slot [28] registered Sep 4 17:39:35.093610 kernel: acpiphp: Slot [29] registered Sep 4 17:39:35.093625 kernel: acpiphp: Slot [30] registered Sep 4 17:39:35.093639 kernel: acpiphp: Slot [31] registered Sep 4 17:39:35.093657 kernel: PCI host bridge to bus 0000:00 Sep 4 17:39:35.093789 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 4 17:39:35.093900 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 4 17:39:35.094008 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 4 17:39:35.094361 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Sep 4 17:39:35.096295 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 4 17:39:35.096469 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Sep 4 17:39:35.096621 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Sep 4 17:39:35.096773 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 Sep 4 17:39:35.096906 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Sep 4 17:39:35.097072 kernel: pci 0000:00:01.3: quirk: [io 0xb100-0xb10f] claimed by PIIX4 SMB Sep 4 17:39:35.097204 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Sep 4 17:39:35.097630 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Sep 4 17:39:35.097889 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Sep 4 17:39:35.098123 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Sep 4 17:39:35.098290 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Sep 4 17:39:35.099862 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Sep 4 17:39:35.100265 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 Sep 4 17:39:35.101088 kernel: pci 0000:00:03.0: reg 0x10: [mem 0xfe400000-0xfe7fffff pref] Sep 4 17:39:35.101234 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Sep 4 17:39:35.102541 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 4 17:39:35.102706 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Sep 4 17:39:35.102843 kernel: pci 0000:00:04.0: reg 0x10: [mem 0xfebf0000-0xfebf3fff] Sep 4 17:39:35.102984 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Sep 4 17:39:35.103116 kernel: pci 0000:00:05.0: reg 0x10: [mem 0xfebf4000-0xfebf7fff] Sep 4 17:39:35.103137 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 4 17:39:35.103154 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 4 17:39:35.103171 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 4 17:39:35.103191 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 4 17:39:35.103206 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Sep 4 17:39:35.103221 kernel: iommu: Default domain type: Translated Sep 4 17:39:35.103237 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 4 17:39:35.103254 kernel: PCI: Using ACPI for IRQ routing Sep 4 17:39:35.103270 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 4 17:39:35.103287 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Sep 4 17:39:35.103302 kernel: e820: reserve RAM buffer [mem 0x7d9ea000-0x7fffffff] Sep 4 17:39:35.103454 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Sep 4 17:39:35.103596 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Sep 4 17:39:35.103729 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 4 17:39:35.103750 kernel: vgaarb: loaded Sep 4 17:39:35.103766 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Sep 4 17:39:35.103781 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Sep 4 17:39:35.103796 kernel: clocksource: Switched to clocksource kvm-clock Sep 4 17:39:35.103812 kernel: VFS: Disk quotas dquot_6.6.0 Sep 4 17:39:35.103828 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 4 17:39:35.103845 kernel: pnp: PnP ACPI init Sep 4 17:39:35.103859 kernel: pnp: PnP ACPI: found 5 devices Sep 4 17:39:35.103873 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 4 17:39:35.103889 kernel: NET: Registered PF_INET protocol family Sep 4 17:39:35.103902 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 4 17:39:35.103914 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Sep 4 17:39:35.103929 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 4 17:39:35.103946 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 4 17:39:35.103961 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 4 17:39:35.103978 kernel: TCP: Hash tables configured (established 16384 bind 16384) Sep 4 17:39:35.103995 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 4 17:39:35.104011 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 4 17:39:35.104026 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 4 17:39:35.104042 kernel: NET: Registered PF_XDP protocol family Sep 4 17:39:35.104182 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 4 17:39:35.104313 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 4 17:39:35.108478 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 4 17:39:35.108635 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Sep 4 17:39:35.108799 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Sep 4 17:39:35.108823 kernel: PCI: CLS 0 bytes, default 64 Sep 4 17:39:35.108842 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 4 17:39:35.108859 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Sep 4 17:39:35.108876 kernel: clocksource: Switched to clocksource tsc Sep 4 17:39:35.108893 kernel: Initialise system trusted keyrings Sep 4 17:39:35.108910 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Sep 4 17:39:35.108930 kernel: Key type asymmetric registered Sep 4 17:39:35.108947 kernel: Asymmetric key parser 'x509' registered Sep 4 17:39:35.108964 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Sep 4 17:39:35.108981 kernel: io scheduler mq-deadline registered Sep 4 17:39:35.108998 kernel: io scheduler kyber registered Sep 4 17:39:35.109015 kernel: io scheduler bfq registered Sep 4 17:39:35.109031 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 4 17:39:35.109047 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 4 17:39:35.109063 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 4 17:39:35.109083 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 4 17:39:35.109100 kernel: i8042: Warning: Keylock active Sep 4 17:39:35.109117 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 4 17:39:35.109133 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 4 17:39:35.110625 kernel: rtc_cmos 00:00: RTC can wake from S4 Sep 4 17:39:35.110775 kernel: rtc_cmos 00:00: registered as rtc0 Sep 4 17:39:35.110904 kernel: rtc_cmos 00:00: setting system clock to 2024-09-04T17:39:34 UTC (1725471574) Sep 4 17:39:35.111030 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Sep 4 17:39:35.111057 kernel: intel_pstate: CPU model not supported Sep 4 17:39:35.111075 kernel: NET: Registered PF_INET6 protocol family Sep 4 17:39:35.111091 kernel: Segment Routing with IPv6 Sep 4 17:39:35.111108 kernel: In-situ OAM (IOAM) with IPv6 Sep 4 17:39:35.111125 kernel: NET: Registered PF_PACKET protocol family Sep 4 17:39:35.111142 kernel: Key type dns_resolver registered Sep 4 17:39:35.111159 kernel: IPI shorthand broadcast: enabled Sep 4 17:39:35.111174 kernel: sched_clock: Marking stable (813010258, 342352499)->(1287037398, -131674641) Sep 4 17:39:35.111191 kernel: registered taskstats version 1 Sep 4 17:39:35.111211 kernel: Loading compiled-in X.509 certificates Sep 4 17:39:35.111228 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.48-flatcar: 8669771ab5e11f458b79e6634fe685dacc266b18' Sep 4 17:39:35.111244 kernel: Key type .fscrypt registered Sep 4 17:39:35.111260 kernel: Key type fscrypt-provisioning registered Sep 4 17:39:35.111278 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 4 17:39:35.111294 kernel: ima: Allocated hash algorithm: sha1 Sep 4 17:39:35.111311 kernel: ima: No architecture policies found Sep 4 17:39:35.111327 kernel: clk: Disabling unused clocks Sep 4 17:39:35.111344 kernel: Freeing unused kernel image (initmem) memory: 42704K Sep 4 17:39:35.111364 kernel: Write protecting the kernel read-only data: 36864k Sep 4 17:39:35.111379 kernel: Freeing unused kernel image (rodata/data gap) memory: 1868K Sep 4 17:39:35.115552 kernel: Run /init as init process Sep 4 17:39:35.115577 kernel: with arguments: Sep 4 17:39:35.115595 kernel: /init Sep 4 17:39:35.115611 kernel: with environment: Sep 4 17:39:35.115719 kernel: HOME=/ Sep 4 17:39:35.115738 kernel: TERM=linux Sep 4 17:39:35.115755 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 4 17:39:35.115788 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 4 17:39:35.115824 systemd[1]: Detected virtualization amazon. Sep 4 17:39:35.115846 systemd[1]: Detected architecture x86-64. Sep 4 17:39:35.115862 systemd[1]: Running in initrd. Sep 4 17:39:35.115879 systemd[1]: No hostname configured, using default hostname. Sep 4 17:39:35.115900 systemd[1]: Hostname set to . Sep 4 17:39:35.115919 systemd[1]: Initializing machine ID from VM UUID. Sep 4 17:39:35.115937 systemd[1]: Queued start job for default target initrd.target. Sep 4 17:39:35.115955 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:39:35.115974 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:39:35.115998 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 4 17:39:35.116017 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 17:39:35.116036 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 4 17:39:35.116057 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 4 17:39:35.116078 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 4 17:39:35.116097 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 4 17:39:35.116115 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:39:35.116134 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:39:35.116152 systemd[1]: Reached target paths.target - Path Units. Sep 4 17:39:35.116170 systemd[1]: Reached target slices.target - Slice Units. Sep 4 17:39:35.116191 systemd[1]: Reached target swap.target - Swaps. Sep 4 17:39:35.116209 systemd[1]: Reached target timers.target - Timer Units. Sep 4 17:39:35.116269 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:39:35.122800 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:39:35.122857 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 4 17:39:35.122912 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 4 17:39:35.122932 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:39:35.122951 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 17:39:35.123014 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:39:35.123035 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 17:39:35.123054 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 4 17:39:35.123080 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 4 17:39:35.123102 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 17:39:35.123120 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 4 17:39:35.123140 systemd[1]: Starting systemd-fsck-usr.service... Sep 4 17:39:35.123162 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 17:39:35.123182 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 17:39:35.123201 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:39:35.123274 systemd-journald[178]: Collecting audit messages is disabled. Sep 4 17:39:35.123317 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 4 17:39:35.123339 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:39:35.123358 systemd[1]: Finished systemd-fsck-usr.service. Sep 4 17:39:35.123425 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 17:39:35.127827 systemd-journald[178]: Journal started Sep 4 17:39:35.127881 systemd-journald[178]: Runtime Journal (/run/log/journal/ec24767fab9b0437dcf06fc9a891202f) is 4.8M, max 38.6M, 33.7M free. Sep 4 17:39:35.098602 systemd-modules-load[179]: Inserted module 'overlay' Sep 4 17:39:35.136810 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 17:39:35.136676 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:39:35.162413 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 4 17:39:35.163540 systemd-modules-load[179]: Inserted module 'br_netfilter' Sep 4 17:39:35.281692 kernel: Bridge firewalling registered Sep 4 17:39:35.164434 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 17:39:35.295340 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 4 17:39:35.298351 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 17:39:35.298662 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:39:35.311048 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:39:35.322676 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:39:35.344097 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:39:35.347970 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:39:35.376447 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:39:35.392648 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 17:39:35.409506 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:39:35.419687 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 4 17:39:35.457089 dracut-cmdline[216]: dracut-dracut-053 Sep 4 17:39:35.460959 systemd-resolved[210]: Positive Trust Anchors: Sep 4 17:39:35.460977 systemd-resolved[210]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 17:39:35.461027 systemd-resolved[210]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 4 17:39:35.475103 dracut-cmdline[216]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=ceda2dd706627da8006bcd6ae77ea155b2a7de6732e2c1c7ab4bed271400663d Sep 4 17:39:35.486590 systemd-resolved[210]: Defaulting to hostname 'linux'. Sep 4 17:39:35.490192 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 17:39:35.492244 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:39:35.583419 kernel: SCSI subsystem initialized Sep 4 17:39:35.594410 kernel: Loading iSCSI transport class v2.0-870. Sep 4 17:39:35.606415 kernel: iscsi: registered transport (tcp) Sep 4 17:39:35.631411 kernel: iscsi: registered transport (qla4xxx) Sep 4 17:39:35.631490 kernel: QLogic iSCSI HBA Driver Sep 4 17:39:35.678420 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 4 17:39:35.683897 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 4 17:39:35.714239 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 4 17:39:35.714360 kernel: device-mapper: uevent: version 1.0.3 Sep 4 17:39:35.714396 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 4 17:39:35.759417 kernel: raid6: avx512x4 gen() 14292 MB/s Sep 4 17:39:35.776424 kernel: raid6: avx512x2 gen() 16470 MB/s Sep 4 17:39:35.793416 kernel: raid6: avx512x1 gen() 16732 MB/s Sep 4 17:39:35.810414 kernel: raid6: avx2x4 gen() 15949 MB/s Sep 4 17:39:35.828423 kernel: raid6: avx2x2 gen() 5705 MB/s Sep 4 17:39:35.845416 kernel: raid6: avx2x1 gen() 5496 MB/s Sep 4 17:39:35.845523 kernel: raid6: using algorithm avx512x1 gen() 16732 MB/s Sep 4 17:39:35.863412 kernel: raid6: .... xor() 7866 MB/s, rmw enabled Sep 4 17:39:35.863495 kernel: raid6: using avx512x2 recovery algorithm Sep 4 17:39:35.890410 kernel: xor: automatically using best checksumming function avx Sep 4 17:39:36.084408 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 4 17:39:36.097584 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:39:36.104653 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:39:36.136586 systemd-udevd[398]: Using default interface naming scheme 'v255'. Sep 4 17:39:36.142993 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:39:36.150579 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 4 17:39:36.175276 dracut-pre-trigger[402]: rd.md=0: removing MD RAID activation Sep 4 17:39:36.211709 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:39:36.218792 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 17:39:36.282111 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:39:36.294944 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 4 17:39:36.378238 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 4 17:39:36.384591 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:39:36.387684 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:39:36.390344 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 17:39:36.400714 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 4 17:39:36.449467 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:39:36.466670 kernel: cryptd: max_cpu_qlen set to 1000 Sep 4 17:39:36.477836 kernel: ena 0000:00:05.0: ENA device version: 0.10 Sep 4 17:39:36.478132 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Sep 4 17:39:36.484402 kernel: AVX2 version of gcm_enc/dec engaged. Sep 4 17:39:36.484474 kernel: AES CTR mode by8 optimization enabled Sep 4 17:39:36.487446 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Sep 4 17:39:36.500410 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem febf4000, mac addr 06:9d:69:14:47:df Sep 4 17:39:36.502761 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:39:36.504125 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:39:36.508737 (udev-worker)[448]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:39:36.510373 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:39:36.516783 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:39:36.517039 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:39:36.518659 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:39:36.542590 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:39:36.551789 kernel: nvme nvme0: pci function 0000:00:04.0 Sep 4 17:39:36.552077 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Sep 4 17:39:36.564490 kernel: nvme nvme0: 2/0/0 default/read/poll queues Sep 4 17:39:36.578423 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 4 17:39:36.578484 kernel: GPT:9289727 != 16777215 Sep 4 17:39:36.578503 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 4 17:39:36.578522 kernel: GPT:9289727 != 16777215 Sep 4 17:39:36.578538 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 4 17:39:36.578556 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 4 17:39:36.691099 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:39:36.697805 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:39:36.737045 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:39:36.782790 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 scanned by (udev-worker) (450) Sep 4 17:39:36.793123 kernel: BTRFS: device fsid 0dc40443-7f77-4fa7-b5e4-579d4bba0772 devid 1 transid 37 /dev/nvme0n1p3 scanned by (udev-worker) (458) Sep 4 17:39:36.820249 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Sep 4 17:39:36.880894 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Sep 4 17:39:36.901855 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Sep 4 17:39:36.902734 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Sep 4 17:39:36.933817 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Sep 4 17:39:36.945285 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 4 17:39:36.961884 disk-uuid[626]: Primary Header is updated. Sep 4 17:39:36.961884 disk-uuid[626]: Secondary Entries is updated. Sep 4 17:39:36.961884 disk-uuid[626]: Secondary Header is updated. Sep 4 17:39:36.972406 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 4 17:39:36.980111 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 4 17:39:36.988670 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 4 17:39:38.000415 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 4 17:39:38.010182 disk-uuid[627]: The operation has completed successfully. Sep 4 17:39:38.206510 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 4 17:39:38.206633 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 4 17:39:38.245774 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 4 17:39:38.265197 sh[970]: Success Sep 4 17:39:38.280406 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 4 17:39:38.426063 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 4 17:39:38.437521 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 4 17:39:38.444103 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 4 17:39:38.490255 kernel: BTRFS info (device dm-0): first mount of filesystem 0dc40443-7f77-4fa7-b5e4-579d4bba0772 Sep 4 17:39:38.490329 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:39:38.490360 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 4 17:39:38.491637 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 4 17:39:38.491668 kernel: BTRFS info (device dm-0): using free space tree Sep 4 17:39:38.580416 kernel: BTRFS info (device dm-0): enabling ssd optimizations Sep 4 17:39:38.609048 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 4 17:39:38.612214 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 4 17:39:38.623607 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 4 17:39:38.629244 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 4 17:39:38.657201 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:39:38.657269 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:39:38.657289 kernel: BTRFS info (device nvme0n1p6): using free space tree Sep 4 17:39:38.664417 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 4 17:39:38.679655 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:39:38.679825 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 4 17:39:38.687773 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 4 17:39:38.700733 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 4 17:39:38.767458 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:39:38.784603 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 17:39:38.868213 systemd-networkd[1162]: lo: Link UP Sep 4 17:39:38.868225 systemd-networkd[1162]: lo: Gained carrier Sep 4 17:39:38.875485 systemd-networkd[1162]: Enumeration completed Sep 4 17:39:38.875950 systemd-networkd[1162]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:39:38.875955 systemd-networkd[1162]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:39:38.877850 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 17:39:38.884627 systemd[1]: Reached target network.target - Network. Sep 4 17:39:38.891464 systemd-networkd[1162]: eth0: Link UP Sep 4 17:39:38.891470 systemd-networkd[1162]: eth0: Gained carrier Sep 4 17:39:38.891486 systemd-networkd[1162]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:39:38.904482 systemd-networkd[1162]: eth0: DHCPv4 address 172.31.20.143/20, gateway 172.31.16.1 acquired from 172.31.16.1 Sep 4 17:39:39.137280 ignition[1087]: Ignition 2.19.0 Sep 4 17:39:39.137295 ignition[1087]: Stage: fetch-offline Sep 4 17:39:39.137589 ignition[1087]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:39:39.137602 ignition[1087]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 4 17:39:39.140855 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:39:39.137980 ignition[1087]: Ignition finished successfully Sep 4 17:39:39.160340 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Sep 4 17:39:39.206223 ignition[1171]: Ignition 2.19.0 Sep 4 17:39:39.206239 ignition[1171]: Stage: fetch Sep 4 17:39:39.207309 ignition[1171]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:39:39.207429 ignition[1171]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 4 17:39:39.207645 ignition[1171]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 4 17:39:39.216440 ignition[1171]: PUT result: OK Sep 4 17:39:39.219528 ignition[1171]: parsed url from cmdline: "" Sep 4 17:39:39.219540 ignition[1171]: no config URL provided Sep 4 17:39:39.219550 ignition[1171]: reading system config file "/usr/lib/ignition/user.ign" Sep 4 17:39:39.219564 ignition[1171]: no config at "/usr/lib/ignition/user.ign" Sep 4 17:39:39.219589 ignition[1171]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 4 17:39:39.236290 ignition[1171]: PUT result: OK Sep 4 17:39:39.236430 ignition[1171]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Sep 4 17:39:39.244826 ignition[1171]: GET result: OK Sep 4 17:39:39.246467 ignition[1171]: parsing config with SHA512: 4afc0433494e0271c4444cb6e347b69439cd31852671b42ad796ab28819284508ff27d82c191c1eca7ab2d33a1cc34247983cbee634e763c22e2f5d7e7ffafed Sep 4 17:39:39.260265 unknown[1171]: fetched base config from "system" Sep 4 17:39:39.260280 unknown[1171]: fetched base config from "system" Sep 4 17:39:39.260288 unknown[1171]: fetched user config from "aws" Sep 4 17:39:39.262511 ignition[1171]: fetch: fetch complete Sep 4 17:39:39.262517 ignition[1171]: fetch: fetch passed Sep 4 17:39:39.262771 ignition[1171]: Ignition finished successfully Sep 4 17:39:39.268499 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Sep 4 17:39:39.276791 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 4 17:39:39.303050 ignition[1177]: Ignition 2.19.0 Sep 4 17:39:39.303064 ignition[1177]: Stage: kargs Sep 4 17:39:39.303571 ignition[1177]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:39:39.303586 ignition[1177]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 4 17:39:39.303755 ignition[1177]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 4 17:39:39.306908 ignition[1177]: PUT result: OK Sep 4 17:39:39.314594 ignition[1177]: kargs: kargs passed Sep 4 17:39:39.314853 ignition[1177]: Ignition finished successfully Sep 4 17:39:39.318939 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 4 17:39:39.329345 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 4 17:39:39.356567 ignition[1183]: Ignition 2.19.0 Sep 4 17:39:39.356580 ignition[1183]: Stage: disks Sep 4 17:39:39.357946 ignition[1183]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:39:39.357962 ignition[1183]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 4 17:39:39.358082 ignition[1183]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 4 17:39:39.361633 ignition[1183]: PUT result: OK Sep 4 17:39:39.366536 ignition[1183]: disks: disks passed Sep 4 17:39:39.366603 ignition[1183]: Ignition finished successfully Sep 4 17:39:39.369038 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 4 17:39:39.376456 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 4 17:39:39.379770 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 4 17:39:39.382624 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 17:39:39.389219 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 17:39:39.393767 systemd[1]: Reached target basic.target - Basic System. Sep 4 17:39:39.411708 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 4 17:39:39.485754 systemd-fsck[1191]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 4 17:39:39.490071 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 4 17:39:39.497540 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 4 17:39:39.648407 kernel: EXT4-fs (nvme0n1p9): mounted filesystem bdbe0f61-2675-40b7-b9ae-5653402e9b23 r/w with ordered data mode. Quota mode: none. Sep 4 17:39:39.649129 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 4 17:39:39.650231 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 4 17:39:39.671024 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:39:39.674692 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 4 17:39:39.677548 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 4 17:39:39.677733 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 4 17:39:39.677767 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:39:39.695806 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/nvme0n1p6 scanned by mount (1210) Sep 4 17:39:39.698337 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:39:39.698407 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:39:39.698427 kernel: BTRFS info (device nvme0n1p6): using free space tree Sep 4 17:39:39.698318 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 4 17:39:39.704603 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 4 17:39:39.718414 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 4 17:39:39.720280 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:39:40.040320 initrd-setup-root[1234]: cut: /sysroot/etc/passwd: No such file or directory Sep 4 17:39:40.071014 initrd-setup-root[1241]: cut: /sysroot/etc/group: No such file or directory Sep 4 17:39:40.081689 initrd-setup-root[1248]: cut: /sysroot/etc/shadow: No such file or directory Sep 4 17:39:40.090565 initrd-setup-root[1255]: cut: /sysroot/etc/gshadow: No such file or directory Sep 4 17:39:40.183532 systemd-networkd[1162]: eth0: Gained IPv6LL Sep 4 17:39:40.443694 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 4 17:39:40.451551 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 4 17:39:40.458489 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 4 17:39:40.467096 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:39:40.466646 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 4 17:39:40.516359 ignition[1323]: INFO : Ignition 2.19.0 Sep 4 17:39:40.520584 ignition[1323]: INFO : Stage: mount Sep 4 17:39:40.520584 ignition[1323]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:39:40.520584 ignition[1323]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 4 17:39:40.520584 ignition[1323]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 4 17:39:40.526519 ignition[1323]: INFO : PUT result: OK Sep 4 17:39:40.533374 ignition[1323]: INFO : mount: mount passed Sep 4 17:39:40.541375 ignition[1323]: INFO : Ignition finished successfully Sep 4 17:39:40.540216 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 4 17:39:40.542078 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 4 17:39:40.554032 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 4 17:39:40.655876 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:39:40.701417 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by mount (1334) Sep 4 17:39:40.703409 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem b2463ce1-c756-4e78-b7f2-401dad24571d Sep 4 17:39:40.703468 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 4 17:39:40.706715 kernel: BTRFS info (device nvme0n1p6): using free space tree Sep 4 17:39:40.713412 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 4 17:39:40.715714 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:39:40.747289 ignition[1351]: INFO : Ignition 2.19.0 Sep 4 17:39:40.747289 ignition[1351]: INFO : Stage: files Sep 4 17:39:40.754460 ignition[1351]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:39:40.754460 ignition[1351]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 4 17:39:40.754460 ignition[1351]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 4 17:39:40.758483 ignition[1351]: INFO : PUT result: OK Sep 4 17:39:40.762062 ignition[1351]: DEBUG : files: compiled without relabeling support, skipping Sep 4 17:39:40.763862 ignition[1351]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 4 17:39:40.763862 ignition[1351]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 4 17:39:40.787242 ignition[1351]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 4 17:39:40.789440 ignition[1351]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 4 17:39:40.793492 ignition[1351]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 4 17:39:40.789857 unknown[1351]: wrote ssh authorized keys file for user: core Sep 4 17:39:40.807787 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Sep 4 17:39:40.810522 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Sep 4 17:39:40.810522 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 4 17:39:40.815843 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Sep 4 17:39:40.868258 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Sep 4 17:39:40.959701 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 4 17:39:40.959701 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 4 17:39:40.964434 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 4 17:39:41.283439 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Sep 4 17:39:41.411426 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Sep 4 17:39:41.414033 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.28.7-x86-64.raw: attempt #1 Sep 4 17:39:41.804171 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(c): GET result: OK Sep 4 17:39:42.185190 ignition[1351]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Sep 4 17:39:42.185190 ignition[1351]: INFO : files: op(d): [started] processing unit "containerd.service" Sep 4 17:39:42.193241 ignition[1351]: INFO : files: op(d): op(e): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(d): [finished] processing unit "containerd.service" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(f): [started] processing unit "prepare-helm.service" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(f): op(10): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(f): [finished] processing unit "prepare-helm.service" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:39:42.204618 ignition[1351]: INFO : files: files passed Sep 4 17:39:42.204618 ignition[1351]: INFO : Ignition finished successfully Sep 4 17:39:42.222934 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 4 17:39:42.262958 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 4 17:39:42.270517 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 4 17:39:42.274254 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 4 17:39:42.274591 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 4 17:39:42.321740 initrd-setup-root-after-ignition[1380]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:39:42.321740 initrd-setup-root-after-ignition[1380]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:39:42.332311 initrd-setup-root-after-ignition[1384]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:39:42.335184 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:39:42.339884 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 4 17:39:42.351795 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 4 17:39:42.429610 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 4 17:39:42.429753 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 4 17:39:42.432611 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 4 17:39:42.438806 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 4 17:39:42.440148 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 4 17:39:42.446559 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 4 17:39:42.473783 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:39:42.483765 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 4 17:39:42.504870 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:39:42.505122 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:39:42.511277 systemd[1]: Stopped target timers.target - Timer Units. Sep 4 17:39:42.513177 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 4 17:39:42.513375 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:39:42.516037 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 4 17:39:42.517339 systemd[1]: Stopped target basic.target - Basic System. Sep 4 17:39:42.524088 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 4 17:39:42.531358 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:39:42.538355 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 4 17:39:42.542163 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 4 17:39:42.545634 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:39:42.546127 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 4 17:39:42.556367 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 4 17:39:42.565691 systemd[1]: Stopped target swap.target - Swaps. Sep 4 17:39:42.568358 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 4 17:39:42.568504 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:39:42.571855 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:39:42.574404 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:39:42.578224 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 4 17:39:42.578310 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:39:42.580836 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 4 17:39:42.580970 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 4 17:39:42.586643 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 4 17:39:42.586810 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:39:42.595002 systemd[1]: ignition-files.service: Deactivated successfully. Sep 4 17:39:42.595304 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 4 17:39:42.610775 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 4 17:39:42.616046 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 4 17:39:42.619480 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 4 17:39:42.619654 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:39:42.625166 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 4 17:39:42.625358 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:39:42.642160 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 4 17:39:42.642465 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 4 17:39:42.671409 ignition[1404]: INFO : Ignition 2.19.0 Sep 4 17:39:42.672829 ignition[1404]: INFO : Stage: umount Sep 4 17:39:42.672829 ignition[1404]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:39:42.672829 ignition[1404]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 4 17:39:42.677794 ignition[1404]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 4 17:39:42.677794 ignition[1404]: INFO : PUT result: OK Sep 4 17:39:42.688226 ignition[1404]: INFO : umount: umount passed Sep 4 17:39:42.688226 ignition[1404]: INFO : Ignition finished successfully Sep 4 17:39:42.687617 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 4 17:39:42.690775 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 4 17:39:42.690952 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 4 17:39:42.693323 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 4 17:39:42.693619 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 4 17:39:42.696072 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 4 17:39:42.696149 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 4 17:39:42.698172 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 4 17:39:42.698239 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Sep 4 17:39:42.701436 systemd[1]: Stopped target network.target - Network. Sep 4 17:39:42.703453 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 4 17:39:42.703619 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:39:42.706690 systemd[1]: Stopped target paths.target - Path Units. Sep 4 17:39:42.707920 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 4 17:39:42.710840 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:39:42.713467 systemd[1]: Stopped target slices.target - Slice Units. Sep 4 17:39:42.715565 systemd[1]: Stopped target sockets.target - Socket Units. Sep 4 17:39:42.727736 systemd[1]: iscsid.socket: Deactivated successfully. Sep 4 17:39:42.727835 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:39:42.737551 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 4 17:39:42.742658 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:39:42.748411 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 4 17:39:42.753132 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 4 17:39:42.760526 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 4 17:39:42.767131 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 4 17:39:42.772515 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 4 17:39:42.776014 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 4 17:39:42.778936 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 4 17:39:42.779061 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 4 17:39:42.792378 systemd-networkd[1162]: eth0: DHCPv6 lease lost Sep 4 17:39:42.797412 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 4 17:39:42.797590 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 4 17:39:42.802661 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 4 17:39:42.802901 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 4 17:39:42.809659 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 4 17:39:42.809740 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:39:42.814199 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 4 17:39:42.814999 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 4 17:39:42.836999 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 4 17:39:42.840371 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 4 17:39:42.840495 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:39:42.842404 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 4 17:39:42.842592 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:39:42.844926 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 4 17:39:42.844999 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 4 17:39:42.848476 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 4 17:39:42.848727 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:39:42.854541 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:39:42.890355 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 4 17:39:42.890736 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:39:42.896112 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 4 17:39:42.896207 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 4 17:39:42.898634 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 4 17:39:42.898688 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:39:42.901157 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 4 17:39:42.901218 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:39:42.908763 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 4 17:39:42.908910 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 4 17:39:42.911473 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:39:42.911538 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:39:42.931645 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 4 17:39:42.939943 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 4 17:39:42.940053 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:39:42.947717 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 4 17:39:42.947821 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:39:42.952590 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 4 17:39:42.952852 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:39:42.969728 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:39:42.969807 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:39:42.978816 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 4 17:39:42.978976 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 4 17:39:42.986908 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 4 17:39:42.989353 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 4 17:39:43.005415 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 4 17:39:43.012734 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 4 17:39:43.068724 systemd[1]: Switching root. Sep 4 17:39:43.102023 systemd-journald[178]: Journal stopped Sep 4 17:39:45.709373 systemd-journald[178]: Received SIGTERM from PID 1 (systemd). Sep 4 17:39:45.709511 kernel: SELinux: policy capability network_peer_controls=1 Sep 4 17:39:45.709538 kernel: SELinux: policy capability open_perms=1 Sep 4 17:39:45.709560 kernel: SELinux: policy capability extended_socket_class=1 Sep 4 17:39:45.709588 kernel: SELinux: policy capability always_check_network=0 Sep 4 17:39:45.709609 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 4 17:39:45.709630 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 4 17:39:45.709651 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 4 17:39:45.709672 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 4 17:39:45.709693 kernel: audit: type=1403 audit(1725471584.229:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 4 17:39:45.709724 systemd[1]: Successfully loaded SELinux policy in 54.211ms. Sep 4 17:39:45.709759 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 13.488ms. Sep 4 17:39:45.709783 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 4 17:39:45.709809 systemd[1]: Detected virtualization amazon. Sep 4 17:39:45.709828 systemd[1]: Detected architecture x86-64. Sep 4 17:39:45.709847 systemd[1]: Detected first boot. Sep 4 17:39:45.709866 systemd[1]: Initializing machine ID from VM UUID. Sep 4 17:39:45.709886 zram_generator::config[1462]: No configuration found. Sep 4 17:39:45.709908 systemd[1]: Populated /etc with preset unit settings. Sep 4 17:39:45.710320 systemd[1]: Queued start job for default target multi-user.target. Sep 4 17:39:45.710414 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Sep 4 17:39:45.710446 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 4 17:39:45.710467 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 4 17:39:45.710487 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 4 17:39:45.710509 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 4 17:39:45.710530 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 4 17:39:45.710554 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 4 17:39:45.710576 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 4 17:39:45.710601 systemd[1]: Created slice user.slice - User and Session Slice. Sep 4 17:39:45.710624 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:39:45.710652 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:39:45.710675 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 4 17:39:45.710697 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 4 17:39:45.710721 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 4 17:39:45.710742 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 17:39:45.710763 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 4 17:39:45.712696 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:39:45.712756 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 4 17:39:45.712780 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:39:45.712814 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 17:39:45.712833 systemd[1]: Reached target slices.target - Slice Units. Sep 4 17:39:45.712855 systemd[1]: Reached target swap.target - Swaps. Sep 4 17:39:45.712874 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 4 17:39:45.712897 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 4 17:39:45.712917 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 4 17:39:45.712937 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 4 17:39:45.712957 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:39:45.712979 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 17:39:45.713000 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:39:45.713021 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 4 17:39:45.713041 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 4 17:39:45.713063 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 4 17:39:45.713103 systemd[1]: Mounting media.mount - External Media Directory... Sep 4 17:39:45.713126 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:39:45.713148 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 4 17:39:45.713171 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 4 17:39:45.713196 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 4 17:39:45.713225 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 4 17:39:45.713249 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:39:45.713270 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 17:39:45.713294 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 4 17:39:45.713318 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:39:45.713341 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 17:39:45.713365 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:39:45.713407 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 4 17:39:45.713425 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:39:45.713444 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 4 17:39:45.713473 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Sep 4 17:39:45.713492 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Sep 4 17:39:45.713509 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 17:39:45.713529 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 17:39:45.713546 kernel: fuse: init (API version 7.39) Sep 4 17:39:45.713564 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 4 17:39:45.713586 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 4 17:39:45.713604 kernel: loop: module loaded Sep 4 17:39:45.713625 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 17:39:45.713651 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:39:45.713725 systemd-journald[1562]: Collecting audit messages is disabled. Sep 4 17:39:45.713765 systemd-journald[1562]: Journal started Sep 4 17:39:45.713809 systemd-journald[1562]: Runtime Journal (/run/log/journal/ec24767fab9b0437dcf06fc9a891202f) is 4.8M, max 38.6M, 33.7M free. Sep 4 17:39:45.718214 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 4 17:39:45.723414 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 17:39:45.726161 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 4 17:39:45.731837 systemd[1]: Mounted media.mount - External Media Directory. Sep 4 17:39:45.734102 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 4 17:39:45.736453 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 4 17:39:45.738081 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 4 17:39:45.739985 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:39:45.743814 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 4 17:39:45.744051 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 4 17:39:45.748560 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:39:45.748989 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:39:45.753417 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:39:45.753652 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:39:45.755593 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 4 17:39:45.755828 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 4 17:39:45.763630 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:39:45.763873 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:39:45.766585 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 4 17:39:45.770789 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 4 17:39:45.783111 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 4 17:39:45.798569 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 4 17:39:45.811535 kernel: ACPI: bus type drm_connector registered Sep 4 17:39:45.825524 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 4 17:39:45.826887 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 4 17:39:45.842770 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 4 17:39:45.853359 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 4 17:39:45.856136 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 17:39:45.865664 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 4 17:39:45.867611 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 17:39:45.890733 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 17:39:45.899118 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 4 17:39:45.901300 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 17:39:45.903748 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 17:39:45.907086 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 17:39:45.909419 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 4 17:39:45.910903 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 4 17:39:45.926514 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 4 17:39:45.935084 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 4 17:39:45.941917 systemd-journald[1562]: Time spent on flushing to /var/log/journal/ec24767fab9b0437dcf06fc9a891202f is 84.506ms for 954 entries. Sep 4 17:39:45.941917 systemd-journald[1562]: System Journal (/var/log/journal/ec24767fab9b0437dcf06fc9a891202f) is 8.0M, max 195.6M, 187.6M free. Sep 4 17:39:46.052966 systemd-journald[1562]: Received client request to flush runtime journal. Sep 4 17:39:45.950430 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:39:45.960436 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:39:45.969698 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 4 17:39:46.020787 udevadm[1622]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Sep 4 17:39:46.047625 systemd-tmpfiles[1606]: ACLs are not supported, ignoring. Sep 4 17:39:46.047649 systemd-tmpfiles[1606]: ACLs are not supported, ignoring. Sep 4 17:39:46.052682 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:39:46.058032 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 4 17:39:46.069789 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:39:46.088724 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 4 17:39:46.158719 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 4 17:39:46.171176 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 17:39:46.209301 systemd-tmpfiles[1636]: ACLs are not supported, ignoring. Sep 4 17:39:46.209332 systemd-tmpfiles[1636]: ACLs are not supported, ignoring. Sep 4 17:39:46.221676 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:39:47.053302 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 4 17:39:47.069686 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:39:47.103189 systemd-udevd[1642]: Using default interface naming scheme 'v255'. Sep 4 17:39:47.157194 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:39:47.171634 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 17:39:47.211471 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 4 17:39:47.235006 (udev-worker)[1655]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:39:47.284419 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1643) Sep 4 17:39:47.316414 kernel: BTRFS info: devid 1 device path /dev/dm-0 changed to /dev/mapper/usr scanned by (udev-worker) (1643) Sep 4 17:39:47.340633 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 4 17:39:47.382184 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Sep 4 17:39:47.452419 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0xb100, revision 255 Sep 4 17:39:47.456435 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input4 Sep 4 17:39:47.460413 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input3 Sep 4 17:39:47.503605 kernel: ACPI: button: Power Button [PWRF] Sep 4 17:39:47.503683 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input5 Sep 4 17:39:47.503723 kernel: ACPI: button: Sleep Button [SLPF] Sep 4 17:39:47.553250 systemd-networkd[1648]: lo: Link UP Sep 4 17:39:47.555074 systemd-networkd[1648]: lo: Gained carrier Sep 4 17:39:47.558114 systemd-networkd[1648]: Enumeration completed Sep 4 17:39:47.558292 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 17:39:47.560972 systemd-networkd[1648]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:39:47.567679 systemd-networkd[1648]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:39:47.571156 systemd-networkd[1648]: eth0: Link UP Sep 4 17:39:47.572295 systemd-networkd[1648]: eth0: Gained carrier Sep 4 17:39:47.572438 systemd-networkd[1648]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:39:47.573689 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 4 17:39:47.579431 kernel: mousedev: PS/2 mouse device common for all mice Sep 4 17:39:47.581473 systemd-networkd[1648]: eth0: DHCPv4 address 172.31.20.143/20, gateway 172.31.16.1 acquired from 172.31.16.1 Sep 4 17:39:47.601502 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:39:47.625411 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (1651) Sep 4 17:39:47.915732 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Sep 4 17:39:47.974317 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 4 17:39:47.977490 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:39:47.986731 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 4 17:39:48.015233 lvm[1766]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 4 17:39:48.058452 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 4 17:39:48.060636 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:39:48.069924 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 4 17:39:48.080580 lvm[1769]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 4 17:39:48.116023 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 4 17:39:48.118150 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 4 17:39:48.120325 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 4 17:39:48.120820 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 17:39:48.122316 systemd[1]: Reached target machines.target - Containers. Sep 4 17:39:48.126125 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 4 17:39:48.133619 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 4 17:39:48.137624 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 4 17:39:48.139809 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:39:48.152627 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 4 17:39:48.166732 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 4 17:39:48.191796 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 4 17:39:48.215296 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 4 17:39:48.252409 kernel: loop0: detected capacity change from 0 to 209816 Sep 4 17:39:48.266978 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 4 17:39:48.292110 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 4 17:39:48.294068 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 4 17:39:48.395423 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 4 17:39:48.437780 kernel: loop1: detected capacity change from 0 to 89336 Sep 4 17:39:48.553550 kernel: loop2: detected capacity change from 0 to 61336 Sep 4 17:39:48.612414 kernel: loop3: detected capacity change from 0 to 140728 Sep 4 17:39:48.755416 kernel: loop4: detected capacity change from 0 to 209816 Sep 4 17:39:48.802633 kernel: loop5: detected capacity change from 0 to 89336 Sep 4 17:39:48.866629 kernel: loop6: detected capacity change from 0 to 61336 Sep 4 17:39:48.897423 kernel: loop7: detected capacity change from 0 to 140728 Sep 4 17:39:48.961571 (sd-merge)[1790]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Sep 4 17:39:48.962472 (sd-merge)[1790]: Merged extensions into '/usr'. Sep 4 17:39:48.979525 systemd[1]: Reloading requested from client PID 1777 ('systemd-sysext') (unit systemd-sysext.service)... Sep 4 17:39:48.979546 systemd[1]: Reloading... Sep 4 17:39:49.136539 zram_generator::config[1819]: No configuration found. Sep 4 17:39:49.207641 systemd-networkd[1648]: eth0: Gained IPv6LL Sep 4 17:39:49.366095 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:39:49.494832 systemd[1]: Reloading finished in 514 ms. Sep 4 17:39:49.515250 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 4 17:39:49.518163 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 4 17:39:49.529788 systemd[1]: Starting ensure-sysext.service... Sep 4 17:39:49.542116 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 4 17:39:49.564540 systemd[1]: Reloading requested from client PID 1872 ('systemctl') (unit ensure-sysext.service)... Sep 4 17:39:49.564566 systemd[1]: Reloading... Sep 4 17:39:49.582298 systemd-tmpfiles[1873]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 4 17:39:49.583949 systemd-tmpfiles[1873]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 4 17:39:49.587713 systemd-tmpfiles[1873]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 4 17:39:49.588254 systemd-tmpfiles[1873]: ACLs are not supported, ignoring. Sep 4 17:39:49.588358 systemd-tmpfiles[1873]: ACLs are not supported, ignoring. Sep 4 17:39:49.600441 systemd-tmpfiles[1873]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 17:39:49.600455 systemd-tmpfiles[1873]: Skipping /boot Sep 4 17:39:49.617995 systemd-tmpfiles[1873]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 17:39:49.618009 systemd-tmpfiles[1873]: Skipping /boot Sep 4 17:39:49.731418 zram_generator::config[1904]: No configuration found. Sep 4 17:39:49.951114 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:39:50.053451 ldconfig[1773]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 4 17:39:50.099334 systemd[1]: Reloading finished in 533 ms. Sep 4 17:39:50.129395 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 4 17:39:50.159437 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 4 17:39:50.191678 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 4 17:39:50.207139 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 4 17:39:50.222843 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 4 17:39:50.231709 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 17:39:50.246652 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 4 17:39:50.271424 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:39:50.271749 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:39:50.282531 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:39:50.301087 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:39:50.317982 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:39:50.321209 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:39:50.321442 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:39:50.328962 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:39:50.329215 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:39:50.355683 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:39:50.356674 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:39:50.369028 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:39:50.371432 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:39:50.400897 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:39:50.402567 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:39:50.409897 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:39:50.430200 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 17:39:50.454600 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:39:50.471470 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:39:50.474725 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:39:50.475249 systemd[1]: Reached target time-set.target - System Time Set. Sep 4 17:39:50.480978 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 4 17:39:50.492796 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 4 17:39:50.498668 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 4 17:39:50.506425 augenrules[1999]: No rules Sep 4 17:39:50.506339 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:39:50.506773 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:39:50.513728 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 4 17:39:50.519555 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 17:39:50.519857 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 17:39:50.523040 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:39:50.523799 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:39:50.526972 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:39:50.527222 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:39:50.548222 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 17:39:50.548918 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 17:39:50.557760 systemd-resolved[1963]: Positive Trust Anchors: Sep 4 17:39:50.557784 systemd-resolved[1963]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 17:39:50.557839 systemd-resolved[1963]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 4 17:39:50.561151 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 4 17:39:50.565902 systemd[1]: Finished ensure-sysext.service. Sep 4 17:39:50.574636 systemd-resolved[1963]: Defaulting to hostname 'linux'. Sep 4 17:39:50.582638 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 17:39:50.585077 systemd[1]: Reached target network.target - Network. Sep 4 17:39:50.587222 systemd[1]: Reached target network-online.target - Network is Online. Sep 4 17:39:50.588834 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:39:50.594697 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 4 17:39:50.597368 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 4 17:39:50.601475 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 4 17:39:50.604862 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 17:39:50.607872 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 4 17:39:50.609595 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 4 17:39:50.611543 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 4 17:39:50.613231 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 4 17:39:50.614676 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 4 17:39:50.618401 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 4 17:39:50.618451 systemd[1]: Reached target paths.target - Path Units. Sep 4 17:39:50.619685 systemd[1]: Reached target timers.target - Timer Units. Sep 4 17:39:50.626466 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 4 17:39:50.634814 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 4 17:39:50.638691 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 4 17:39:50.641584 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 4 17:39:50.642892 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 17:39:50.643959 systemd[1]: Reached target basic.target - Basic System. Sep 4 17:39:50.646596 systemd[1]: System is tainted: cgroupsv1 Sep 4 17:39:50.646723 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 4 17:39:50.646748 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 4 17:39:50.648072 systemd[1]: Starting containerd.service - containerd container runtime... Sep 4 17:39:50.657809 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Sep 4 17:39:50.665467 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 4 17:39:50.668500 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 4 17:39:50.687611 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 4 17:39:50.690424 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 4 17:39:50.695729 jq[2027]: false Sep 4 17:39:50.697879 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:39:50.714964 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 4 17:39:50.732610 systemd[1]: Started ntpd.service - Network Time Service. Sep 4 17:39:50.737601 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 4 17:39:50.745564 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 4 17:39:50.760561 systemd[1]: Starting setup-oem.service - Setup OEM... Sep 4 17:39:50.784655 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 4 17:39:50.802731 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 4 17:39:50.826546 extend-filesystems[2028]: Found loop4 Sep 4 17:39:50.825039 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 4 17:39:50.844787 extend-filesystems[2028]: Found loop5 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found loop6 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found loop7 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1p1 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1p2 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1p3 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found usr Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1p4 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1p6 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1p7 Sep 4 17:39:50.844787 extend-filesystems[2028]: Found nvme0n1p9 Sep 4 17:39:50.844787 extend-filesystems[2028]: Checking size of /dev/nvme0n1p9 Sep 4 17:39:50.828105 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 4 17:39:50.884790 systemd[1]: Starting update-engine.service - Update Engine... Sep 4 17:39:50.895363 dbus-daemon[2025]: [system] SELinux support is enabled Sep 4 17:39:50.924185 dbus-daemon[2025]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1648 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Sep 4 17:39:50.932962 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 4 17:39:50.942926 extend-filesystems[2028]: Resized partition /dev/nvme0n1p9 Sep 4 17:39:50.940229 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 4 17:39:50.980421 extend-filesystems[2064]: resize2fs 1.47.1 (20-May-2024) Sep 4 17:39:50.984499 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Sep 4 17:39:50.990882 ntpd[2033]: ntpd 4.2.8p17@1.4004-o Wed Sep 4 15:17:38 UTC 2024 (1): Starting Sep 4 17:39:50.991917 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: ntpd 4.2.8p17@1.4004-o Wed Sep 4 15:17:38 UTC 2024 (1): Starting Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: ---------------------------------------------------- Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: ntp-4 is maintained by Network Time Foundation, Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: corporation. Support and training for ntp-4 are Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: available at https://www.nwtime.org/support Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: ---------------------------------------------------- Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: proto: precision = 0.065 usec (-24) Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: basedate set to 2024-08-23 Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:50 ntpd[2033]: gps base set to 2024-08-25 (week 2329) Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: Listen and drop on 0 v6wildcard [::]:123 Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: Listen normally on 2 lo 127.0.0.1:123 Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: Listen normally on 3 eth0 172.31.20.143:123 Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: Listen normally on 4 lo [::1]:123 Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: Listen normally on 5 eth0 [fe80::49d:69ff:fe14:47df%2]:123 Sep 4 17:39:51.016097 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: Listening on routing socket on fd #22 for interface updates Sep 4 17:39:50.990922 ntpd[2033]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Sep 4 17:39:50.992405 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 4 17:39:51.061600 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 4 17:39:51.061600 ntpd[2033]: 4 Sep 17:39:51 ntpd[2033]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 4 17:39:50.990934 ntpd[2033]: ---------------------------------------------------- Sep 4 17:39:51.052131 systemd[1]: motdgen.service: Deactivated successfully. Sep 4 17:39:50.990944 ntpd[2033]: ntp-4 is maintained by Network Time Foundation, Sep 4 17:39:51.053756 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 4 17:39:50.990954 ntpd[2033]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Sep 4 17:39:50.990963 ntpd[2033]: corporation. Support and training for ntp-4 are Sep 4 17:39:50.990972 ntpd[2033]: available at https://www.nwtime.org/support Sep 4 17:39:50.990981 ntpd[2033]: ---------------------------------------------------- Sep 4 17:39:50.995092 ntpd[2033]: proto: precision = 0.065 usec (-24) Sep 4 17:39:50.999053 ntpd[2033]: basedate set to 2024-08-23 Sep 4 17:39:50.999077 ntpd[2033]: gps base set to 2024-08-25 (week 2329) Sep 4 17:39:51.001794 ntpd[2033]: Listen and drop on 0 v6wildcard [::]:123 Sep 4 17:39:51.001858 ntpd[2033]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Sep 4 17:39:51.002078 ntpd[2033]: Listen normally on 2 lo 127.0.0.1:123 Sep 4 17:39:51.002130 ntpd[2033]: Listen normally on 3 eth0 172.31.20.143:123 Sep 4 17:39:51.002177 ntpd[2033]: Listen normally on 4 lo [::1]:123 Sep 4 17:39:51.002261 ntpd[2033]: Listen normally on 5 eth0 [fe80::49d:69ff:fe14:47df%2]:123 Sep 4 17:39:51.002303 ntpd[2033]: Listening on routing socket on fd #22 for interface updates Sep 4 17:39:51.016670 ntpd[2033]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 4 17:39:51.016704 ntpd[2033]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 4 17:39:51.065263 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 4 17:39:51.069440 update_engine[2052]: I0904 17:39:51.069341 2052 main.cc:92] Flatcar Update Engine starting Sep 4 17:39:51.080969 update_engine[2052]: I0904 17:39:51.074131 2052 update_check_scheduler.cc:74] Next update check in 4m36s Sep 4 17:39:51.077937 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 4 17:39:51.078330 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 4 17:39:51.131657 jq[2061]: true Sep 4 17:39:51.138268 (ntainerd)[2079]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 4 17:39:51.191503 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 4 17:39:51.196802 coreos-metadata[2024]: Sep 04 17:39:51.196 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Sep 4 17:39:51.191562 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 4 17:39:51.193145 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 4 17:39:51.193172 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 4 17:39:51.211620 tar[2075]: linux-amd64/helm Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.198 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.199 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.199 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.200 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.200 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.201 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.202 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.202 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.202 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.203 INFO Fetch failed with 404: resource not found Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.203 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.204 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.204 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.205 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.205 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.206 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.206 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.208 INFO Fetch successful Sep 4 17:39:51.212022 coreos-metadata[2024]: Sep 04 17:39:51.208 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Sep 4 17:39:51.245792 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Sep 4 17:39:51.219276 dbus-daemon[2025]: [system] Successfully activated service 'org.freedesktop.systemd1' Sep 4 17:39:51.246367 coreos-metadata[2024]: Sep 04 17:39:51.231 INFO Fetch successful Sep 4 17:39:51.246447 jq[2087]: true Sep 4 17:39:51.268710 extend-filesystems[2064]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Sep 4 17:39:51.268710 extend-filesystems[2064]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 4 17:39:51.268710 extend-filesystems[2064]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Sep 4 17:39:51.281488 extend-filesystems[2028]: Resized filesystem in /dev/nvme0n1p9 Sep 4 17:39:51.284477 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 4 17:39:51.293157 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 4 17:39:51.313907 systemd[1]: Finished setup-oem.service - Setup OEM. Sep 4 17:39:51.324032 systemd[1]: Started update-engine.service - Update Engine. Sep 4 17:39:51.339609 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Sep 4 17:39:51.346742 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Sep 4 17:39:51.349064 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 4 17:39:51.363434 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (2114) Sep 4 17:39:51.370149 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 4 17:39:51.438977 bash[2148]: Updated "/home/core/.ssh/authorized_keys" Sep 4 17:39:51.447016 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 4 17:39:51.453066 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Sep 4 17:39:51.463132 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 4 17:39:51.474094 systemd-logind[2045]: Watching system buttons on /dev/input/event2 (Power Button) Sep 4 17:39:51.474123 systemd-logind[2045]: Watching system buttons on /dev/input/event3 (Sleep Button) Sep 4 17:39:51.474150 systemd-logind[2045]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 4 17:39:51.474427 systemd[1]: Starting sshkeys.service... Sep 4 17:39:51.482074 systemd-logind[2045]: New seat seat0. Sep 4 17:39:51.491195 systemd[1]: Started systemd-logind.service - User Login Management. Sep 4 17:39:51.548153 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Sep 4 17:39:51.558036 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Sep 4 17:39:51.698486 amazon-ssm-agent[2127]: Initializing new seelog logger Sep 4 17:39:51.698911 amazon-ssm-agent[2127]: New Seelog Logger Creation Complete Sep 4 17:39:51.698911 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.698911 amazon-ssm-agent[2127]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 processing appconfig overrides Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 processing appconfig overrides Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 processing appconfig overrides Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024-09-04 17:39:51 INFO Proxy environment variables: Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 4 17:39:51.703420 amazon-ssm-agent[2127]: 2024/09/04 17:39:51 processing appconfig overrides Sep 4 17:39:51.741941 sshd_keygen[2076]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 4 17:39:51.857867 amazon-ssm-agent[2127]: 2024-09-04 17:39:51 INFO https_proxy: Sep 4 17:39:51.938082 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 4 17:39:51.953461 amazon-ssm-agent[2127]: 2024-09-04 17:39:51 INFO http_proxy: Sep 4 17:39:51.955568 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 4 17:39:51.997803 systemd[1]: issuegen.service: Deactivated successfully. Sep 4 17:39:51.998171 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 4 17:39:52.014712 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 4 17:39:52.035047 coreos-metadata[2165]: Sep 04 17:39:52.035 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Sep 4 17:39:52.039329 coreos-metadata[2165]: Sep 04 17:39:52.035 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Sep 4 17:39:52.042410 coreos-metadata[2165]: Sep 04 17:39:52.042 INFO Fetch successful Sep 4 17:39:52.042410 coreos-metadata[2165]: Sep 04 17:39:52.042 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Sep 4 17:39:52.043633 coreos-metadata[2165]: Sep 04 17:39:52.043 INFO Fetch successful Sep 4 17:39:52.052932 unknown[2165]: wrote ssh authorized keys file for user: core Sep 4 17:39:52.056434 amazon-ssm-agent[2127]: 2024-09-04 17:39:51 INFO no_proxy: Sep 4 17:39:52.138560 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 4 17:39:52.152832 amazon-ssm-agent[2127]: 2024-09-04 17:39:51 INFO Checking if agent identity type OnPrem can be assumed Sep 4 17:39:52.160067 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 4 17:39:52.194157 update-ssh-keys[2250]: Updated "/home/core/.ssh/authorized_keys" Sep 4 17:39:52.200426 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 4 17:39:52.209742 systemd[1]: Reached target getty.target - Login Prompts. Sep 4 17:39:52.217958 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Sep 4 17:39:52.228050 systemd[1]: Finished sshkeys.service. Sep 4 17:39:52.254518 amazon-ssm-agent[2127]: 2024-09-04 17:39:51 INFO Checking if agent identity type EC2 can be assumed Sep 4 17:39:52.347851 dbus-daemon[2025]: [system] Successfully activated service 'org.freedesktop.hostname1' Sep 4 17:39:52.348565 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Sep 4 17:39:52.354660 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO Agent will take identity from EC2 Sep 4 17:39:52.354654 dbus-daemon[2025]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.7' (uid=0 pid=2128 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Sep 4 17:39:52.363008 locksmithd[2129]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 4 17:39:52.368481 systemd[1]: Starting polkit.service - Authorization Manager... Sep 4 17:39:52.432249 polkitd[2277]: Started polkitd version 121 Sep 4 17:39:52.470785 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [amazon-ssm-agent] using named pipe channel for IPC Sep 4 17:39:52.540332 polkitd[2277]: Loading rules from directory /etc/polkit-1/rules.d Sep 4 17:39:52.540462 polkitd[2277]: Loading rules from directory /usr/share/polkit-1/rules.d Sep 4 17:39:52.547870 polkitd[2277]: Finished loading, compiling and executing 2 rules Sep 4 17:39:52.552631 dbus-daemon[2025]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Sep 4 17:39:52.552835 systemd[1]: Started polkit.service - Authorization Manager. Sep 4 17:39:52.558566 polkitd[2277]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Sep 4 17:39:52.574178 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [amazon-ssm-agent] using named pipe channel for IPC Sep 4 17:39:52.616746 systemd-hostnamed[2128]: Hostname set to (transient) Sep 4 17:39:52.617826 systemd-resolved[1963]: System hostname changed to 'ip-172-31-20-143'. Sep 4 17:39:52.675403 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [amazon-ssm-agent] using named pipe channel for IPC Sep 4 17:39:52.763454 containerd[2079]: time="2024-09-04T17:39:52.763266800Z" level=info msg="starting containerd" revision=8ccfc03e4e2b73c22899202ae09d0caf906d3863 version=v1.7.20 Sep 4 17:39:52.773162 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Sep 4 17:39:52.873300 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Sep 4 17:39:52.878434 containerd[2079]: time="2024-09-04T17:39:52.878337511Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:39:52.882860 containerd[2079]: time="2024-09-04T17:39:52.882806971Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.48-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:39:52.883223 containerd[2079]: time="2024-09-04T17:39:52.883200367Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 4 17:39:52.884363 containerd[2079]: time="2024-09-04T17:39:52.883315086Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 4 17:39:52.884363 containerd[2079]: time="2024-09-04T17:39:52.883529039Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 4 17:39:52.884363 containerd[2079]: time="2024-09-04T17:39:52.883553278Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 4 17:39:52.884363 containerd[2079]: time="2024-09-04T17:39:52.883625774Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:39:52.884363 containerd[2079]: time="2024-09-04T17:39:52.883647043Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:39:52.884955 containerd[2079]: time="2024-09-04T17:39:52.884925508Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:39:52.885509 containerd[2079]: time="2024-09-04T17:39:52.885483921Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 4 17:39:52.886524 containerd[2079]: time="2024-09-04T17:39:52.885870160Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:39:52.886524 containerd[2079]: time="2024-09-04T17:39:52.885895314Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 4 17:39:52.886524 containerd[2079]: time="2024-09-04T17:39:52.886027532Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:39:52.886524 containerd[2079]: time="2024-09-04T17:39:52.886276344Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:39:52.887580 containerd[2079]: time="2024-09-04T17:39:52.887132243Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:39:52.887580 containerd[2079]: time="2024-09-04T17:39:52.887160210Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 4 17:39:52.887580 containerd[2079]: time="2024-09-04T17:39:52.887274559Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 4 17:39:52.887580 containerd[2079]: time="2024-09-04T17:39:52.887339925Z" level=info msg="metadata content store policy set" policy=shared Sep 4 17:39:52.897321 containerd[2079]: time="2024-09-04T17:39:52.897232373Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 4 17:39:52.897531 containerd[2079]: time="2024-09-04T17:39:52.897491476Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 4 17:39:52.898406 containerd[2079]: time="2024-09-04T17:39:52.897726905Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 4 17:39:52.898406 containerd[2079]: time="2024-09-04T17:39:52.897757927Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 4 17:39:52.898406 containerd[2079]: time="2024-09-04T17:39:52.897779701Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 4 17:39:52.898406 containerd[2079]: time="2024-09-04T17:39:52.898045572Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 4 17:39:52.899469 containerd[2079]: time="2024-09-04T17:39:52.899443755Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900048785Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900080280Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900100801Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900121903Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900142122Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900162925Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900185583Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900208085Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900231092Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900250910Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900279242Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900314946Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900335567Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.900759 containerd[2079]: time="2024-09-04T17:39:52.900354563Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.901437 containerd[2079]: time="2024-09-04T17:39:52.900374462Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902068754Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902103448Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902123775Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902144964Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902165659Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902189527Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902210434Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902229356Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902249679Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902275406Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902308062Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902326086Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902345747Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 4 17:39:52.903774 containerd[2079]: time="2024-09-04T17:39:52.902509903Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 4 17:39:52.904397 containerd[2079]: time="2024-09-04T17:39:52.902538265Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 4 17:39:52.904397 containerd[2079]: time="2024-09-04T17:39:52.902556144Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 4 17:39:52.904397 containerd[2079]: time="2024-09-04T17:39:52.902574656Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 4 17:39:52.904397 containerd[2079]: time="2024-09-04T17:39:52.902590120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.904397 containerd[2079]: time="2024-09-04T17:39:52.902608616Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 4 17:39:52.904397 containerd[2079]: time="2024-09-04T17:39:52.902627623Z" level=info msg="NRI interface is disabled by configuration." Sep 4 17:39:52.904397 containerd[2079]: time="2024-09-04T17:39:52.902644158Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 4 17:39:52.904678 containerd[2079]: time="2024-09-04T17:39:52.903084280Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 4 17:39:52.904678 containerd[2079]: time="2024-09-04T17:39:52.903234381Z" level=info msg="Connect containerd service" Sep 4 17:39:52.904678 containerd[2079]: time="2024-09-04T17:39:52.903292383Z" level=info msg="using legacy CRI server" Sep 4 17:39:52.904678 containerd[2079]: time="2024-09-04T17:39:52.903303589Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 4 17:39:52.907865 containerd[2079]: time="2024-09-04T17:39:52.905729450Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 4 17:39:52.907865 containerd[2079]: time="2024-09-04T17:39:52.907259342Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 4 17:39:52.908635 containerd[2079]: time="2024-09-04T17:39:52.908610362Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 4 17:39:52.908782 containerd[2079]: time="2024-09-04T17:39:52.908767824Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 4 17:39:52.910175 containerd[2079]: time="2024-09-04T17:39:52.908977242Z" level=info msg="Start subscribing containerd event" Sep 4 17:39:52.910175 containerd[2079]: time="2024-09-04T17:39:52.909027075Z" level=info msg="Start recovering state" Sep 4 17:39:52.910175 containerd[2079]: time="2024-09-04T17:39:52.909103535Z" level=info msg="Start event monitor" Sep 4 17:39:52.910175 containerd[2079]: time="2024-09-04T17:39:52.909120700Z" level=info msg="Start snapshots syncer" Sep 4 17:39:52.910175 containerd[2079]: time="2024-09-04T17:39:52.909133421Z" level=info msg="Start cni network conf syncer for default" Sep 4 17:39:52.910175 containerd[2079]: time="2024-09-04T17:39:52.909144267Z" level=info msg="Start streaming server" Sep 4 17:39:52.909358 systemd[1]: Started containerd.service - containerd container runtime. Sep 4 17:39:52.910732 containerd[2079]: time="2024-09-04T17:39:52.910711328Z" level=info msg="containerd successfully booted in 0.151121s" Sep 4 17:39:52.972721 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [amazon-ssm-agent] Starting Core Agent Sep 4 17:39:53.018105 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [amazon-ssm-agent] registrar detected. Attempting registration Sep 4 17:39:53.018105 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [Registrar] Starting registrar module Sep 4 17:39:53.018105 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Sep 4 17:39:53.018105 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [EC2Identity] EC2 registration was successful. Sep 4 17:39:53.018105 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [CredentialRefresher] credentialRefresher has started Sep 4 17:39:53.018105 amazon-ssm-agent[2127]: 2024-09-04 17:39:52 INFO [CredentialRefresher] Starting credentials refresher loop Sep 4 17:39:53.018105 amazon-ssm-agent[2127]: 2024-09-04 17:39:53 INFO EC2RoleProvider Successfully connected with instance profile role credentials Sep 4 17:39:53.074407 amazon-ssm-agent[2127]: 2024-09-04 17:39:53 INFO [CredentialRefresher] Next credential rotation will be in 31.574994816566665 minutes Sep 4 17:39:53.090997 tar[2075]: linux-amd64/LICENSE Sep 4 17:39:53.095764 tar[2075]: linux-amd64/README.md Sep 4 17:39:53.110321 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 4 17:39:54.036460 amazon-ssm-agent[2127]: 2024-09-04 17:39:54 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Sep 4 17:39:54.137428 amazon-ssm-agent[2127]: 2024-09-04 17:39:54 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2314) started Sep 4 17:39:54.238466 amazon-ssm-agent[2127]: 2024-09-04 17:39:54 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Sep 4 17:39:54.572613 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:39:54.575239 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 4 17:39:54.577804 systemd[1]: Startup finished in 10.460s (kernel) + 10.400s (userspace) = 20.860s. Sep 4 17:39:54.580027 (kubelet)[2332]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:39:56.349890 kubelet[2332]: E0904 17:39:56.349805 2332 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:39:56.353442 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:39:56.353742 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:39:59.059393 systemd-resolved[1963]: Clock change detected. Flushing caches. Sep 4 17:39:59.499134 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 4 17:39:59.512745 systemd[1]: Started sshd@0-172.31.20.143:22-139.178.68.195:32858.service - OpenSSH per-connection server daemon (139.178.68.195:32858). Sep 4 17:39:59.722694 sshd[2345]: Accepted publickey for core from 139.178.68.195 port 32858 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:39:59.727601 sshd[2345]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:39:59.747224 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 4 17:39:59.752715 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 4 17:39:59.757158 systemd-logind[2045]: New session 1 of user core. Sep 4 17:39:59.782134 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 4 17:39:59.794678 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 4 17:39:59.809146 (systemd)[2351]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:39:59.998605 systemd[2351]: Queued start job for default target default.target. Sep 4 17:39:59.999233 systemd[2351]: Created slice app.slice - User Application Slice. Sep 4 17:39:59.999282 systemd[2351]: Reached target paths.target - Paths. Sep 4 17:39:59.999301 systemd[2351]: Reached target timers.target - Timers. Sep 4 17:40:00.040518 systemd[2351]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 4 17:40:00.139323 systemd[2351]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 4 17:40:00.139415 systemd[2351]: Reached target sockets.target - Sockets. Sep 4 17:40:00.139436 systemd[2351]: Reached target basic.target - Basic System. Sep 4 17:40:00.139503 systemd[2351]: Reached target default.target - Main User Target. Sep 4 17:40:00.139547 systemd[2351]: Startup finished in 298ms. Sep 4 17:40:00.142941 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 4 17:40:00.162777 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 4 17:40:00.348126 systemd[1]: Started sshd@1-172.31.20.143:22-139.178.68.195:32872.service - OpenSSH per-connection server daemon (139.178.68.195:32872). Sep 4 17:40:00.601285 sshd[2363]: Accepted publickey for core from 139.178.68.195 port 32872 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:40:00.602922 sshd[2363]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:40:00.610428 systemd-logind[2045]: New session 2 of user core. Sep 4 17:40:00.619623 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 4 17:40:00.744816 sshd[2363]: pam_unix(sshd:session): session closed for user core Sep 4 17:40:00.749761 systemd[1]: sshd@1-172.31.20.143:22-139.178.68.195:32872.service: Deactivated successfully. Sep 4 17:40:00.754583 systemd[1]: session-2.scope: Deactivated successfully. Sep 4 17:40:00.755985 systemd-logind[2045]: Session 2 logged out. Waiting for processes to exit. Sep 4 17:40:00.757891 systemd-logind[2045]: Removed session 2. Sep 4 17:40:00.772907 systemd[1]: Started sshd@2-172.31.20.143:22-139.178.68.195:32882.service - OpenSSH per-connection server daemon (139.178.68.195:32882). Sep 4 17:40:00.929959 sshd[2371]: Accepted publickey for core from 139.178.68.195 port 32882 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:40:00.931784 sshd[2371]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:40:00.950358 systemd-logind[2045]: New session 3 of user core. Sep 4 17:40:00.963069 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 4 17:40:01.097131 sshd[2371]: pam_unix(sshd:session): session closed for user core Sep 4 17:40:01.103415 systemd[1]: sshd@2-172.31.20.143:22-139.178.68.195:32882.service: Deactivated successfully. Sep 4 17:40:01.116728 systemd-logind[2045]: Session 3 logged out. Waiting for processes to exit. Sep 4 17:40:01.117082 systemd[1]: session-3.scope: Deactivated successfully. Sep 4 17:40:01.127070 systemd[1]: Started sshd@3-172.31.20.143:22-139.178.68.195:32888.service - OpenSSH per-connection server daemon (139.178.68.195:32888). Sep 4 17:40:01.128928 systemd-logind[2045]: Removed session 3. Sep 4 17:40:01.294358 sshd[2379]: Accepted publickey for core from 139.178.68.195 port 32888 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:40:01.295665 sshd[2379]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:40:01.305756 systemd-logind[2045]: New session 4 of user core. Sep 4 17:40:01.318272 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 4 17:40:01.505076 sshd[2379]: pam_unix(sshd:session): session closed for user core Sep 4 17:40:01.521505 systemd[1]: sshd@3-172.31.20.143:22-139.178.68.195:32888.service: Deactivated successfully. Sep 4 17:40:01.562282 systemd-logind[2045]: Session 4 logged out. Waiting for processes to exit. Sep 4 17:40:01.565779 systemd[1]: session-4.scope: Deactivated successfully. Sep 4 17:40:01.620541 systemd[1]: Started sshd@4-172.31.20.143:22-139.178.68.195:32896.service - OpenSSH per-connection server daemon (139.178.68.195:32896). Sep 4 17:40:01.621755 systemd-logind[2045]: Removed session 4. Sep 4 17:40:01.885674 sshd[2387]: Accepted publickey for core from 139.178.68.195 port 32896 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:40:01.890522 sshd[2387]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:40:01.932126 systemd-logind[2045]: New session 5 of user core. Sep 4 17:40:01.942038 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 4 17:40:02.092681 sudo[2391]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 4 17:40:02.093311 sudo[2391]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:40:02.170605 sudo[2391]: pam_unix(sudo:session): session closed for user root Sep 4 17:40:02.204622 sshd[2387]: pam_unix(sshd:session): session closed for user core Sep 4 17:40:02.213983 systemd[1]: sshd@4-172.31.20.143:22-139.178.68.195:32896.service: Deactivated successfully. Sep 4 17:40:02.230016 systemd-logind[2045]: Session 5 logged out. Waiting for processes to exit. Sep 4 17:40:02.235423 systemd[1]: session-5.scope: Deactivated successfully. Sep 4 17:40:02.257959 systemd[1]: Started sshd@5-172.31.20.143:22-139.178.68.195:32906.service - OpenSSH per-connection server daemon (139.178.68.195:32906). Sep 4 17:40:02.260638 systemd-logind[2045]: Removed session 5. Sep 4 17:40:02.475440 sshd[2396]: Accepted publickey for core from 139.178.68.195 port 32906 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:40:02.483615 sshd[2396]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:40:02.512804 systemd-logind[2045]: New session 6 of user core. Sep 4 17:40:02.522141 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 4 17:40:02.650832 sudo[2401]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 4 17:40:02.651414 sudo[2401]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:40:02.673588 sudo[2401]: pam_unix(sudo:session): session closed for user root Sep 4 17:40:02.714146 sudo[2400]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 4 17:40:02.729575 sudo[2400]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:40:02.823009 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 4 17:40:02.869203 auditctl[2404]: No rules Sep 4 17:40:02.869979 systemd[1]: audit-rules.service: Deactivated successfully. Sep 4 17:40:02.870419 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 4 17:40:02.896704 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 4 17:40:03.079842 augenrules[2423]: No rules Sep 4 17:40:03.080976 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 4 17:40:03.087165 sudo[2400]: pam_unix(sudo:session): session closed for user root Sep 4 17:40:03.133872 sshd[2396]: pam_unix(sshd:session): session closed for user core Sep 4 17:40:03.172815 systemd[1]: Started sshd@6-172.31.20.143:22-139.178.68.195:32922.service - OpenSSH per-connection server daemon (139.178.68.195:32922). Sep 4 17:40:03.176292 systemd[1]: sshd@5-172.31.20.143:22-139.178.68.195:32906.service: Deactivated successfully. Sep 4 17:40:03.193290 systemd[1]: session-6.scope: Deactivated successfully. Sep 4 17:40:03.215201 systemd-logind[2045]: Session 6 logged out. Waiting for processes to exit. Sep 4 17:40:03.259191 systemd-logind[2045]: Removed session 6. Sep 4 17:40:03.455680 sshd[2430]: Accepted publickey for core from 139.178.68.195 port 32922 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:40:03.457884 sshd[2430]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:40:03.471751 systemd-logind[2045]: New session 7 of user core. Sep 4 17:40:03.482727 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 4 17:40:03.602501 sudo[2436]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 4 17:40:03.602939 sudo[2436]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 4 17:40:03.992970 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 4 17:40:03.994077 (dockerd)[2446]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 4 17:40:04.940137 dockerd[2446]: time="2024-09-04T17:40:04.940064427Z" level=info msg="Starting up" Sep 4 17:40:05.124164 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1529065623-merged.mount: Deactivated successfully. Sep 4 17:40:05.691825 dockerd[2446]: time="2024-09-04T17:40:05.691537534Z" level=info msg="Loading containers: start." Sep 4 17:40:05.920930 kernel: Initializing XFRM netlink socket Sep 4 17:40:05.978641 (udev-worker)[2466]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:40:06.075264 systemd-networkd[1648]: docker0: Link UP Sep 4 17:40:06.099617 dockerd[2446]: time="2024-09-04T17:40:06.099567384Z" level=info msg="Loading containers: done." Sep 4 17:40:06.151352 dockerd[2446]: time="2024-09-04T17:40:06.151304060Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 4 17:40:06.151561 dockerd[2446]: time="2024-09-04T17:40:06.151455783Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Sep 4 17:40:06.151653 dockerd[2446]: time="2024-09-04T17:40:06.151622584Z" level=info msg="Daemon has completed initialization" Sep 4 17:40:06.208883 dockerd[2446]: time="2024-09-04T17:40:06.208214270Z" level=info msg="API listen on /run/docker.sock" Sep 4 17:40:06.208740 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 4 17:40:07.628051 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 4 17:40:07.673723 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:40:07.972669 containerd[2079]: time="2024-09-04T17:40:07.967872816Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.13\"" Sep 4 17:40:08.216648 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:40:08.241808 (kubelet)[2602]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:40:08.319581 kubelet[2602]: E0904 17:40:08.319478 2602 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:40:08.333935 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:40:08.334091 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:40:08.930512 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3820066239.mount: Deactivated successfully. Sep 4 17:40:10.973432 containerd[2079]: time="2024-09-04T17:40:10.973375300Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.28.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:10.975433 containerd[2079]: time="2024-09-04T17:40:10.975383262Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.28.13: active requests=0, bytes read=34530735" Sep 4 17:40:10.978823 containerd[2079]: time="2024-09-04T17:40:10.978777331Z" level=info msg="ImageCreate event name:\"sha256:5447bb21fa283749e558782cbef636f1991732f1b8f345296a5204ccf0b5f7b7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:10.987616 containerd[2079]: time="2024-09-04T17:40:10.987546852Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:7d2c9256ad576a0b3745b749efe7f4fa8b276ec7ef448fc0f45794ca78eb8625\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:10.989276 containerd[2079]: time="2024-09-04T17:40:10.989114185Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.28.13\" with image id \"sha256:5447bb21fa283749e558782cbef636f1991732f1b8f345296a5204ccf0b5f7b7\", repo tag \"registry.k8s.io/kube-apiserver:v1.28.13\", repo digest \"registry.k8s.io/kube-apiserver@sha256:7d2c9256ad576a0b3745b749efe7f4fa8b276ec7ef448fc0f45794ca78eb8625\", size \"34527535\" in 3.021176616s" Sep 4 17:40:10.989276 containerd[2079]: time="2024-09-04T17:40:10.989169791Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.13\" returns image reference \"sha256:5447bb21fa283749e558782cbef636f1991732f1b8f345296a5204ccf0b5f7b7\"" Sep 4 17:40:11.018630 containerd[2079]: time="2024-09-04T17:40:11.018523121Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.13\"" Sep 4 17:40:13.216494 containerd[2079]: time="2024-09-04T17:40:13.216442069Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.28.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:13.218323 containerd[2079]: time="2024-09-04T17:40:13.218088196Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.28.13: active requests=0, bytes read=31849709" Sep 4 17:40:13.220763 containerd[2079]: time="2024-09-04T17:40:13.220703901Z" level=info msg="ImageCreate event name:\"sha256:f1a0a396058d414b391ade9dba6e95d7a71ee665b09fc0fc420126ac21c155a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:13.224120 containerd[2079]: time="2024-09-04T17:40:13.224058192Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:e7b44c1741fe1802d159ffdbd0d1f78d48a4185d7fb1cdf8a112fbb50696f7e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:13.225697 containerd[2079]: time="2024-09-04T17:40:13.225126781Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.28.13\" with image id \"sha256:f1a0a396058d414b391ade9dba6e95d7a71ee665b09fc0fc420126ac21c155a5\", repo tag \"registry.k8s.io/kube-controller-manager:v1.28.13\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:e7b44c1741fe1802d159ffdbd0d1f78d48a4185d7fb1cdf8a112fbb50696f7e1\", size \"33399655\" in 2.206496926s" Sep 4 17:40:13.225697 containerd[2079]: time="2024-09-04T17:40:13.225172886Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.13\" returns image reference \"sha256:f1a0a396058d414b391ade9dba6e95d7a71ee665b09fc0fc420126ac21c155a5\"" Sep 4 17:40:13.251352 containerd[2079]: time="2024-09-04T17:40:13.251309390Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.13\"" Sep 4 17:40:14.683915 containerd[2079]: time="2024-09-04T17:40:14.683857036Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.28.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:14.685884 containerd[2079]: time="2024-09-04T17:40:14.685692241Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.28.13: active requests=0, bytes read=17097777" Sep 4 17:40:14.689549 containerd[2079]: time="2024-09-04T17:40:14.689036157Z" level=info msg="ImageCreate event name:\"sha256:a60f64c0f37d085a5fcafef1b2a7adc9be95184dae7d8a5d1dbf6ca4681d328a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:14.693017 containerd[2079]: time="2024-09-04T17:40:14.692973404Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:efeb791718f4b9c62bd683f5b403da520f3651cb36ad9f800e0f98b595beafa4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:14.694083 containerd[2079]: time="2024-09-04T17:40:14.694044577Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.28.13\" with image id \"sha256:a60f64c0f37d085a5fcafef1b2a7adc9be95184dae7d8a5d1dbf6ca4681d328a\", repo tag \"registry.k8s.io/kube-scheduler:v1.28.13\", repo digest \"registry.k8s.io/kube-scheduler@sha256:efeb791718f4b9c62bd683f5b403da520f3651cb36ad9f800e0f98b595beafa4\", size \"18647741\" in 1.442685901s" Sep 4 17:40:14.694265 containerd[2079]: time="2024-09-04T17:40:14.694222686Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.13\" returns image reference \"sha256:a60f64c0f37d085a5fcafef1b2a7adc9be95184dae7d8a5d1dbf6ca4681d328a\"" Sep 4 17:40:14.719026 containerd[2079]: time="2024-09-04T17:40:14.718982745Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.13\"" Sep 4 17:40:15.891059 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount550289446.mount: Deactivated successfully. Sep 4 17:40:16.729664 containerd[2079]: time="2024-09-04T17:40:16.729609137Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.28.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:16.731319 containerd[2079]: time="2024-09-04T17:40:16.731013122Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.28.13: active requests=0, bytes read=28303449" Sep 4 17:40:16.732901 containerd[2079]: time="2024-09-04T17:40:16.732842016Z" level=info msg="ImageCreate event name:\"sha256:31fde28e72a31599555ab5aba850caa90b9254b760b1007bfb662d086bb672fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:16.738955 containerd[2079]: time="2024-09-04T17:40:16.737529195Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:537633f399f87ce85d44fc8471ece97a83632198f99b3f7e08770beca95e9fa1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:16.740382 containerd[2079]: time="2024-09-04T17:40:16.740143172Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.28.13\" with image id \"sha256:31fde28e72a31599555ab5aba850caa90b9254b760b1007bfb662d086bb672fc\", repo tag \"registry.k8s.io/kube-proxy:v1.28.13\", repo digest \"registry.k8s.io/kube-proxy@sha256:537633f399f87ce85d44fc8471ece97a83632198f99b3f7e08770beca95e9fa1\", size \"28302468\" in 2.021109442s" Sep 4 17:40:16.740382 containerd[2079]: time="2024-09-04T17:40:16.740197475Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.13\" returns image reference \"sha256:31fde28e72a31599555ab5aba850caa90b9254b760b1007bfb662d086bb672fc\"" Sep 4 17:40:16.777694 containerd[2079]: time="2024-09-04T17:40:16.777594092Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Sep 4 17:40:17.320685 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount309278816.mount: Deactivated successfully. Sep 4 17:40:17.337384 containerd[2079]: time="2024-09-04T17:40:17.337337364Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:17.340749 containerd[2079]: time="2024-09-04T17:40:17.340376476Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322290" Sep 4 17:40:17.344285 containerd[2079]: time="2024-09-04T17:40:17.343428902Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:17.348291 containerd[2079]: time="2024-09-04T17:40:17.347580069Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:17.349026 containerd[2079]: time="2024-09-04T17:40:17.348984263Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 571.258939ms" Sep 4 17:40:17.349026 containerd[2079]: time="2024-09-04T17:40:17.349023254Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Sep 4 17:40:17.381651 containerd[2079]: time="2024-09-04T17:40:17.381609158Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\"" Sep 4 17:40:18.001762 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1197303698.mount: Deactivated successfully. Sep 4 17:40:18.376355 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 4 17:40:18.388380 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:40:19.502466 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:40:19.503412 (kubelet)[2746]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:40:19.596266 kubelet[2746]: E0904 17:40:19.594928 2746 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:40:19.602134 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:40:19.602397 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:40:21.614286 containerd[2079]: time="2024-09-04T17:40:21.614147301Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.10-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:21.617705 containerd[2079]: time="2024-09-04T17:40:21.617633998Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.10-0: active requests=0, bytes read=56651625" Sep 4 17:40:21.619734 containerd[2079]: time="2024-09-04T17:40:21.619592001Z" level=info msg="ImageCreate event name:\"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:21.625143 containerd[2079]: time="2024-09-04T17:40:21.624890932Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:21.636795 containerd[2079]: time="2024-09-04T17:40:21.636740399Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.10-0\" with image id \"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\", repo tag \"registry.k8s.io/etcd:3.5.10-0\", repo digest \"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\", size \"56649232\" in 4.255095542s" Sep 4 17:40:21.636795 containerd[2079]: time="2024-09-04T17:40:21.636789655Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\" returns image reference \"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\"" Sep 4 17:40:21.693952 containerd[2079]: time="2024-09-04T17:40:21.693912469Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\"" Sep 4 17:40:22.276647 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2947025150.mount: Deactivated successfully. Sep 4 17:40:23.020632 containerd[2079]: time="2024-09-04T17:40:23.020575394Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:23.022785 containerd[2079]: time="2024-09-04T17:40:23.022543127Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.10.1: active requests=0, bytes read=16191749" Sep 4 17:40:23.024888 containerd[2079]: time="2024-09-04T17:40:23.024846715Z" level=info msg="ImageCreate event name:\"sha256:ead0a4a53df89fd173874b46093b6e62d8c72967bbf606d672c9e8c9b601a4fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:23.028198 containerd[2079]: time="2024-09-04T17:40:23.028106427Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:23.029315 containerd[2079]: time="2024-09-04T17:40:23.029115281Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.10.1\" with image id \"sha256:ead0a4a53df89fd173874b46093b6e62d8c72967bbf606d672c9e8c9b601a4fc\", repo tag \"registry.k8s.io/coredns/coredns:v1.10.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e\", size \"16190758\" in 1.335157251s" Sep 4 17:40:23.029315 containerd[2079]: time="2024-09-04T17:40:23.029163004Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\" returns image reference \"sha256:ead0a4a53df89fd173874b46093b6e62d8c72967bbf606d672c9e8c9b601a4fc\"" Sep 4 17:40:23.721770 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Sep 4 17:40:27.116267 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:40:27.124612 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:40:27.159464 systemd[1]: Reloading requested from client PID 2862 ('systemctl') (unit session-7.scope)... Sep 4 17:40:27.159483 systemd[1]: Reloading... Sep 4 17:40:27.316281 zram_generator::config[2900]: No configuration found. Sep 4 17:40:27.502857 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:40:27.594776 systemd[1]: Reloading finished in 434 ms. Sep 4 17:40:27.663131 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 4 17:40:27.663341 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 4 17:40:27.663941 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:40:27.671737 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:40:27.936497 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:40:27.949883 (kubelet)[2972]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 17:40:28.032783 kubelet[2972]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:40:28.033482 kubelet[2972]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 4 17:40:28.033482 kubelet[2972]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:40:28.033482 kubelet[2972]: I0904 17:40:28.033345 2972 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 17:40:28.661258 kubelet[2972]: I0904 17:40:28.661204 2972 server.go:467] "Kubelet version" kubeletVersion="v1.28.7" Sep 4 17:40:28.661258 kubelet[2972]: I0904 17:40:28.661235 2972 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 17:40:28.661707 kubelet[2972]: I0904 17:40:28.661685 2972 server.go:895] "Client rotation is on, will bootstrap in background" Sep 4 17:40:28.697843 kubelet[2972]: I0904 17:40:28.697237 2972 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 17:40:28.710340 kubelet[2972]: E0904 17:40:28.710313 2972 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://172.31.20.143:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.730967 kubelet[2972]: I0904 17:40:28.730934 2972 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 17:40:28.733289 kubelet[2972]: I0904 17:40:28.733233 2972 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 17:40:28.733528 kubelet[2972]: I0904 17:40:28.733499 2972 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Sep 4 17:40:28.733682 kubelet[2972]: I0904 17:40:28.733532 2972 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 17:40:28.733682 kubelet[2972]: I0904 17:40:28.733548 2972 container_manager_linux.go:301] "Creating device plugin manager" Sep 4 17:40:28.734977 kubelet[2972]: I0904 17:40:28.734938 2972 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:40:28.739117 kubelet[2972]: I0904 17:40:28.739074 2972 kubelet.go:393] "Attempting to sync node with API server" Sep 4 17:40:28.739117 kubelet[2972]: I0904 17:40:28.739119 2972 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 17:40:28.739288 kubelet[2972]: I0904 17:40:28.739156 2972 kubelet.go:309] "Adding apiserver pod source" Sep 4 17:40:28.739288 kubelet[2972]: I0904 17:40:28.739176 2972 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 17:40:28.742276 kubelet[2972]: W0904 17:40:28.741139 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://172.31.20.143:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.742276 kubelet[2972]: E0904 17:40:28.741198 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://172.31.20.143:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.742276 kubelet[2972]: W0904 17:40:28.741874 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://172.31.20.143:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-143&limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.742276 kubelet[2972]: E0904 17:40:28.741931 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://172.31.20.143:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-143&limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.742276 kubelet[2972]: I0904 17:40:28.742030 2972 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="v1.7.20" apiVersion="v1" Sep 4 17:40:28.746851 kubelet[2972]: W0904 17:40:28.746814 2972 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 4 17:40:28.747830 kubelet[2972]: I0904 17:40:28.747763 2972 server.go:1232] "Started kubelet" Sep 4 17:40:28.749534 kubelet[2972]: I0904 17:40:28.748356 2972 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Sep 4 17:40:28.749534 kubelet[2972]: I0904 17:40:28.748735 2972 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 17:40:28.749534 kubelet[2972]: I0904 17:40:28.748789 2972 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 17:40:28.749534 kubelet[2972]: I0904 17:40:28.749529 2972 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 17:40:28.753266 kubelet[2972]: I0904 17:40:28.752808 2972 server.go:462] "Adding debug handlers to kubelet server" Sep 4 17:40:28.756036 kubelet[2972]: E0904 17:40:28.756010 2972 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Sep 4 17:40:28.756271 kubelet[2972]: E0904 17:40:28.756237 2972 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 17:40:28.758626 kubelet[2972]: I0904 17:40:28.758606 2972 volume_manager.go:291] "Starting Kubelet Volume Manager" Sep 4 17:40:28.763751 kubelet[2972]: I0904 17:40:28.763721 2972 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Sep 4 17:40:28.766822 kubelet[2972]: I0904 17:40:28.763981 2972 reconciler_new.go:29] "Reconciler: start to sync state" Sep 4 17:40:28.767622 kubelet[2972]: W0904 17:40:28.767571 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://172.31.20.143:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.767765 kubelet[2972]: E0904 17:40:28.767755 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://172.31.20.143:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.770021 kubelet[2972]: E0904 17:40:28.769046 2972 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.143:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-143?timeout=10s\": dial tcp 172.31.20.143:6443: connect: connection refused" interval="200ms" Sep 4 17:40:28.770021 kubelet[2972]: E0904 17:40:28.769109 2972 event.go:289] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"ip-172-31-20-143.17f21b4ded85eb0a", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ip-172-31-20-143", UID:"ip-172-31-20-143", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"ip-172-31-20-143"}, FirstTimestamp:time.Date(2024, time.September, 4, 17, 40, 28, 747729674, time.Local), LastTimestamp:time.Date(2024, time.September, 4, 17, 40, 28, 747729674, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"ip-172-31-20-143"}': 'Post "https://172.31.20.143:6443/api/v1/namespaces/default/events": dial tcp 172.31.20.143:6443: connect: connection refused'(may retry after sleeping) Sep 4 17:40:28.804521 kubelet[2972]: I0904 17:40:28.804491 2972 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 17:40:28.806694 kubelet[2972]: I0904 17:40:28.806669 2972 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 17:40:28.806919 kubelet[2972]: I0904 17:40:28.806905 2972 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 4 17:40:28.807019 kubelet[2972]: I0904 17:40:28.807010 2972 kubelet.go:2303] "Starting kubelet main sync loop" Sep 4 17:40:28.807175 kubelet[2972]: E0904 17:40:28.807165 2972 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 17:40:28.817448 kubelet[2972]: W0904 17:40:28.817384 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://172.31.20.143:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.817448 kubelet[2972]: E0904 17:40:28.817451 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://172.31.20.143:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:28.846717 kubelet[2972]: I0904 17:40:28.846640 2972 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 4 17:40:28.846717 kubelet[2972]: I0904 17:40:28.846662 2972 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 4 17:40:28.846717 kubelet[2972]: I0904 17:40:28.846682 2972 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:40:28.849915 kubelet[2972]: I0904 17:40:28.849886 2972 policy_none.go:49] "None policy: Start" Sep 4 17:40:28.850908 kubelet[2972]: I0904 17:40:28.850885 2972 memory_manager.go:169] "Starting memorymanager" policy="None" Sep 4 17:40:28.851024 kubelet[2972]: I0904 17:40:28.850919 2972 state_mem.go:35] "Initializing new in-memory state store" Sep 4 17:40:28.858469 kubelet[2972]: I0904 17:40:28.857585 2972 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 17:40:28.858469 kubelet[2972]: I0904 17:40:28.857908 2972 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 17:40:28.861724 kubelet[2972]: E0904 17:40:28.861697 2972 eviction_manager.go:258] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-20-143\" not found" Sep 4 17:40:28.869839 kubelet[2972]: I0904 17:40:28.869550 2972 kubelet_node_status.go:70] "Attempting to register node" node="ip-172-31-20-143" Sep 4 17:40:28.869839 kubelet[2972]: E0904 17:40:28.869821 2972 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://172.31.20.143:6443/api/v1/nodes\": dial tcp 172.31.20.143:6443: connect: connection refused" node="ip-172-31-20-143" Sep 4 17:40:28.908922 kubelet[2972]: I0904 17:40:28.908838 2972 topology_manager.go:215] "Topology Admit Handler" podUID="0a59b5a908b97dc850b40c944a13c186" podNamespace="kube-system" podName="kube-apiserver-ip-172-31-20-143" Sep 4 17:40:28.910563 kubelet[2972]: I0904 17:40:28.910536 2972 topology_manager.go:215] "Topology Admit Handler" podUID="44ef5d5ac47f06aa98b4f00afbce8a61" podNamespace="kube-system" podName="kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:28.913688 kubelet[2972]: I0904 17:40:28.913425 2972 topology_manager.go:215] "Topology Admit Handler" podUID="6b2842a9bd421343e3cb87d167be4d72" podNamespace="kube-system" podName="kube-scheduler-ip-172-31-20-143" Sep 4 17:40:28.970496 kubelet[2972]: E0904 17:40:28.970454 2972 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.143:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-143?timeout=10s\": dial tcp 172.31.20.143:6443: connect: connection refused" interval="400ms" Sep 4 17:40:29.065986 kubelet[2972]: I0904 17:40:29.065939 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0a59b5a908b97dc850b40c944a13c186-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-143\" (UID: \"0a59b5a908b97dc850b40c944a13c186\") " pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:29.065986 kubelet[2972]: I0904 17:40:29.066000 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:29.066687 kubelet[2972]: I0904 17:40:29.066035 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:29.066687 kubelet[2972]: I0904 17:40:29.066063 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0a59b5a908b97dc850b40c944a13c186-ca-certs\") pod \"kube-apiserver-ip-172-31-20-143\" (UID: \"0a59b5a908b97dc850b40c944a13c186\") " pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:29.066687 kubelet[2972]: I0904 17:40:29.066092 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0a59b5a908b97dc850b40c944a13c186-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-143\" (UID: \"0a59b5a908b97dc850b40c944a13c186\") " pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:29.066687 kubelet[2972]: I0904 17:40:29.066162 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:29.066687 kubelet[2972]: I0904 17:40:29.066190 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:29.066818 kubelet[2972]: I0904 17:40:29.066221 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6b2842a9bd421343e3cb87d167be4d72-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-143\" (UID: \"6b2842a9bd421343e3cb87d167be4d72\") " pod="kube-system/kube-scheduler-ip-172-31-20-143" Sep 4 17:40:29.066818 kubelet[2972]: I0904 17:40:29.066269 2972 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:29.071635 kubelet[2972]: I0904 17:40:29.071608 2972 kubelet_node_status.go:70] "Attempting to register node" node="ip-172-31-20-143" Sep 4 17:40:29.072006 kubelet[2972]: E0904 17:40:29.071981 2972 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://172.31.20.143:6443/api/v1/nodes\": dial tcp 172.31.20.143:6443: connect: connection refused" node="ip-172-31-20-143" Sep 4 17:40:29.219454 containerd[2079]: time="2024-09-04T17:40:29.218904612Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-143,Uid:0a59b5a908b97dc850b40c944a13c186,Namespace:kube-system,Attempt:0,}" Sep 4 17:40:29.236436 containerd[2079]: time="2024-09-04T17:40:29.236049534Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-143,Uid:6b2842a9bd421343e3cb87d167be4d72,Namespace:kube-system,Attempt:0,}" Sep 4 17:40:29.236436 containerd[2079]: time="2024-09-04T17:40:29.236048586Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-143,Uid:44ef5d5ac47f06aa98b4f00afbce8a61,Namespace:kube-system,Attempt:0,}" Sep 4 17:40:29.371340 kubelet[2972]: E0904 17:40:29.371306 2972 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.143:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-143?timeout=10s\": dial tcp 172.31.20.143:6443: connect: connection refused" interval="800ms" Sep 4 17:40:29.474607 kubelet[2972]: I0904 17:40:29.474510 2972 kubelet_node_status.go:70] "Attempting to register node" node="ip-172-31-20-143" Sep 4 17:40:29.475123 kubelet[2972]: E0904 17:40:29.475096 2972 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://172.31.20.143:6443/api/v1/nodes\": dial tcp 172.31.20.143:6443: connect: connection refused" node="ip-172-31-20-143" Sep 4 17:40:29.743388 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2487750692.mount: Deactivated successfully. Sep 4 17:40:29.756329 containerd[2079]: time="2024-09-04T17:40:29.756271834Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:40:29.757622 containerd[2079]: time="2024-09-04T17:40:29.757581134Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:40:29.758885 containerd[2079]: time="2024-09-04T17:40:29.758836284Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Sep 4 17:40:29.759862 containerd[2079]: time="2024-09-04T17:40:29.759812480Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 4 17:40:29.761200 containerd[2079]: time="2024-09-04T17:40:29.761167202Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:40:29.762949 containerd[2079]: time="2024-09-04T17:40:29.762912132Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:40:29.764017 containerd[2079]: time="2024-09-04T17:40:29.763934271Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 4 17:40:29.766473 containerd[2079]: time="2024-09-04T17:40:29.766370443Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:40:29.770654 containerd[2079]: time="2024-09-04T17:40:29.769171873Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 532.763651ms" Sep 4 17:40:29.776272 containerd[2079]: time="2024-09-04T17:40:29.775880913Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 556.881229ms" Sep 4 17:40:29.776272 containerd[2079]: time="2024-09-04T17:40:29.776127082Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 539.962328ms" Sep 4 17:40:29.955356 kubelet[2972]: W0904 17:40:29.950987 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://172.31.20.143:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:29.955356 kubelet[2972]: E0904 17:40:29.953405 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://172.31.20.143:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:29.990213 containerd[2079]: time="2024-09-04T17:40:29.989860744Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:40:29.990213 containerd[2079]: time="2024-09-04T17:40:29.989925054Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:40:29.990213 containerd[2079]: time="2024-09-04T17:40:29.989941222Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:29.990213 containerd[2079]: time="2024-09-04T17:40:29.990122447Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:29.997323 containerd[2079]: time="2024-09-04T17:40:29.996683927Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:40:29.997323 containerd[2079]: time="2024-09-04T17:40:29.996749925Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:40:29.997323 containerd[2079]: time="2024-09-04T17:40:29.996765884Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:29.997323 containerd[2079]: time="2024-09-04T17:40:29.997000397Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:30.035376 containerd[2079]: time="2024-09-04T17:40:30.029925903Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:40:30.035376 containerd[2079]: time="2024-09-04T17:40:30.030018301Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:40:30.035376 containerd[2079]: time="2024-09-04T17:40:30.030042166Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:30.036583 containerd[2079]: time="2024-09-04T17:40:30.036382338Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:30.091545 kubelet[2972]: W0904 17:40:30.090745 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://172.31.20.143:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-143&limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:30.091545 kubelet[2972]: E0904 17:40:30.090828 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://172.31.20.143:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-143&limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:30.148567 kubelet[2972]: W0904 17:40:30.148498 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://172.31.20.143:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:30.148567 kubelet[2972]: E0904 17:40:30.148575 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://172.31.20.143:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:30.173372 kubelet[2972]: E0904 17:40:30.172067 2972 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.143:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-143?timeout=10s\": dial tcp 172.31.20.143:6443: connect: connection refused" interval="1.6s" Sep 4 17:40:30.227376 containerd[2079]: time="2024-09-04T17:40:30.227333345Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-143,Uid:0a59b5a908b97dc850b40c944a13c186,Namespace:kube-system,Attempt:0,} returns sandbox id \"f26d30efbe450aa04f9c2cbcd5f5b2d1de54435543aba8d5f475cce1703f82d3\"" Sep 4 17:40:30.228360 containerd[2079]: time="2024-09-04T17:40:30.228312356Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-143,Uid:44ef5d5ac47f06aa98b4f00afbce8a61,Namespace:kube-system,Attempt:0,} returns sandbox id \"7ab0f5a3d699b3896b1cc70c93f3bff768f6532092bdd932c8cde4c766f2dad6\"" Sep 4 17:40:30.228885 containerd[2079]: time="2024-09-04T17:40:30.228860642Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-143,Uid:6b2842a9bd421343e3cb87d167be4d72,Namespace:kube-system,Attempt:0,} returns sandbox id \"6cac7bb63ccc218e1f530235f4cc5808c3fca5e0ccfcf89580de0053fb4c38c1\"" Sep 4 17:40:30.241574 containerd[2079]: time="2024-09-04T17:40:30.241526141Z" level=info msg="CreateContainer within sandbox \"f26d30efbe450aa04f9c2cbcd5f5b2d1de54435543aba8d5f475cce1703f82d3\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 4 17:40:30.242721 containerd[2079]: time="2024-09-04T17:40:30.242609117Z" level=info msg="CreateContainer within sandbox \"6cac7bb63ccc218e1f530235f4cc5808c3fca5e0ccfcf89580de0053fb4c38c1\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 4 17:40:30.247181 containerd[2079]: time="2024-09-04T17:40:30.247117465Z" level=info msg="CreateContainer within sandbox \"7ab0f5a3d699b3896b1cc70c93f3bff768f6532092bdd932c8cde4c766f2dad6\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 4 17:40:30.276991 kubelet[2972]: I0904 17:40:30.276908 2972 kubelet_node_status.go:70] "Attempting to register node" node="ip-172-31-20-143" Sep 4 17:40:30.278011 kubelet[2972]: E0904 17:40:30.277996 2972 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://172.31.20.143:6443/api/v1/nodes\": dial tcp 172.31.20.143:6443: connect: connection refused" node="ip-172-31-20-143" Sep 4 17:40:30.279462 containerd[2079]: time="2024-09-04T17:40:30.279417791Z" level=info msg="CreateContainer within sandbox \"7ab0f5a3d699b3896b1cc70c93f3bff768f6532092bdd932c8cde4c766f2dad6\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"bee494728ab0a9cd7334dce1badaac45e0d9da4c73409fa45d941ba0f971126a\"" Sep 4 17:40:30.280893 containerd[2079]: time="2024-09-04T17:40:30.280027192Z" level=info msg="StartContainer for \"bee494728ab0a9cd7334dce1badaac45e0d9da4c73409fa45d941ba0f971126a\"" Sep 4 17:40:30.282393 containerd[2079]: time="2024-09-04T17:40:30.282355743Z" level=info msg="CreateContainer within sandbox \"6cac7bb63ccc218e1f530235f4cc5808c3fca5e0ccfcf89580de0053fb4c38c1\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"0e089d9cb36b6332e22ad835626d20962d57f0fa01363b032ef03a4af90aee42\"" Sep 4 17:40:30.283104 containerd[2079]: time="2024-09-04T17:40:30.283073865Z" level=info msg="StartContainer for \"0e089d9cb36b6332e22ad835626d20962d57f0fa01363b032ef03a4af90aee42\"" Sep 4 17:40:30.287511 containerd[2079]: time="2024-09-04T17:40:30.287461970Z" level=info msg="CreateContainer within sandbox \"f26d30efbe450aa04f9c2cbcd5f5b2d1de54435543aba8d5f475cce1703f82d3\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"757ff7008b0ac8203a1ffe2e29bb3342b9c714d657306a896d18ecec04d61fcc\"" Sep 4 17:40:30.288163 containerd[2079]: time="2024-09-04T17:40:30.288114822Z" level=info msg="StartContainer for \"757ff7008b0ac8203a1ffe2e29bb3342b9c714d657306a896d18ecec04d61fcc\"" Sep 4 17:40:30.296418 kubelet[2972]: W0904 17:40:30.296344 2972 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://172.31.20.143:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:30.296418 kubelet[2972]: E0904 17:40:30.296431 2972 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://172.31.20.143:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:30.442156 containerd[2079]: time="2024-09-04T17:40:30.442106791Z" level=info msg="StartContainer for \"bee494728ab0a9cd7334dce1badaac45e0d9da4c73409fa45d941ba0f971126a\" returns successfully" Sep 4 17:40:30.507740 containerd[2079]: time="2024-09-04T17:40:30.507306311Z" level=info msg="StartContainer for \"0e089d9cb36b6332e22ad835626d20962d57f0fa01363b032ef03a4af90aee42\" returns successfully" Sep 4 17:40:30.509471 containerd[2079]: time="2024-09-04T17:40:30.509228849Z" level=info msg="StartContainer for \"757ff7008b0ac8203a1ffe2e29bb3342b9c714d657306a896d18ecec04d61fcc\" returns successfully" Sep 4 17:40:30.841754 kubelet[2972]: E0904 17:40:30.841649 2972 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://172.31.20.143:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 172.31.20.143:6443: connect: connection refused Sep 4 17:40:31.216351 kubelet[2972]: E0904 17:40:31.216029 2972 event.go:289] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"ip-172-31-20-143.17f21b4ded85eb0a", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ip-172-31-20-143", UID:"ip-172-31-20-143", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"ip-172-31-20-143"}, FirstTimestamp:time.Date(2024, time.September, 4, 17, 40, 28, 747729674, time.Local), LastTimestamp:time.Date(2024, time.September, 4, 17, 40, 28, 747729674, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"ip-172-31-20-143"}': 'Post "https://172.31.20.143:6443/api/v1/namespaces/default/events": dial tcp 172.31.20.143:6443: connect: connection refused'(may retry after sleeping) Sep 4 17:40:31.886321 kubelet[2972]: I0904 17:40:31.882932 2972 kubelet_node_status.go:70] "Attempting to register node" node="ip-172-31-20-143" Sep 4 17:40:34.228588 kubelet[2972]: E0904 17:40:34.228461 2972 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-20-143\" not found" node="ip-172-31-20-143" Sep 4 17:40:34.364407 kubelet[2972]: I0904 17:40:34.364207 2972 kubelet_node_status.go:73] "Successfully registered node" node="ip-172-31-20-143" Sep 4 17:40:34.751561 kubelet[2972]: I0904 17:40:34.751486 2972 apiserver.go:52] "Watching apiserver" Sep 4 17:40:34.764612 kubelet[2972]: I0904 17:40:34.764555 2972 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Sep 4 17:40:34.956827 kubelet[2972]: E0904 17:40:34.956665 2972 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ip-172-31-20-143\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:37.139430 systemd[1]: Reloading requested from client PID 3247 ('systemctl') (unit session-7.scope)... Sep 4 17:40:37.139495 systemd[1]: Reloading... Sep 4 17:40:37.295316 zram_generator::config[3291]: No configuration found. Sep 4 17:40:37.517118 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:40:37.668495 systemd[1]: Reloading finished in 528 ms. Sep 4 17:40:37.710470 kubelet[2972]: I0904 17:40:37.710434 2972 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 17:40:37.710914 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:40:37.725316 systemd[1]: kubelet.service: Deactivated successfully. Sep 4 17:40:37.726739 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:40:37.738777 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:40:37.797899 update_engine[2052]: I0904 17:40:37.796712 2052 update_attempter.cc:509] Updating boot flags... Sep 4 17:40:37.896266 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (3358) Sep 4 17:40:38.277281 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (3362) Sep 4 17:40:38.284124 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:40:38.317304 (kubelet)[3455]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 17:40:38.548430 sudo[3492]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 4 17:40:38.548992 sudo[3492]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Sep 4 17:40:38.652494 kubelet[3455]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:40:38.655725 kubelet[3455]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 4 17:40:38.655725 kubelet[3455]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:40:38.655725 kubelet[3455]: I0904 17:40:38.653002 3455 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 17:40:38.720391 kubelet[3455]: I0904 17:40:38.715799 3455 server.go:467] "Kubelet version" kubeletVersion="v1.28.7" Sep 4 17:40:38.720391 kubelet[3455]: I0904 17:40:38.720346 3455 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 17:40:38.720794 kubelet[3455]: I0904 17:40:38.720770 3455 server.go:895] "Client rotation is on, will bootstrap in background" Sep 4 17:40:38.749383 kubelet[3455]: I0904 17:40:38.748002 3455 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 4 17:40:38.773728 kubelet[3455]: I0904 17:40:38.773698 3455 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 17:40:38.816758 kubelet[3455]: I0904 17:40:38.816654 3455 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 17:40:38.817604 kubelet[3455]: I0904 17:40:38.817585 3455 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 17:40:38.817963 kubelet[3455]: I0904 17:40:38.817946 3455 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Sep 4 17:40:38.818164 kubelet[3455]: I0904 17:40:38.818151 3455 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 17:40:38.818444 kubelet[3455]: I0904 17:40:38.818254 3455 container_manager_linux.go:301] "Creating device plugin manager" Sep 4 17:40:38.818444 kubelet[3455]: I0904 17:40:38.818325 3455 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:40:38.818640 kubelet[3455]: I0904 17:40:38.818604 3455 kubelet.go:393] "Attempting to sync node with API server" Sep 4 17:40:38.818945 kubelet[3455]: I0904 17:40:38.818827 3455 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 17:40:38.818945 kubelet[3455]: I0904 17:40:38.818866 3455 kubelet.go:309] "Adding apiserver pod source" Sep 4 17:40:38.818945 kubelet[3455]: I0904 17:40:38.818885 3455 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 17:40:38.822284 kubelet[3455]: I0904 17:40:38.821165 3455 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="v1.7.20" apiVersion="v1" Sep 4 17:40:38.822284 kubelet[3455]: I0904 17:40:38.821994 3455 server.go:1232] "Started kubelet" Sep 4 17:40:38.829959 kubelet[3455]: I0904 17:40:38.829929 3455 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 17:40:38.846197 kubelet[3455]: I0904 17:40:38.846165 3455 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 17:40:38.889372 kubelet[3455]: I0904 17:40:38.889313 3455 server.go:462] "Adding debug handlers to kubelet server" Sep 4 17:40:38.903758 kubelet[3455]: E0904 17:40:38.903723 3455 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Sep 4 17:40:38.903758 kubelet[3455]: E0904 17:40:38.903768 3455 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 17:40:38.908232 kubelet[3455]: I0904 17:40:38.846235 3455 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Sep 4 17:40:38.908232 kubelet[3455]: I0904 17:40:38.908197 3455 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 17:40:38.908232 kubelet[3455]: I0904 17:40:38.854279 3455 volume_manager.go:291] "Starting Kubelet Volume Manager" Sep 4 17:40:38.908487 kubelet[3455]: I0904 17:40:38.854300 3455 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Sep 4 17:40:38.908487 kubelet[3455]: I0904 17:40:38.908484 3455 reconciler_new.go:29] "Reconciler: start to sync state" Sep 4 17:40:38.999787 kubelet[3455]: I0904 17:40:38.999721 3455 kubelet_node_status.go:70] "Attempting to register node" node="ip-172-31-20-143" Sep 4 17:40:39.012210 kubelet[3455]: I0904 17:40:39.012089 3455 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 17:40:39.024279 kubelet[3455]: I0904 17:40:39.024070 3455 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 17:40:39.024279 kubelet[3455]: I0904 17:40:39.024116 3455 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 4 17:40:39.024279 kubelet[3455]: I0904 17:40:39.024175 3455 kubelet.go:2303] "Starting kubelet main sync loop" Sep 4 17:40:39.025564 kubelet[3455]: E0904 17:40:39.025176 3455 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 17:40:39.039218 kubelet[3455]: I0904 17:40:39.034813 3455 kubelet_node_status.go:108] "Node was previously registered" node="ip-172-31-20-143" Sep 4 17:40:39.039218 kubelet[3455]: I0904 17:40:39.035145 3455 kubelet_node_status.go:73] "Successfully registered node" node="ip-172-31-20-143" Sep 4 17:40:39.126376 kubelet[3455]: E0904 17:40:39.126337 3455 kubelet.go:2327] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 4 17:40:39.275162 kubelet[3455]: I0904 17:40:39.274717 3455 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 4 17:40:39.275162 kubelet[3455]: I0904 17:40:39.274749 3455 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 4 17:40:39.275162 kubelet[3455]: I0904 17:40:39.274774 3455 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:40:39.275390 kubelet[3455]: I0904 17:40:39.275209 3455 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 4 17:40:39.275390 kubelet[3455]: I0904 17:40:39.275260 3455 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 4 17:40:39.275390 kubelet[3455]: I0904 17:40:39.275271 3455 policy_none.go:49] "None policy: Start" Sep 4 17:40:39.283231 kubelet[3455]: I0904 17:40:39.279914 3455 memory_manager.go:169] "Starting memorymanager" policy="None" Sep 4 17:40:39.283231 kubelet[3455]: I0904 17:40:39.279952 3455 state_mem.go:35] "Initializing new in-memory state store" Sep 4 17:40:39.283231 kubelet[3455]: I0904 17:40:39.280432 3455 state_mem.go:75] "Updated machine memory state" Sep 4 17:40:39.288284 kubelet[3455]: I0904 17:40:39.284369 3455 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 17:40:39.289717 kubelet[3455]: I0904 17:40:39.289380 3455 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 17:40:39.329762 kubelet[3455]: I0904 17:40:39.329728 3455 topology_manager.go:215] "Topology Admit Handler" podUID="0a59b5a908b97dc850b40c944a13c186" podNamespace="kube-system" podName="kube-apiserver-ip-172-31-20-143" Sep 4 17:40:39.329910 kubelet[3455]: I0904 17:40:39.329855 3455 topology_manager.go:215] "Topology Admit Handler" podUID="44ef5d5ac47f06aa98b4f00afbce8a61" podNamespace="kube-system" podName="kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:39.331493 kubelet[3455]: I0904 17:40:39.331405 3455 topology_manager.go:215] "Topology Admit Handler" podUID="6b2842a9bd421343e3cb87d167be4d72" podNamespace="kube-system" podName="kube-scheduler-ip-172-31-20-143" Sep 4 17:40:39.351536 kubelet[3455]: E0904 17:40:39.350739 3455 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ip-172-31-20-143\" already exists" pod="kube-system/kube-scheduler-ip-172-31-20-143" Sep 4 17:40:39.421866 kubelet[3455]: I0904 17:40:39.420922 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:39.423358 kubelet[3455]: I0904 17:40:39.422941 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:39.423649 kubelet[3455]: I0904 17:40:39.423479 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6b2842a9bd421343e3cb87d167be4d72-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-143\" (UID: \"6b2842a9bd421343e3cb87d167be4d72\") " pod="kube-system/kube-scheduler-ip-172-31-20-143" Sep 4 17:40:39.423649 kubelet[3455]: I0904 17:40:39.423534 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0a59b5a908b97dc850b40c944a13c186-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-143\" (UID: \"0a59b5a908b97dc850b40c944a13c186\") " pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:39.423649 kubelet[3455]: I0904 17:40:39.423555 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0a59b5a908b97dc850b40c944a13c186-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-143\" (UID: \"0a59b5a908b97dc850b40c944a13c186\") " pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:39.423649 kubelet[3455]: I0904 17:40:39.423599 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:39.423649 kubelet[3455]: I0904 17:40:39.423619 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:39.423801 kubelet[3455]: I0904 17:40:39.423638 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0a59b5a908b97dc850b40c944a13c186-ca-certs\") pod \"kube-apiserver-ip-172-31-20-143\" (UID: \"0a59b5a908b97dc850b40c944a13c186\") " pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:39.423894 kubelet[3455]: I0904 17:40:39.423855 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/44ef5d5ac47f06aa98b4f00afbce8a61-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-143\" (UID: \"44ef5d5ac47f06aa98b4f00afbce8a61\") " pod="kube-system/kube-controller-manager-ip-172-31-20-143" Sep 4 17:40:39.824029 kubelet[3455]: I0904 17:40:39.823691 3455 apiserver.go:52] "Watching apiserver" Sep 4 17:40:39.909134 kubelet[3455]: I0904 17:40:39.909092 3455 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Sep 4 17:40:40.109324 sudo[3492]: pam_unix(sudo:session): session closed for user root Sep 4 17:40:40.154035 kubelet[3455]: E0904 17:40:40.153445 3455 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ip-172-31-20-143\" already exists" pod="kube-system/kube-apiserver-ip-172-31-20-143" Sep 4 17:40:40.176884 kubelet[3455]: I0904 17:40:40.176847 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-20-143" podStartSLOduration=4.17679893 podCreationTimestamp="2024-09-04 17:40:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:40:40.176508122 +0000 UTC m=+1.825716895" watchObservedRunningTime="2024-09-04 17:40:40.17679893 +0000 UTC m=+1.826007702" Sep 4 17:40:40.205502 kubelet[3455]: I0904 17:40:40.204839 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-20-143" podStartSLOduration=1.204774111 podCreationTimestamp="2024-09-04 17:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:40:40.190155987 +0000 UTC m=+1.839364756" watchObservedRunningTime="2024-09-04 17:40:40.204774111 +0000 UTC m=+1.853982880" Sep 4 17:40:40.205502 kubelet[3455]: I0904 17:40:40.204963 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-20-143" podStartSLOduration=1.204936146 podCreationTimestamp="2024-09-04 17:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:40:40.204876787 +0000 UTC m=+1.854085560" watchObservedRunningTime="2024-09-04 17:40:40.204936146 +0000 UTC m=+1.854144917" Sep 4 17:40:42.489327 sudo[2436]: pam_unix(sudo:session): session closed for user root Sep 4 17:40:42.515302 sshd[2430]: pam_unix(sshd:session): session closed for user core Sep 4 17:40:42.523041 systemd[1]: sshd@6-172.31.20.143:22-139.178.68.195:32922.service: Deactivated successfully. Sep 4 17:40:42.528688 systemd[1]: session-7.scope: Deactivated successfully. Sep 4 17:40:42.530537 systemd-logind[2045]: Session 7 logged out. Waiting for processes to exit. Sep 4 17:40:42.532086 systemd-logind[2045]: Removed session 7. Sep 4 17:40:50.156596 kubelet[3455]: I0904 17:40:50.156551 3455 topology_manager.go:215] "Topology Admit Handler" podUID="4f107353-ab1b-4d6a-8602-8ddafd64f039" podNamespace="kube-system" podName="kube-proxy-bsmvq" Sep 4 17:40:50.175279 kubelet[3455]: I0904 17:40:50.173301 3455 topology_manager.go:215] "Topology Admit Handler" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" podNamespace="kube-system" podName="cilium-rhgnt" Sep 4 17:40:50.184266 kubelet[3455]: I0904 17:40:50.182489 3455 kuberuntime_manager.go:1528] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 4 17:40:50.186026 containerd[2079]: time="2024-09-04T17:40:50.185976690Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 4 17:40:50.187596 kubelet[3455]: I0904 17:40:50.187084 3455 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 4 17:40:50.209264 kubelet[3455]: I0904 17:40:50.206793 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-bpf-maps\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209264 kubelet[3455]: I0904 17:40:50.206860 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/4f107353-ab1b-4d6a-8602-8ddafd64f039-kube-proxy\") pod \"kube-proxy-bsmvq\" (UID: \"4f107353-ab1b-4d6a-8602-8ddafd64f039\") " pod="kube-system/kube-proxy-bsmvq" Sep 4 17:40:50.209264 kubelet[3455]: I0904 17:40:50.206894 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-etc-cni-netd\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209264 kubelet[3455]: I0904 17:40:50.206939 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-lib-modules\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209264 kubelet[3455]: I0904 17:40:50.206968 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4f107353-ab1b-4d6a-8602-8ddafd64f039-lib-modules\") pod \"kube-proxy-bsmvq\" (UID: \"4f107353-ab1b-4d6a-8602-8ddafd64f039\") " pod="kube-system/kube-proxy-bsmvq" Sep 4 17:40:50.209264 kubelet[3455]: I0904 17:40:50.207010 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-xtables-lock\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209840 kubelet[3455]: I0904 17:40:50.207041 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-run\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209840 kubelet[3455]: I0904 17:40:50.207068 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-hubble-tls\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209840 kubelet[3455]: I0904 17:40:50.207114 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-cgroup\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209840 kubelet[3455]: I0904 17:40:50.207140 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/63a1952e-55aa-4c14-8998-d2fece4d52cc-clustermesh-secrets\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209840 kubelet[3455]: I0904 17:40:50.207209 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-config-path\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.209840 kubelet[3455]: I0904 17:40:50.207282 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5vdd\" (UniqueName: \"kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-kube-api-access-t5vdd\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.210024 kubelet[3455]: I0904 17:40:50.207316 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-net\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.210024 kubelet[3455]: I0904 17:40:50.207361 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cni-path\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.210024 kubelet[3455]: I0904 17:40:50.207390 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-kernel\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.210024 kubelet[3455]: I0904 17:40:50.207536 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4f107353-ab1b-4d6a-8602-8ddafd64f039-xtables-lock\") pod \"kube-proxy-bsmvq\" (UID: \"4f107353-ab1b-4d6a-8602-8ddafd64f039\") " pod="kube-system/kube-proxy-bsmvq" Sep 4 17:40:50.210024 kubelet[3455]: I0904 17:40:50.207586 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgs9w\" (UniqueName: \"kubernetes.io/projected/4f107353-ab1b-4d6a-8602-8ddafd64f039-kube-api-access-hgs9w\") pod \"kube-proxy-bsmvq\" (UID: \"4f107353-ab1b-4d6a-8602-8ddafd64f039\") " pod="kube-system/kube-proxy-bsmvq" Sep 4 17:40:50.210170 kubelet[3455]: I0904 17:40:50.207623 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-hostproc\") pod \"cilium-rhgnt\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " pod="kube-system/cilium-rhgnt" Sep 4 17:40:50.425054 kubelet[3455]: E0904 17:40:50.421013 3455 projected.go:292] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 4 17:40:50.425054 kubelet[3455]: E0904 17:40:50.421092 3455 projected.go:198] Error preparing data for projected volume kube-api-access-t5vdd for pod kube-system/cilium-rhgnt: configmap "kube-root-ca.crt" not found Sep 4 17:40:50.425054 kubelet[3455]: E0904 17:40:50.424690 3455 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-kube-api-access-t5vdd podName:63a1952e-55aa-4c14-8998-d2fece4d52cc nodeName:}" failed. No retries permitted until 2024-09-04 17:40:50.921393463 +0000 UTC m=+12.570602229 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-t5vdd" (UniqueName: "kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-kube-api-access-t5vdd") pod "cilium-rhgnt" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc") : configmap "kube-root-ca.crt" not found Sep 4 17:40:50.427959 kubelet[3455]: E0904 17:40:50.427931 3455 projected.go:292] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 4 17:40:50.430515 kubelet[3455]: E0904 17:40:50.430492 3455 projected.go:198] Error preparing data for projected volume kube-api-access-hgs9w for pod kube-system/kube-proxy-bsmvq: configmap "kube-root-ca.crt" not found Sep 4 17:40:50.430752 kubelet[3455]: E0904 17:40:50.430710 3455 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f107353-ab1b-4d6a-8602-8ddafd64f039-kube-api-access-hgs9w podName:4f107353-ab1b-4d6a-8602-8ddafd64f039 nodeName:}" failed. No retries permitted until 2024-09-04 17:40:50.930681681 +0000 UTC m=+12.579890450 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hgs9w" (UniqueName: "kubernetes.io/projected/4f107353-ab1b-4d6a-8602-8ddafd64f039-kube-api-access-hgs9w") pod "kube-proxy-bsmvq" (UID: "4f107353-ab1b-4d6a-8602-8ddafd64f039") : configmap "kube-root-ca.crt" not found Sep 4 17:40:50.466986 kubelet[3455]: I0904 17:40:50.462821 3455 topology_manager.go:215] "Topology Admit Handler" podUID="1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e" podNamespace="kube-system" podName="cilium-operator-6bc8ccdb58-bnw7c" Sep 4 17:40:50.533840 kubelet[3455]: I0904 17:40:50.528161 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-cilium-config-path\") pod \"cilium-operator-6bc8ccdb58-bnw7c\" (UID: \"1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e\") " pod="kube-system/cilium-operator-6bc8ccdb58-bnw7c" Sep 4 17:40:50.534606 kubelet[3455]: I0904 17:40:50.534584 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk7nc\" (UniqueName: \"kubernetes.io/projected/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-kube-api-access-vk7nc\") pod \"cilium-operator-6bc8ccdb58-bnw7c\" (UID: \"1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e\") " pod="kube-system/cilium-operator-6bc8ccdb58-bnw7c" Sep 4 17:40:50.778101 containerd[2079]: time="2024-09-04T17:40:50.777627331Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-bnw7c,Uid:1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e,Namespace:kube-system,Attempt:0,}" Sep 4 17:40:50.843159 containerd[2079]: time="2024-09-04T17:40:50.842792178Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:40:50.843362 containerd[2079]: time="2024-09-04T17:40:50.843136072Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:40:50.843362 containerd[2079]: time="2024-09-04T17:40:50.843162215Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:50.843572 containerd[2079]: time="2024-09-04T17:40:50.843407061Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:50.973531 containerd[2079]: time="2024-09-04T17:40:50.973484987Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-bnw7c,Uid:1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e,Namespace:kube-system,Attempt:0,} returns sandbox id \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\"" Sep 4 17:40:50.976532 containerd[2079]: time="2024-09-04T17:40:50.976497696Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 4 17:40:51.086044 containerd[2079]: time="2024-09-04T17:40:51.085907886Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bsmvq,Uid:4f107353-ab1b-4d6a-8602-8ddafd64f039,Namespace:kube-system,Attempt:0,}" Sep 4 17:40:51.108456 containerd[2079]: time="2024-09-04T17:40:51.107350405Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-rhgnt,Uid:63a1952e-55aa-4c14-8998-d2fece4d52cc,Namespace:kube-system,Attempt:0,}" Sep 4 17:40:51.137830 containerd[2079]: time="2024-09-04T17:40:51.137691174Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:40:51.137830 containerd[2079]: time="2024-09-04T17:40:51.137771714Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:40:51.137830 containerd[2079]: time="2024-09-04T17:40:51.137794177Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:51.140016 containerd[2079]: time="2024-09-04T17:40:51.137928604Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:51.176589 containerd[2079]: time="2024-09-04T17:40:51.176093507Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:40:51.176589 containerd[2079]: time="2024-09-04T17:40:51.176197924Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:40:51.176589 containerd[2079]: time="2024-09-04T17:40:51.176223902Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:51.179042 containerd[2079]: time="2024-09-04T17:40:51.178914619Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:40:51.228203 containerd[2079]: time="2024-09-04T17:40:51.228151223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bsmvq,Uid:4f107353-ab1b-4d6a-8602-8ddafd64f039,Namespace:kube-system,Attempt:0,} returns sandbox id \"86e090c05e14c1210a013e19e33f49e23fc4d6c5d252cb0422509e227146de18\"" Sep 4 17:40:51.244935 containerd[2079]: time="2024-09-04T17:40:51.241938661Z" level=info msg="CreateContainer within sandbox \"86e090c05e14c1210a013e19e33f49e23fc4d6c5d252cb0422509e227146de18\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 4 17:40:51.278204 containerd[2079]: time="2024-09-04T17:40:51.278083397Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-rhgnt,Uid:63a1952e-55aa-4c14-8998-d2fece4d52cc,Namespace:kube-system,Attempt:0,} returns sandbox id \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\"" Sep 4 17:40:51.283710 containerd[2079]: time="2024-09-04T17:40:51.283597662Z" level=info msg="CreateContainer within sandbox \"86e090c05e14c1210a013e19e33f49e23fc4d6c5d252cb0422509e227146de18\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"cf909000742ecb0ddfc1303bfd74e69bd18a8191bb0a7781ed3e3206ce3ded22\"" Sep 4 17:40:51.286417 containerd[2079]: time="2024-09-04T17:40:51.284649470Z" level=info msg="StartContainer for \"cf909000742ecb0ddfc1303bfd74e69bd18a8191bb0a7781ed3e3206ce3ded22\"" Sep 4 17:40:51.439220 containerd[2079]: time="2024-09-04T17:40:51.439186811Z" level=info msg="StartContainer for \"cf909000742ecb0ddfc1303bfd74e69bd18a8191bb0a7781ed3e3206ce3ded22\" returns successfully" Sep 4 17:40:52.239892 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount574295161.mount: Deactivated successfully. Sep 4 17:40:52.445795 kubelet[3455]: I0904 17:40:52.441458 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-bsmvq" podStartSLOduration=2.441400572 podCreationTimestamp="2024-09-04 17:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:40:52.43379159 +0000 UTC m=+14.083000365" watchObservedRunningTime="2024-09-04 17:40:52.441400572 +0000 UTC m=+14.090609345" Sep 4 17:40:53.437990 containerd[2079]: time="2024-09-04T17:40:53.437933647Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:53.439270 containerd[2079]: time="2024-09-04T17:40:53.439128709Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18907217" Sep 4 17:40:53.440785 containerd[2079]: time="2024-09-04T17:40:53.440729138Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:40:53.442558 containerd[2079]: time="2024-09-04T17:40:53.442368755Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 2.465828002s" Sep 4 17:40:53.442558 containerd[2079]: time="2024-09-04T17:40:53.442415466Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 4 17:40:53.444265 containerd[2079]: time="2024-09-04T17:40:53.444182314Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 4 17:40:53.445496 containerd[2079]: time="2024-09-04T17:40:53.445463253Z" level=info msg="CreateContainer within sandbox \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 4 17:40:53.466270 containerd[2079]: time="2024-09-04T17:40:53.466208127Z" level=info msg="CreateContainer within sandbox \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\"" Sep 4 17:40:53.468335 containerd[2079]: time="2024-09-04T17:40:53.468052866Z" level=info msg="StartContainer for \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\"" Sep 4 17:40:53.554621 containerd[2079]: time="2024-09-04T17:40:53.554568346Z" level=info msg="StartContainer for \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\" returns successfully" Sep 4 17:40:54.402534 systemd-resolved[1963]: Under memory pressure, flushing caches. Sep 4 17:40:54.402612 systemd-resolved[1963]: Flushed all caches. Sep 4 17:40:54.406268 systemd-journald[1562]: Under memory pressure, flushing caches. Sep 4 17:40:59.058999 kubelet[3455]: I0904 17:40:59.058307 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-6bc8ccdb58-bnw7c" podStartSLOduration=6.590133596 podCreationTimestamp="2024-09-04 17:40:50 +0000 UTC" firstStartedPulling="2024-09-04 17:40:50.974859054 +0000 UTC m=+12.624067820" lastFinishedPulling="2024-09-04 17:40:53.442960714 +0000 UTC m=+15.092169483" observedRunningTime="2024-09-04 17:40:54.521309858 +0000 UTC m=+16.170518627" watchObservedRunningTime="2024-09-04 17:40:59.058235259 +0000 UTC m=+20.707444032" Sep 4 17:41:00.854189 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2600297780.mount: Deactivated successfully. Sep 4 17:41:05.387174 containerd[2079]: time="2024-09-04T17:41:05.387114101Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:05.389848 containerd[2079]: time="2024-09-04T17:41:05.389762390Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166735291" Sep 4 17:41:05.390364 containerd[2079]: time="2024-09-04T17:41:05.390332364Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:41:05.392222 containerd[2079]: time="2024-09-04T17:41:05.392077692Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 11.947839411s" Sep 4 17:41:05.392222 containerd[2079]: time="2024-09-04T17:41:05.392122602Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 4 17:41:05.401842 containerd[2079]: time="2024-09-04T17:41:05.401790792Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 4 17:41:05.497059 containerd[2079]: time="2024-09-04T17:41:05.497000696Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\"" Sep 4 17:41:05.497870 containerd[2079]: time="2024-09-04T17:41:05.497837573Z" level=info msg="StartContainer for \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\"" Sep 4 17:41:05.818852 containerd[2079]: time="2024-09-04T17:41:05.815467551Z" level=info msg="StartContainer for \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\" returns successfully" Sep 4 17:41:05.944054 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681-rootfs.mount: Deactivated successfully. Sep 4 17:41:06.165605 containerd[2079]: time="2024-09-04T17:41:06.121070470Z" level=info msg="shim disconnected" id=881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681 namespace=k8s.io Sep 4 17:41:06.165605 containerd[2079]: time="2024-09-04T17:41:06.165603827Z" level=warning msg="cleaning up after shim disconnected" id=881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681 namespace=k8s.io Sep 4 17:41:06.165977 containerd[2079]: time="2024-09-04T17:41:06.165626954Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:41:06.425075 containerd[2079]: time="2024-09-04T17:41:06.423720200Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 4 17:41:06.517112 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3951979762.mount: Deactivated successfully. Sep 4 17:41:06.525542 containerd[2079]: time="2024-09-04T17:41:06.525492996Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\"" Sep 4 17:41:06.527490 containerd[2079]: time="2024-09-04T17:41:06.526467176Z" level=info msg="StartContainer for \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\"" Sep 4 17:41:06.652859 containerd[2079]: time="2024-09-04T17:41:06.652810896Z" level=info msg="StartContainer for \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\" returns successfully" Sep 4 17:41:06.676963 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 4 17:41:06.678695 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:41:06.678795 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:41:06.693093 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:41:06.750471 containerd[2079]: time="2024-09-04T17:41:06.750378640Z" level=info msg="shim disconnected" id=aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa namespace=k8s.io Sep 4 17:41:06.750471 containerd[2079]: time="2024-09-04T17:41:06.750454975Z" level=warning msg="cleaning up after shim disconnected" id=aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa namespace=k8s.io Sep 4 17:41:06.750471 containerd[2079]: time="2024-09-04T17:41:06.750467871Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:41:06.779978 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:41:06.788397 containerd[2079]: time="2024-09-04T17:41:06.788234775Z" level=warning msg="cleanup warnings time=\"2024-09-04T17:41:06Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 4 17:41:07.406503 containerd[2079]: time="2024-09-04T17:41:07.406320603Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 4 17:41:07.452503 containerd[2079]: time="2024-09-04T17:41:07.452453524Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\"" Sep 4 17:41:07.454217 containerd[2079]: time="2024-09-04T17:41:07.453231047Z" level=info msg="StartContainer for \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\"" Sep 4 17:41:07.492469 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa-rootfs.mount: Deactivated successfully. Sep 4 17:41:07.573488 containerd[2079]: time="2024-09-04T17:41:07.573440827Z" level=info msg="StartContainer for \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\" returns successfully" Sep 4 17:41:07.622683 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58-rootfs.mount: Deactivated successfully. Sep 4 17:41:07.649172 containerd[2079]: time="2024-09-04T17:41:07.648857730Z" level=info msg="shim disconnected" id=3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58 namespace=k8s.io Sep 4 17:41:07.649172 containerd[2079]: time="2024-09-04T17:41:07.648970470Z" level=warning msg="cleaning up after shim disconnected" id=3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58 namespace=k8s.io Sep 4 17:41:07.649172 containerd[2079]: time="2024-09-04T17:41:07.648984337Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:41:08.434294 containerd[2079]: time="2024-09-04T17:41:08.433894760Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 4 17:41:08.492029 containerd[2079]: time="2024-09-04T17:41:08.490944750Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\"" Sep 4 17:41:08.500998 containerd[2079]: time="2024-09-04T17:41:08.499896319Z" level=info msg="StartContainer for \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\"" Sep 4 17:41:08.500561 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1016648249.mount: Deactivated successfully. Sep 4 17:41:08.609254 containerd[2079]: time="2024-09-04T17:41:08.608136484Z" level=info msg="StartContainer for \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\" returns successfully" Sep 4 17:41:08.628209 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d-rootfs.mount: Deactivated successfully. Sep 4 17:41:08.637053 containerd[2079]: time="2024-09-04T17:41:08.636904644Z" level=info msg="shim disconnected" id=ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d namespace=k8s.io Sep 4 17:41:08.637303 containerd[2079]: time="2024-09-04T17:41:08.637045497Z" level=warning msg="cleaning up after shim disconnected" id=ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d namespace=k8s.io Sep 4 17:41:08.637303 containerd[2079]: time="2024-09-04T17:41:08.637092234Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:41:08.657292 containerd[2079]: time="2024-09-04T17:41:08.656701120Z" level=warning msg="cleanup warnings time=\"2024-09-04T17:41:08Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Sep 4 17:41:09.422270 containerd[2079]: time="2024-09-04T17:41:09.422210925Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 4 17:41:09.481184 containerd[2079]: time="2024-09-04T17:41:09.481083710Z" level=info msg="CreateContainer within sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\"" Sep 4 17:41:09.487530 containerd[2079]: time="2024-09-04T17:41:09.487473537Z" level=info msg="StartContainer for \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\"" Sep 4 17:41:09.614844 containerd[2079]: time="2024-09-04T17:41:09.608285070Z" level=info msg="StartContainer for \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\" returns successfully" Sep 4 17:41:10.198007 kubelet[3455]: I0904 17:41:10.196445 3455 kubelet_node_status.go:493] "Fast updating node status as it just became ready" Sep 4 17:41:10.242348 kubelet[3455]: I0904 17:41:10.240748 3455 topology_manager.go:215] "Topology Admit Handler" podUID="e9ebf3ef-0547-4332-846a-0d5bf557b87d" podNamespace="kube-system" podName="coredns-5dd5756b68-bgp4z" Sep 4 17:41:10.249022 kubelet[3455]: I0904 17:41:10.248326 3455 topology_manager.go:215] "Topology Admit Handler" podUID="65db96a2-61d5-43db-9fe6-bfe2798e3644" podNamespace="kube-system" podName="coredns-5dd5756b68-x9jwk" Sep 4 17:41:10.422666 kubelet[3455]: I0904 17:41:10.422631 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/65db96a2-61d5-43db-9fe6-bfe2798e3644-config-volume\") pod \"coredns-5dd5756b68-x9jwk\" (UID: \"65db96a2-61d5-43db-9fe6-bfe2798e3644\") " pod="kube-system/coredns-5dd5756b68-x9jwk" Sep 4 17:41:10.423228 kubelet[3455]: I0904 17:41:10.423202 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q95ff\" (UniqueName: \"kubernetes.io/projected/e9ebf3ef-0547-4332-846a-0d5bf557b87d-kube-api-access-q95ff\") pod \"coredns-5dd5756b68-bgp4z\" (UID: \"e9ebf3ef-0547-4332-846a-0d5bf557b87d\") " pod="kube-system/coredns-5dd5756b68-bgp4z" Sep 4 17:41:10.423370 kubelet[3455]: I0904 17:41:10.423270 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2k4r\" (UniqueName: \"kubernetes.io/projected/65db96a2-61d5-43db-9fe6-bfe2798e3644-kube-api-access-z2k4r\") pod \"coredns-5dd5756b68-x9jwk\" (UID: \"65db96a2-61d5-43db-9fe6-bfe2798e3644\") " pod="kube-system/coredns-5dd5756b68-x9jwk" Sep 4 17:41:10.423370 kubelet[3455]: I0904 17:41:10.423308 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9ebf3ef-0547-4332-846a-0d5bf557b87d-config-volume\") pod \"coredns-5dd5756b68-bgp4z\" (UID: \"e9ebf3ef-0547-4332-846a-0d5bf557b87d\") " pod="kube-system/coredns-5dd5756b68-bgp4z" Sep 4 17:41:10.460362 kubelet[3455]: I0904 17:41:10.460235 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-rhgnt" podStartSLOduration=6.347467867 podCreationTimestamp="2024-09-04 17:40:50 +0000 UTC" firstStartedPulling="2024-09-04 17:40:51.279844493 +0000 UTC m=+12.929053245" lastFinishedPulling="2024-09-04 17:41:05.392562128 +0000 UTC m=+27.041770890" observedRunningTime="2024-09-04 17:41:10.459576429 +0000 UTC m=+32.108785225" watchObservedRunningTime="2024-09-04 17:41:10.460185512 +0000 UTC m=+32.109394289" Sep 4 17:41:10.466352 systemd-resolved[1963]: Under memory pressure, flushing caches. Sep 4 17:41:10.468335 systemd-journald[1562]: Under memory pressure, flushing caches. Sep 4 17:41:10.466398 systemd-resolved[1963]: Flushed all caches. Sep 4 17:41:10.561586 containerd[2079]: time="2024-09-04T17:41:10.559956610Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-bgp4z,Uid:e9ebf3ef-0547-4332-846a-0d5bf557b87d,Namespace:kube-system,Attempt:0,}" Sep 4 17:41:10.581144 containerd[2079]: time="2024-09-04T17:41:10.580551474Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-x9jwk,Uid:65db96a2-61d5-43db-9fe6-bfe2798e3644,Namespace:kube-system,Attempt:0,}" Sep 4 17:41:12.377958 systemd-networkd[1648]: cilium_host: Link UP Sep 4 17:41:12.378143 systemd-networkd[1648]: cilium_net: Link UP Sep 4 17:41:12.378148 systemd-networkd[1648]: cilium_net: Gained carrier Sep 4 17:41:12.383833 (udev-worker)[4295]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:41:12.386837 systemd-networkd[1648]: cilium_host: Gained carrier Sep 4 17:41:12.387520 systemd-networkd[1648]: cilium_host: Gained IPv6LL Sep 4 17:41:12.400352 (udev-worker)[4359]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:41:12.558094 (udev-worker)[4385]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:41:12.568183 systemd-networkd[1648]: cilium_vxlan: Link UP Sep 4 17:41:12.568193 systemd-networkd[1648]: cilium_vxlan: Gained carrier Sep 4 17:41:12.634406 systemd-networkd[1648]: cilium_net: Gained IPv6LL Sep 4 17:41:13.497465 kernel: NET: Registered PF_ALG protocol family Sep 4 17:41:13.858491 systemd-networkd[1648]: cilium_vxlan: Gained IPv6LL Sep 4 17:41:14.471659 systemd-networkd[1648]: lxc_health: Link UP Sep 4 17:41:14.476101 systemd-networkd[1648]: lxc_health: Gained carrier Sep 4 17:41:15.284650 systemd-networkd[1648]: lxc7279d704b53a: Link UP Sep 4 17:41:15.294147 kernel: eth0: renamed from tmpf062e Sep 4 17:41:15.303876 (udev-worker)[4384]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:41:15.308400 systemd-networkd[1648]: lxc7279d704b53a: Gained carrier Sep 4 17:41:15.319349 systemd-networkd[1648]: lxc1d835ff61e6e: Link UP Sep 4 17:41:15.330235 kernel: eth0: renamed from tmpdb60c Sep 4 17:41:15.341033 systemd-networkd[1648]: lxc1d835ff61e6e: Gained carrier Sep 4 17:41:16.228335 systemd-networkd[1648]: lxc_health: Gained IPv6LL Sep 4 17:41:16.738781 systemd-networkd[1648]: lxc1d835ff61e6e: Gained IPv6LL Sep 4 17:41:17.061098 systemd-networkd[1648]: lxc7279d704b53a: Gained IPv6LL Sep 4 17:41:19.059354 ntpd[2033]: Listen normally on 6 cilium_host 192.168.0.109:123 Sep 4 17:41:19.059923 ntpd[2033]: Listen normally on 7 cilium_net [fe80::5cfb:2fff:feb0:98f5%4]:123 Sep 4 17:41:19.061064 ntpd[2033]: 4 Sep 17:41:19 ntpd[2033]: Listen normally on 6 cilium_host 192.168.0.109:123 Sep 4 17:41:19.061064 ntpd[2033]: 4 Sep 17:41:19 ntpd[2033]: Listen normally on 7 cilium_net [fe80::5cfb:2fff:feb0:98f5%4]:123 Sep 4 17:41:19.061064 ntpd[2033]: 4 Sep 17:41:19 ntpd[2033]: Listen normally on 8 cilium_host [fe80::38eb:e3ff:fef3:595%5]:123 Sep 4 17:41:19.061064 ntpd[2033]: 4 Sep 17:41:19 ntpd[2033]: Listen normally on 9 cilium_vxlan [fe80::7c85:31ff:fe60:16d6%6]:123 Sep 4 17:41:19.061064 ntpd[2033]: 4 Sep 17:41:19 ntpd[2033]: Listen normally on 10 lxc_health [fe80::e0d2:2eff:feca:184%8]:123 Sep 4 17:41:19.061064 ntpd[2033]: 4 Sep 17:41:19 ntpd[2033]: Listen normally on 11 lxc1d835ff61e6e [fe80::fc92:8dff:fee1:c330%10]:123 Sep 4 17:41:19.061064 ntpd[2033]: 4 Sep 17:41:19 ntpd[2033]: Listen normally on 12 lxc7279d704b53a [fe80::3485:d8ff:fe74:36cd%12]:123 Sep 4 17:41:19.059979 ntpd[2033]: Listen normally on 8 cilium_host [fe80::38eb:e3ff:fef3:595%5]:123 Sep 4 17:41:19.060020 ntpd[2033]: Listen normally on 9 cilium_vxlan [fe80::7c85:31ff:fe60:16d6%6]:123 Sep 4 17:41:19.060059 ntpd[2033]: Listen normally on 10 lxc_health [fe80::e0d2:2eff:feca:184%8]:123 Sep 4 17:41:19.060106 ntpd[2033]: Listen normally on 11 lxc1d835ff61e6e [fe80::fc92:8dff:fee1:c330%10]:123 Sep 4 17:41:19.060142 ntpd[2033]: Listen normally on 12 lxc7279d704b53a [fe80::3485:d8ff:fe74:36cd%12]:123 Sep 4 17:41:21.247066 containerd[2079]: time="2024-09-04T17:41:21.223925468Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:41:21.247066 containerd[2079]: time="2024-09-04T17:41:21.224044272Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:41:21.247066 containerd[2079]: time="2024-09-04T17:41:21.224069426Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:41:21.247066 containerd[2079]: time="2024-09-04T17:41:21.224196239Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:41:21.320960 containerd[2079]: time="2024-09-04T17:41:21.318185005Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:41:21.320960 containerd[2079]: time="2024-09-04T17:41:21.319749549Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:41:21.320960 containerd[2079]: time="2024-09-04T17:41:21.319775733Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:41:21.320960 containerd[2079]: time="2024-09-04T17:41:21.320162517Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:41:21.441495 systemd[1]: run-containerd-runc-k8s.io-db60c380a08f58178c48833f0c61c696fb9be36055b3b4aadde3089c6aba8b00-runc.gGqYCU.mount: Deactivated successfully. Sep 4 17:41:21.468998 containerd[2079]: time="2024-09-04T17:41:21.468266585Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-x9jwk,Uid:65db96a2-61d5-43db-9fe6-bfe2798e3644,Namespace:kube-system,Attempt:0,} returns sandbox id \"f062e49ef092f61252c60baba648a2b035007627e29359cf98cc493c48fdb751\"" Sep 4 17:41:21.548588 containerd[2079]: time="2024-09-04T17:41:21.546502071Z" level=info msg="CreateContainer within sandbox \"f062e49ef092f61252c60baba648a2b035007627e29359cf98cc493c48fdb751\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 17:41:21.549811 containerd[2079]: time="2024-09-04T17:41:21.549774822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-bgp4z,Uid:e9ebf3ef-0547-4332-846a-0d5bf557b87d,Namespace:kube-system,Attempt:0,} returns sandbox id \"db60c380a08f58178c48833f0c61c696fb9be36055b3b4aadde3089c6aba8b00\"" Sep 4 17:41:21.587179 containerd[2079]: time="2024-09-04T17:41:21.587038280Z" level=info msg="CreateContainer within sandbox \"db60c380a08f58178c48833f0c61c696fb9be36055b3b4aadde3089c6aba8b00\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 17:41:21.591605 containerd[2079]: time="2024-09-04T17:41:21.591572185Z" level=info msg="CreateContainer within sandbox \"f062e49ef092f61252c60baba648a2b035007627e29359cf98cc493c48fdb751\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"7c54273cd492404dc85ad26d8110cd1fd30d39ddbb94ea1ba26eeeb17e260fc9\"" Sep 4 17:41:21.606560 containerd[2079]: time="2024-09-04T17:41:21.606518948Z" level=info msg="StartContainer for \"7c54273cd492404dc85ad26d8110cd1fd30d39ddbb94ea1ba26eeeb17e260fc9\"" Sep 4 17:41:21.612840 containerd[2079]: time="2024-09-04T17:41:21.612792487Z" level=info msg="CreateContainer within sandbox \"db60c380a08f58178c48833f0c61c696fb9be36055b3b4aadde3089c6aba8b00\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"0159395689d2a6667d8eef09fdd78bb146b2cb83bfdb0bbaa4d15a07963a035f\"" Sep 4 17:41:21.618370 containerd[2079]: time="2024-09-04T17:41:21.617965940Z" level=info msg="StartContainer for \"0159395689d2a6667d8eef09fdd78bb146b2cb83bfdb0bbaa4d15a07963a035f\"" Sep 4 17:41:21.778435 containerd[2079]: time="2024-09-04T17:41:21.777453490Z" level=info msg="StartContainer for \"7c54273cd492404dc85ad26d8110cd1fd30d39ddbb94ea1ba26eeeb17e260fc9\" returns successfully" Sep 4 17:41:21.799696 containerd[2079]: time="2024-09-04T17:41:21.799566846Z" level=info msg="StartContainer for \"0159395689d2a6667d8eef09fdd78bb146b2cb83bfdb0bbaa4d15a07963a035f\" returns successfully" Sep 4 17:41:22.546057 kubelet[3455]: I0904 17:41:22.534650 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-x9jwk" podStartSLOduration=32.534596436 podCreationTimestamp="2024-09-04 17:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:41:22.525968118 +0000 UTC m=+44.175176891" watchObservedRunningTime="2024-09-04 17:41:22.534596436 +0000 UTC m=+44.183805207" Sep 4 17:41:22.578141 kubelet[3455]: I0904 17:41:22.577887 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-bgp4z" podStartSLOduration=32.577838737 podCreationTimestamp="2024-09-04 17:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:41:22.577453499 +0000 UTC m=+44.226662272" watchObservedRunningTime="2024-09-04 17:41:22.577838737 +0000 UTC m=+44.227047509" Sep 4 17:41:24.851885 systemd[1]: Started sshd@7-172.31.20.143:22-139.178.68.195:55320.service - OpenSSH per-connection server daemon (139.178.68.195:55320). Sep 4 17:41:25.090918 sshd[4886]: Accepted publickey for core from 139.178.68.195 port 55320 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:25.097105 sshd[4886]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:25.156549 systemd-logind[2045]: New session 8 of user core. Sep 4 17:41:25.164677 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 4 17:41:26.190850 sshd[4886]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:26.195604 systemd[1]: sshd@7-172.31.20.143:22-139.178.68.195:55320.service: Deactivated successfully. Sep 4 17:41:26.202261 systemd-logind[2045]: Session 8 logged out. Waiting for processes to exit. Sep 4 17:41:26.203038 systemd[1]: session-8.scope: Deactivated successfully. Sep 4 17:41:26.205754 systemd-logind[2045]: Removed session 8. Sep 4 17:41:31.221997 systemd[1]: Started sshd@8-172.31.20.143:22-139.178.68.195:53920.service - OpenSSH per-connection server daemon (139.178.68.195:53920). Sep 4 17:41:31.396776 sshd[4902]: Accepted publickey for core from 139.178.68.195 port 53920 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:31.398491 sshd[4902]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:31.406209 systemd-logind[2045]: New session 9 of user core. Sep 4 17:41:31.410717 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 4 17:41:31.679000 sshd[4902]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:31.684743 systemd[1]: sshd@8-172.31.20.143:22-139.178.68.195:53920.service: Deactivated successfully. Sep 4 17:41:31.699298 systemd-logind[2045]: Session 9 logged out. Waiting for processes to exit. Sep 4 17:41:31.699679 systemd[1]: session-9.scope: Deactivated successfully. Sep 4 17:41:31.701597 systemd-logind[2045]: Removed session 9. Sep 4 17:41:36.710786 systemd[1]: Started sshd@9-172.31.20.143:22-139.178.68.195:43948.service - OpenSSH per-connection server daemon (139.178.68.195:43948). Sep 4 17:41:36.894193 sshd[4917]: Accepted publickey for core from 139.178.68.195 port 43948 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:36.895976 sshd[4917]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:36.903040 systemd-logind[2045]: New session 10 of user core. Sep 4 17:41:36.916838 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 4 17:41:37.189373 sshd[4917]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:37.193687 systemd[1]: sshd@9-172.31.20.143:22-139.178.68.195:43948.service: Deactivated successfully. Sep 4 17:41:37.204836 systemd[1]: session-10.scope: Deactivated successfully. Sep 4 17:41:37.206814 systemd-logind[2045]: Session 10 logged out. Waiting for processes to exit. Sep 4 17:41:37.208652 systemd-logind[2045]: Removed session 10. Sep 4 17:41:42.219640 systemd[1]: Started sshd@10-172.31.20.143:22-139.178.68.195:43952.service - OpenSSH per-connection server daemon (139.178.68.195:43952). Sep 4 17:41:42.395768 sshd[4934]: Accepted publickey for core from 139.178.68.195 port 43952 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:42.397915 sshd[4934]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:42.406882 systemd-logind[2045]: New session 11 of user core. Sep 4 17:41:42.415640 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 4 17:41:42.634837 sshd[4934]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:42.644784 systemd[1]: sshd@10-172.31.20.143:22-139.178.68.195:43952.service: Deactivated successfully. Sep 4 17:41:42.649214 systemd[1]: session-11.scope: Deactivated successfully. Sep 4 17:41:42.651655 systemd-logind[2045]: Session 11 logged out. Waiting for processes to exit. Sep 4 17:41:42.653102 systemd-logind[2045]: Removed session 11. Sep 4 17:41:42.665676 systemd[1]: Started sshd@11-172.31.20.143:22-139.178.68.195:43956.service - OpenSSH per-connection server daemon (139.178.68.195:43956). Sep 4 17:41:42.861420 sshd[4949]: Accepted publickey for core from 139.178.68.195 port 43956 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:42.862104 sshd[4949]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:42.867714 systemd-logind[2045]: New session 12 of user core. Sep 4 17:41:42.873951 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 4 17:41:44.331435 sshd[4949]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:44.371870 systemd[1]: Started sshd@12-172.31.20.143:22-139.178.68.195:43964.service - OpenSSH per-connection server daemon (139.178.68.195:43964). Sep 4 17:41:44.375330 systemd[1]: sshd@11-172.31.20.143:22-139.178.68.195:43956.service: Deactivated successfully. Sep 4 17:41:44.391510 systemd[1]: session-12.scope: Deactivated successfully. Sep 4 17:41:44.415773 systemd-logind[2045]: Session 12 logged out. Waiting for processes to exit. Sep 4 17:41:44.428992 systemd-logind[2045]: Removed session 12. Sep 4 17:41:44.597954 sshd[4958]: Accepted publickey for core from 139.178.68.195 port 43964 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:44.601062 sshd[4958]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:44.608603 systemd-logind[2045]: New session 13 of user core. Sep 4 17:41:44.613612 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 4 17:41:44.873104 sshd[4958]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:44.878873 systemd[1]: sshd@12-172.31.20.143:22-139.178.68.195:43964.service: Deactivated successfully. Sep 4 17:41:44.885723 systemd[1]: session-13.scope: Deactivated successfully. Sep 4 17:41:44.889369 systemd-logind[2045]: Session 13 logged out. Waiting for processes to exit. Sep 4 17:41:44.891351 systemd-logind[2045]: Removed session 13. Sep 4 17:41:49.910644 systemd[1]: Started sshd@13-172.31.20.143:22-139.178.68.195:57924.service - OpenSSH per-connection server daemon (139.178.68.195:57924). Sep 4 17:41:50.111124 sshd[4976]: Accepted publickey for core from 139.178.68.195 port 57924 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:50.114013 sshd[4976]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:50.130105 systemd-logind[2045]: New session 14 of user core. Sep 4 17:41:50.136719 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 4 17:41:50.352775 sshd[4976]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:50.358164 systemd[1]: sshd@13-172.31.20.143:22-139.178.68.195:57924.service: Deactivated successfully. Sep 4 17:41:50.363824 systemd-logind[2045]: Session 14 logged out. Waiting for processes to exit. Sep 4 17:41:50.365211 systemd[1]: session-14.scope: Deactivated successfully. Sep 4 17:41:50.368306 systemd-logind[2045]: Removed session 14. Sep 4 17:41:55.380216 systemd[1]: Started sshd@14-172.31.20.143:22-139.178.68.195:57940.service - OpenSSH per-connection server daemon (139.178.68.195:57940). Sep 4 17:41:55.539181 sshd[4992]: Accepted publickey for core from 139.178.68.195 port 57940 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:41:55.540985 sshd[4992]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:41:55.547629 systemd-logind[2045]: New session 15 of user core. Sep 4 17:41:55.552647 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 4 17:41:55.745877 sshd[4992]: pam_unix(sshd:session): session closed for user core Sep 4 17:41:55.749823 systemd[1]: sshd@14-172.31.20.143:22-139.178.68.195:57940.service: Deactivated successfully. Sep 4 17:41:55.756311 systemd[1]: session-15.scope: Deactivated successfully. Sep 4 17:41:55.758636 systemd-logind[2045]: Session 15 logged out. Waiting for processes to exit. Sep 4 17:41:55.760149 systemd-logind[2045]: Removed session 15. Sep 4 17:42:00.777965 systemd[1]: Started sshd@15-172.31.20.143:22-139.178.68.195:35126.service - OpenSSH per-connection server daemon (139.178.68.195:35126). Sep 4 17:42:00.948042 sshd[5007]: Accepted publickey for core from 139.178.68.195 port 35126 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:00.957334 sshd[5007]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:00.965159 systemd-logind[2045]: New session 16 of user core. Sep 4 17:42:00.971977 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 4 17:42:01.499738 sshd[5007]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:01.518500 systemd[1]: sshd@15-172.31.20.143:22-139.178.68.195:35126.service: Deactivated successfully. Sep 4 17:42:01.555331 systemd[1]: session-16.scope: Deactivated successfully. Sep 4 17:42:01.574557 systemd-logind[2045]: Session 16 logged out. Waiting for processes to exit. Sep 4 17:42:01.616598 systemd[1]: Started sshd@16-172.31.20.143:22-139.178.68.195:35142.service - OpenSSH per-connection server daemon (139.178.68.195:35142). Sep 4 17:42:01.629985 systemd-logind[2045]: Removed session 16. Sep 4 17:42:01.908787 sshd[5022]: Accepted publickey for core from 139.178.68.195 port 35142 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:01.922599 sshd[5022]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:02.001301 systemd-logind[2045]: New session 17 of user core. Sep 4 17:42:02.048937 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 4 17:42:03.259773 sshd[5022]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:03.270933 systemd[1]: sshd@16-172.31.20.143:22-139.178.68.195:35142.service: Deactivated successfully. Sep 4 17:42:03.283869 systemd[1]: session-17.scope: Deactivated successfully. Sep 4 17:42:03.289741 systemd-logind[2045]: Session 17 logged out. Waiting for processes to exit. Sep 4 17:42:03.316326 systemd[1]: Started sshd@17-172.31.20.143:22-139.178.68.195:35154.service - OpenSSH per-connection server daemon (139.178.68.195:35154). Sep 4 17:42:03.320888 systemd-logind[2045]: Removed session 17. Sep 4 17:42:03.553815 sshd[5034]: Accepted publickey for core from 139.178.68.195 port 35154 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:03.555152 sshd[5034]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:03.571180 systemd-logind[2045]: New session 18 of user core. Sep 4 17:42:03.577331 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 4 17:42:05.044561 sshd[5034]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:05.070773 systemd[1]: sshd@17-172.31.20.143:22-139.178.68.195:35154.service: Deactivated successfully. Sep 4 17:42:05.090622 systemd[1]: session-18.scope: Deactivated successfully. Sep 4 17:42:05.098231 systemd-logind[2045]: Session 18 logged out. Waiting for processes to exit. Sep 4 17:42:05.111388 systemd[1]: Started sshd@18-172.31.20.143:22-139.178.68.195:35168.service - OpenSSH per-connection server daemon (139.178.68.195:35168). Sep 4 17:42:05.117785 systemd-logind[2045]: Removed session 18. Sep 4 17:42:05.310121 sshd[5053]: Accepted publickey for core from 139.178.68.195 port 35168 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:05.311167 sshd[5053]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:05.319847 systemd-logind[2045]: New session 19 of user core. Sep 4 17:42:05.322548 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 4 17:42:05.818004 sshd[5053]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:05.822882 systemd-logind[2045]: Session 19 logged out. Waiting for processes to exit. Sep 4 17:42:05.825782 systemd[1]: sshd@18-172.31.20.143:22-139.178.68.195:35168.service: Deactivated successfully. Sep 4 17:42:05.829952 systemd[1]: session-19.scope: Deactivated successfully. Sep 4 17:42:05.831482 systemd-logind[2045]: Removed session 19. Sep 4 17:42:05.844730 systemd[1]: Started sshd@19-172.31.20.143:22-139.178.68.195:35176.service - OpenSSH per-connection server daemon (139.178.68.195:35176). Sep 4 17:42:06.010933 sshd[5066]: Accepted publickey for core from 139.178.68.195 port 35176 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:06.013657 sshd[5066]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:06.018569 systemd-logind[2045]: New session 20 of user core. Sep 4 17:42:06.021569 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 4 17:42:06.244321 sshd[5066]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:06.250086 systemd[1]: sshd@19-172.31.20.143:22-139.178.68.195:35176.service: Deactivated successfully. Sep 4 17:42:06.263008 systemd[1]: session-20.scope: Deactivated successfully. Sep 4 17:42:06.264070 systemd-logind[2045]: Session 20 logged out. Waiting for processes to exit. Sep 4 17:42:06.265199 systemd-logind[2045]: Removed session 20. Sep 4 17:42:11.291037 systemd[1]: Started sshd@20-172.31.20.143:22-139.178.68.195:36388.service - OpenSSH per-connection server daemon (139.178.68.195:36388). Sep 4 17:42:11.468093 sshd[5080]: Accepted publickey for core from 139.178.68.195 port 36388 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:11.469748 sshd[5080]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:11.477059 systemd-logind[2045]: New session 21 of user core. Sep 4 17:42:11.483325 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 4 17:42:11.695756 sshd[5080]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:11.702772 systemd[1]: sshd@20-172.31.20.143:22-139.178.68.195:36388.service: Deactivated successfully. Sep 4 17:42:11.710744 systemd[1]: session-21.scope: Deactivated successfully. Sep 4 17:42:11.711154 systemd-logind[2045]: Session 21 logged out. Waiting for processes to exit. Sep 4 17:42:11.715580 systemd-logind[2045]: Removed session 21. Sep 4 17:42:16.726705 systemd[1]: Started sshd@21-172.31.20.143:22-139.178.68.195:35626.service - OpenSSH per-connection server daemon (139.178.68.195:35626). Sep 4 17:42:16.905028 sshd[5097]: Accepted publickey for core from 139.178.68.195 port 35626 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:16.906848 sshd[5097]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:16.918358 systemd-logind[2045]: New session 22 of user core. Sep 4 17:42:16.923805 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 4 17:42:17.186349 sshd[5097]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:17.192254 systemd-logind[2045]: Session 22 logged out. Waiting for processes to exit. Sep 4 17:42:17.193839 systemd[1]: sshd@21-172.31.20.143:22-139.178.68.195:35626.service: Deactivated successfully. Sep 4 17:42:17.201129 systemd[1]: session-22.scope: Deactivated successfully. Sep 4 17:42:17.203424 systemd-logind[2045]: Removed session 22. Sep 4 17:42:22.233469 systemd[1]: Started sshd@22-172.31.20.143:22-139.178.68.195:35634.service - OpenSSH per-connection server daemon (139.178.68.195:35634). Sep 4 17:42:22.450278 sshd[5111]: Accepted publickey for core from 139.178.68.195 port 35634 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:22.452018 sshd[5111]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:22.460796 systemd-logind[2045]: New session 23 of user core. Sep 4 17:42:22.467057 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 4 17:42:22.686473 sshd[5111]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:22.705703 systemd[1]: sshd@22-172.31.20.143:22-139.178.68.195:35634.service: Deactivated successfully. Sep 4 17:42:22.711833 systemd[1]: session-23.scope: Deactivated successfully. Sep 4 17:42:22.712717 systemd-logind[2045]: Session 23 logged out. Waiting for processes to exit. Sep 4 17:42:22.714849 systemd-logind[2045]: Removed session 23. Sep 4 17:42:27.719145 systemd[1]: Started sshd@23-172.31.20.143:22-139.178.68.195:45354.service - OpenSSH per-connection server daemon (139.178.68.195:45354). Sep 4 17:42:27.935001 sshd[5127]: Accepted publickey for core from 139.178.68.195 port 45354 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:27.937972 sshd[5127]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:27.947334 systemd-logind[2045]: New session 24 of user core. Sep 4 17:42:27.954615 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 4 17:42:28.281882 sshd[5127]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:28.292460 systemd[1]: sshd@23-172.31.20.143:22-139.178.68.195:45354.service: Deactivated successfully. Sep 4 17:42:28.295402 systemd-logind[2045]: Session 24 logged out. Waiting for processes to exit. Sep 4 17:42:28.309099 systemd[1]: session-24.scope: Deactivated successfully. Sep 4 17:42:28.312080 systemd-logind[2045]: Removed session 24. Sep 4 17:42:28.317670 systemd[1]: Started sshd@24-172.31.20.143:22-139.178.68.195:45370.service - OpenSSH per-connection server daemon (139.178.68.195:45370). Sep 4 17:42:28.540653 sshd[5141]: Accepted publickey for core from 139.178.68.195 port 45370 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:28.544556 sshd[5141]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:28.554140 systemd-logind[2045]: New session 25 of user core. Sep 4 17:42:28.559731 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 4 17:42:30.316653 containerd[2079]: time="2024-09-04T17:42:30.316466455Z" level=info msg="StopContainer for \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\" with timeout 30 (s)" Sep 4 17:42:30.322479 containerd[2079]: time="2024-09-04T17:42:30.322415268Z" level=info msg="Stop container \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\" with signal terminated" Sep 4 17:42:30.335025 containerd[2079]: time="2024-09-04T17:42:30.334950198Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 4 17:42:30.342575 containerd[2079]: time="2024-09-04T17:42:30.342535973Z" level=info msg="StopContainer for \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\" with timeout 2 (s)" Sep 4 17:42:30.344616 containerd[2079]: time="2024-09-04T17:42:30.342829848Z" level=info msg="Stop container \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\" with signal terminated" Sep 4 17:42:30.362071 systemd-networkd[1648]: lxc_health: Link DOWN Sep 4 17:42:30.362080 systemd-networkd[1648]: lxc_health: Lost carrier Sep 4 17:42:30.409017 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf-rootfs.mount: Deactivated successfully. Sep 4 17:42:30.423591 containerd[2079]: time="2024-09-04T17:42:30.423488779Z" level=info msg="shim disconnected" id=e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf namespace=k8s.io Sep 4 17:42:30.423591 containerd[2079]: time="2024-09-04T17:42:30.423545930Z" level=warning msg="cleaning up after shim disconnected" id=e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf namespace=k8s.io Sep 4 17:42:30.423591 containerd[2079]: time="2024-09-04T17:42:30.423560250Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:30.431361 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664-rootfs.mount: Deactivated successfully. Sep 4 17:42:30.445260 containerd[2079]: time="2024-09-04T17:42:30.444782108Z" level=info msg="shim disconnected" id=6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664 namespace=k8s.io Sep 4 17:42:30.445260 containerd[2079]: time="2024-09-04T17:42:30.444900240Z" level=warning msg="cleaning up after shim disconnected" id=6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664 namespace=k8s.io Sep 4 17:42:30.445260 containerd[2079]: time="2024-09-04T17:42:30.444914802Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:30.456903 containerd[2079]: time="2024-09-04T17:42:30.456833023Z" level=info msg="StopContainer for \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\" returns successfully" Sep 4 17:42:30.458299 containerd[2079]: time="2024-09-04T17:42:30.458265822Z" level=info msg="StopPodSandbox for \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\"" Sep 4 17:42:30.466025 containerd[2079]: time="2024-09-04T17:42:30.465978895Z" level=info msg="Container to stop \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 17:42:30.471833 containerd[2079]: time="2024-09-04T17:42:30.468772269Z" level=info msg="StopContainer for \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\" returns successfully" Sep 4 17:42:30.471833 containerd[2079]: time="2024-09-04T17:42:30.471654602Z" level=info msg="StopPodSandbox for \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\"" Sep 4 17:42:30.471833 containerd[2079]: time="2024-09-04T17:42:30.471697198Z" level=info msg="Container to stop \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 17:42:30.471833 containerd[2079]: time="2024-09-04T17:42:30.471708064Z" level=info msg="Container to stop \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 17:42:30.471833 containerd[2079]: time="2024-09-04T17:42:30.471721954Z" level=info msg="Container to stop \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 17:42:30.471833 containerd[2079]: time="2024-09-04T17:42:30.471740365Z" level=info msg="Container to stop \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 17:42:30.471833 containerd[2079]: time="2024-09-04T17:42:30.471750134Z" level=info msg="Container to stop \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 4 17:42:30.470187 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4-shm.mount: Deactivated successfully. Sep 4 17:42:30.478942 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af-shm.mount: Deactivated successfully. Sep 4 17:42:30.558926 containerd[2079]: time="2024-09-04T17:42:30.558680308Z" level=info msg="shim disconnected" id=5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4 namespace=k8s.io Sep 4 17:42:30.558926 containerd[2079]: time="2024-09-04T17:42:30.558744491Z" level=warning msg="cleaning up after shim disconnected" id=5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4 namespace=k8s.io Sep 4 17:42:30.558926 containerd[2079]: time="2024-09-04T17:42:30.558756743Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:30.568486 containerd[2079]: time="2024-09-04T17:42:30.558831743Z" level=info msg="shim disconnected" id=de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af namespace=k8s.io Sep 4 17:42:30.568486 containerd[2079]: time="2024-09-04T17:42:30.559375732Z" level=warning msg="cleaning up after shim disconnected" id=de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af namespace=k8s.io Sep 4 17:42:30.568486 containerd[2079]: time="2024-09-04T17:42:30.559390468Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:30.632281 containerd[2079]: time="2024-09-04T17:42:30.631647831Z" level=info msg="TearDown network for sandbox \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" successfully" Sep 4 17:42:30.633441 containerd[2079]: time="2024-09-04T17:42:30.632410932Z" level=info msg="StopPodSandbox for \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" returns successfully" Sep 4 17:42:30.633441 containerd[2079]: time="2024-09-04T17:42:30.632368867Z" level=info msg="TearDown network for sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" successfully" Sep 4 17:42:30.633441 containerd[2079]: time="2024-09-04T17:42:30.632672560Z" level=info msg="StopPodSandbox for \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" returns successfully" Sep 4 17:42:30.729817 kubelet[3455]: I0904 17:42:30.729719 3455 scope.go:117] "RemoveContainer" containerID="e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf" Sep 4 17:42:30.754497 kubelet[3455]: I0904 17:42:30.754454 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-xtables-lock\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.754761 kubelet[3455]: I0904 17:42:30.754606 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-net\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.754761 kubelet[3455]: I0904 17:42:30.754641 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-kernel\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.754761 kubelet[3455]: I0904 17:42:30.754674 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5vdd\" (UniqueName: \"kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-kube-api-access-t5vdd\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.754761 kubelet[3455]: I0904 17:42:30.754697 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-etc-cni-netd\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.754761 kubelet[3455]: I0904 17:42:30.754726 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-bpf-maps\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.754761 kubelet[3455]: I0904 17:42:30.754754 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-config-path\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.755148 kubelet[3455]: I0904 17:42:30.754784 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-run\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.755148 kubelet[3455]: I0904 17:42:30.754853 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-cgroup\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.755148 kubelet[3455]: I0904 17:42:30.754885 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-hostproc\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.755148 kubelet[3455]: I0904 17:42:30.754918 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-cilium-config-path\") pod \"1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e\" (UID: \"1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e\") " Sep 4 17:42:30.755148 kubelet[3455]: I0904 17:42:30.754958 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk7nc\" (UniqueName: \"kubernetes.io/projected/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-kube-api-access-vk7nc\") pod \"1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e\" (UID: \"1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e\") " Sep 4 17:42:30.755148 kubelet[3455]: I0904 17:42:30.755088 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-hubble-tls\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.755546 kubelet[3455]: I0904 17:42:30.755122 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cni-path\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.755546 kubelet[3455]: I0904 17:42:30.755176 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/63a1952e-55aa-4c14-8998-d2fece4d52cc-clustermesh-secrets\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.755546 kubelet[3455]: I0904 17:42:30.755228 3455 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-lib-modules\") pod \"63a1952e-55aa-4c14-8998-d2fece4d52cc\" (UID: \"63a1952e-55aa-4c14-8998-d2fece4d52cc\") " Sep 4 17:42:30.774263 kubelet[3455]: I0904 17:42:30.771008 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.774263 kubelet[3455]: I0904 17:42:30.770724 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.774263 kubelet[3455]: I0904 17:42:30.772821 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.774263 kubelet[3455]: I0904 17:42:30.772849 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.774263 kubelet[3455]: I0904 17:42:30.773302 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-hostproc" (OuterVolumeSpecName: "hostproc") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.774709 kubelet[3455]: I0904 17:42:30.773338 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.774709 kubelet[3455]: I0904 17:42:30.773358 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.777309 kubelet[3455]: I0904 17:42:30.777272 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 4 17:42:30.778727 kubelet[3455]: I0904 17:42:30.777483 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.778727 kubelet[3455]: I0904 17:42:30.777530 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.780437 containerd[2079]: time="2024-09-04T17:42:30.780207624Z" level=info msg="RemoveContainer for \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\"" Sep 4 17:42:30.794219 containerd[2079]: time="2024-09-04T17:42:30.793520900Z" level=info msg="RemoveContainer for \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\" returns successfully" Sep 4 17:42:30.797206 kubelet[3455]: I0904 17:42:30.796815 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cni-path" (OuterVolumeSpecName: "cni-path") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 4 17:42:30.802956 kubelet[3455]: I0904 17:42:30.802895 3455 scope.go:117] "RemoveContainer" containerID="e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf" Sep 4 17:42:30.807617 kubelet[3455]: I0904 17:42:30.807577 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 4 17:42:30.813969 kubelet[3455]: I0904 17:42:30.813419 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63a1952e-55aa-4c14-8998-d2fece4d52cc-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 4 17:42:30.814377 kubelet[3455]: I0904 17:42:30.814348 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-kube-api-access-t5vdd" (OuterVolumeSpecName: "kube-api-access-t5vdd") pod "63a1952e-55aa-4c14-8998-d2fece4d52cc" (UID: "63a1952e-55aa-4c14-8998-d2fece4d52cc"). InnerVolumeSpecName "kube-api-access-t5vdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 4 17:42:30.814765 kubelet[3455]: I0904 17:42:30.814738 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e" (UID: "1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 4 17:42:30.820276 kubelet[3455]: I0904 17:42:30.818574 3455 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-kube-api-access-vk7nc" (OuterVolumeSpecName: "kube-api-access-vk7nc") pod "1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e" (UID: "1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e"). InnerVolumeSpecName "kube-api-access-vk7nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 4 17:42:30.820501 containerd[2079]: time="2024-09-04T17:42:30.804728285Z" level=error msg="ContainerStatus for \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\": not found" Sep 4 17:42:30.825476 kubelet[3455]: E0904 17:42:30.825440 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\": not found" containerID="e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf" Sep 4 17:42:30.840104 kubelet[3455]: I0904 17:42:30.840049 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf"} err="failed to get container status \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\": rpc error: code = NotFound desc = an error occurred when try to find container \"e1836a0d2f88e315f4d05f545fc9ceddbe9885c82367124a6daa8612985745cf\": not found" Sep 4 17:42:30.840104 kubelet[3455]: I0904 17:42:30.840112 3455 scope.go:117] "RemoveContainer" containerID="6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664" Sep 4 17:42:30.842027 containerd[2079]: time="2024-09-04T17:42:30.841853690Z" level=info msg="RemoveContainer for \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\"" Sep 4 17:42:30.852095 containerd[2079]: time="2024-09-04T17:42:30.851879160Z" level=info msg="RemoveContainer for \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\" returns successfully" Sep 4 17:42:30.852578 kubelet[3455]: I0904 17:42:30.852500 3455 scope.go:117] "RemoveContainer" containerID="ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d" Sep 4 17:42:30.855024 containerd[2079]: time="2024-09-04T17:42:30.854990964Z" level=info msg="RemoveContainer for \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\"" Sep 4 17:42:30.855798 kubelet[3455]: I0904 17:42:30.855772 3455 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-xtables-lock\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855804 3455 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-t5vdd\" (UniqueName: \"kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-kube-api-access-t5vdd\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855821 3455 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-net\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855836 3455 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-host-proc-sys-kernel\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855851 3455 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-bpf-maps\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855864 3455 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-etc-cni-netd\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855878 3455 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-run\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855893 3455 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-config-path\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.855987 kubelet[3455]: I0904 17:42:30.855906 3455 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-hostproc\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.856598 kubelet[3455]: I0904 17:42:30.855978 3455 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cilium-cgroup\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.856598 kubelet[3455]: I0904 17:42:30.855993 3455 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/63a1952e-55aa-4c14-8998-d2fece4d52cc-hubble-tls\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.856598 kubelet[3455]: I0904 17:42:30.856113 3455 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-cni-path\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.856598 kubelet[3455]: I0904 17:42:30.856137 3455 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-cilium-config-path\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.856598 kubelet[3455]: I0904 17:42:30.856152 3455 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-vk7nc\" (UniqueName: \"kubernetes.io/projected/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e-kube-api-access-vk7nc\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.856598 kubelet[3455]: I0904 17:42:30.856168 3455 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/63a1952e-55aa-4c14-8998-d2fece4d52cc-lib-modules\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.856598 kubelet[3455]: I0904 17:42:30.856185 3455 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/63a1952e-55aa-4c14-8998-d2fece4d52cc-clustermesh-secrets\") on node \"ip-172-31-20-143\" DevicePath \"\"" Sep 4 17:42:30.859505 containerd[2079]: time="2024-09-04T17:42:30.859472412Z" level=info msg="RemoveContainer for \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\" returns successfully" Sep 4 17:42:30.859836 kubelet[3455]: I0904 17:42:30.859710 3455 scope.go:117] "RemoveContainer" containerID="3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58" Sep 4 17:42:30.861416 containerd[2079]: time="2024-09-04T17:42:30.861380094Z" level=info msg="RemoveContainer for \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\"" Sep 4 17:42:30.865583 containerd[2079]: time="2024-09-04T17:42:30.865467769Z" level=info msg="RemoveContainer for \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\" returns successfully" Sep 4 17:42:30.865722 kubelet[3455]: I0904 17:42:30.865701 3455 scope.go:117] "RemoveContainer" containerID="aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa" Sep 4 17:42:30.867869 containerd[2079]: time="2024-09-04T17:42:30.867838863Z" level=info msg="RemoveContainer for \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\"" Sep 4 17:42:30.873326 containerd[2079]: time="2024-09-04T17:42:30.873286257Z" level=info msg="RemoveContainer for \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\" returns successfully" Sep 4 17:42:30.873566 kubelet[3455]: I0904 17:42:30.873522 3455 scope.go:117] "RemoveContainer" containerID="881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681" Sep 4 17:42:30.875041 containerd[2079]: time="2024-09-04T17:42:30.875007047Z" level=info msg="RemoveContainer for \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\"" Sep 4 17:42:30.880003 containerd[2079]: time="2024-09-04T17:42:30.879897667Z" level=info msg="RemoveContainer for \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\" returns successfully" Sep 4 17:42:30.880406 kubelet[3455]: I0904 17:42:30.880309 3455 scope.go:117] "RemoveContainer" containerID="6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664" Sep 4 17:42:30.880810 containerd[2079]: time="2024-09-04T17:42:30.880761540Z" level=error msg="ContainerStatus for \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\": not found" Sep 4 17:42:30.880961 kubelet[3455]: E0904 17:42:30.880939 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\": not found" containerID="6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664" Sep 4 17:42:30.881029 kubelet[3455]: I0904 17:42:30.880986 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664"} err="failed to get container status \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\": rpc error: code = NotFound desc = an error occurred when try to find container \"6d974b78da7bac9bfb44405fa0661dc3ab3ae3032645b58e233e9100bfbc8664\": not found" Sep 4 17:42:30.881029 kubelet[3455]: I0904 17:42:30.881002 3455 scope.go:117] "RemoveContainer" containerID="ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d" Sep 4 17:42:30.881587 containerd[2079]: time="2024-09-04T17:42:30.881453994Z" level=error msg="ContainerStatus for \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\": not found" Sep 4 17:42:30.881842 kubelet[3455]: E0904 17:42:30.881821 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\": not found" containerID="ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d" Sep 4 17:42:30.881958 kubelet[3455]: I0904 17:42:30.881862 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d"} err="failed to get container status \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\": rpc error: code = NotFound desc = an error occurred when try to find container \"ce894f5843d293681c7f3558e113f06ed1835f9f2becc661ace58e4ff7bdd92d\": not found" Sep 4 17:42:30.881958 kubelet[3455]: I0904 17:42:30.881930 3455 scope.go:117] "RemoveContainer" containerID="3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58" Sep 4 17:42:30.882402 containerd[2079]: time="2024-09-04T17:42:30.882351795Z" level=error msg="ContainerStatus for \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\": not found" Sep 4 17:42:30.882696 kubelet[3455]: E0904 17:42:30.882586 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\": not found" containerID="3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58" Sep 4 17:42:30.882782 kubelet[3455]: I0904 17:42:30.882697 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58"} err="failed to get container status \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\": rpc error: code = NotFound desc = an error occurred when try to find container \"3f9e098ee3a04e6c74a13bf73509346463b7ce1b261ff17760baaf176a329b58\": not found" Sep 4 17:42:30.882782 kubelet[3455]: I0904 17:42:30.882711 3455 scope.go:117] "RemoveContainer" containerID="aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa" Sep 4 17:42:30.882934 containerd[2079]: time="2024-09-04T17:42:30.882897794Z" level=error msg="ContainerStatus for \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\": not found" Sep 4 17:42:30.883165 kubelet[3455]: E0904 17:42:30.883073 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\": not found" containerID="aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa" Sep 4 17:42:30.883165 kubelet[3455]: I0904 17:42:30.883152 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa"} err="failed to get container status \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\": rpc error: code = NotFound desc = an error occurred when try to find container \"aaf57f627ec4f501f6854149726d8a0f8dd2c00cfdb14da74a5215c96c84e3aa\": not found" Sep 4 17:42:30.883165 kubelet[3455]: I0904 17:42:30.883167 3455 scope.go:117] "RemoveContainer" containerID="881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681" Sep 4 17:42:30.883424 containerd[2079]: time="2024-09-04T17:42:30.883373935Z" level=error msg="ContainerStatus for \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\": not found" Sep 4 17:42:30.883580 kubelet[3455]: E0904 17:42:30.883547 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\": not found" containerID="881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681" Sep 4 17:42:30.883651 kubelet[3455]: I0904 17:42:30.883583 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681"} err="failed to get container status \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\": rpc error: code = NotFound desc = an error occurred when try to find container \"881f0dcd1ca43b0ac8553392233f595b2363721b9819178bc913529d952ff681\": not found" Sep 4 17:42:31.290355 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af-rootfs.mount: Deactivated successfully. Sep 4 17:42:31.290571 systemd[1]: var-lib-kubelet-pods-63a1952e\x2d55aa\x2d4c14\x2d8998\x2dd2fece4d52cc-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dt5vdd.mount: Deactivated successfully. Sep 4 17:42:31.290855 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4-rootfs.mount: Deactivated successfully. Sep 4 17:42:31.293608 systemd[1]: var-lib-kubelet-pods-1dbb0a38\x2d5c13\x2d4611\x2d9fcd\x2df3ff69be0c9e-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dvk7nc.mount: Deactivated successfully. Sep 4 17:42:31.293943 systemd[1]: var-lib-kubelet-pods-63a1952e\x2d55aa\x2d4c14\x2d8998\x2dd2fece4d52cc-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 4 17:42:31.294280 systemd[1]: var-lib-kubelet-pods-63a1952e\x2d55aa\x2d4c14\x2d8998\x2dd2fece4d52cc-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 4 17:42:32.032309 sshd[5141]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:32.066769 systemd[1]: Started sshd@25-172.31.20.143:22-139.178.68.195:45384.service - OpenSSH per-connection server daemon (139.178.68.195:45384). Sep 4 17:42:32.073379 systemd[1]: sshd@24-172.31.20.143:22-139.178.68.195:45370.service: Deactivated successfully. Sep 4 17:42:32.089556 systemd[1]: session-25.scope: Deactivated successfully. Sep 4 17:42:32.104417 systemd-logind[2045]: Session 25 logged out. Waiting for processes to exit. Sep 4 17:42:32.113410 systemd-logind[2045]: Removed session 25. Sep 4 17:42:32.290223 sshd[5305]: Accepted publickey for core from 139.178.68.195 port 45384 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:32.292362 sshd[5305]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:32.314835 systemd-logind[2045]: New session 26 of user core. Sep 4 17:42:32.323616 systemd[1]: Started session-26.scope - Session 26 of User core. Sep 4 17:42:32.453285 systemd-journald[1562]: Under memory pressure, flushing caches. Sep 4 17:42:32.452304 systemd-resolved[1963]: Under memory pressure, flushing caches. Sep 4 17:42:32.452334 systemd-resolved[1963]: Flushed all caches. Sep 4 17:42:33.042812 kubelet[3455]: I0904 17:42:33.042567 3455 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e" path="/var/lib/kubelet/pods/1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e/volumes" Sep 4 17:42:33.044965 kubelet[3455]: I0904 17:42:33.044287 3455 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" path="/var/lib/kubelet/pods/63a1952e-55aa-4c14-8998-d2fece4d52cc/volumes" Sep 4 17:42:33.059402 ntpd[2033]: Deleting interface #10 lxc_health, fe80::e0d2:2eff:feca:184%8#123, interface stats: received=0, sent=0, dropped=0, active_time=74 secs Sep 4 17:42:33.060485 ntpd[2033]: 4 Sep 17:42:33 ntpd[2033]: Deleting interface #10 lxc_health, fe80::e0d2:2eff:feca:184%8#123, interface stats: received=0, sent=0, dropped=0, active_time=74 secs Sep 4 17:42:33.513489 sshd[5305]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:33.522804 systemd-logind[2045]: Session 26 logged out. Waiting for processes to exit. Sep 4 17:42:33.525188 systemd[1]: sshd@25-172.31.20.143:22-139.178.68.195:45384.service: Deactivated successfully. Sep 4 17:42:33.550863 systemd[1]: Started sshd@26-172.31.20.143:22-139.178.68.195:45398.service - OpenSSH per-connection server daemon (139.178.68.195:45398). Sep 4 17:42:33.551723 systemd[1]: session-26.scope: Deactivated successfully. Sep 4 17:42:33.559523 systemd-logind[2045]: Removed session 26. Sep 4 17:42:33.594076 kubelet[3455]: I0904 17:42:33.594025 3455 topology_manager.go:215] "Topology Admit Handler" podUID="8f36ed50-36c7-41b5-9b6e-9e72cf5d8047" podNamespace="kube-system" podName="cilium-hvd5t" Sep 4 17:42:33.599264 kubelet[3455]: E0904 17:42:33.596977 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" containerName="apply-sysctl-overwrites" Sep 4 17:42:33.599264 kubelet[3455]: E0904 17:42:33.597039 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" containerName="mount-bpf-fs" Sep 4 17:42:33.599264 kubelet[3455]: E0904 17:42:33.597055 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e" containerName="cilium-operator" Sep 4 17:42:33.599264 kubelet[3455]: E0904 17:42:33.597066 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" containerName="mount-cgroup" Sep 4 17:42:33.599264 kubelet[3455]: E0904 17:42:33.597077 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" containerName="clean-cilium-state" Sep 4 17:42:33.599264 kubelet[3455]: E0904 17:42:33.597087 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" containerName="cilium-agent" Sep 4 17:42:33.599264 kubelet[3455]: I0904 17:42:33.597298 3455 memory_manager.go:346] "RemoveStaleState removing state" podUID="1dbb0a38-5c13-4611-9fcd-f3ff69be0c9e" containerName="cilium-operator" Sep 4 17:42:33.599264 kubelet[3455]: I0904 17:42:33.597316 3455 memory_manager.go:346] "RemoveStaleState removing state" podUID="63a1952e-55aa-4c14-8998-d2fece4d52cc" containerName="cilium-agent" Sep 4 17:42:33.711280 kubelet[3455]: I0904 17:42:33.708073 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-cni-path\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721271 kubelet[3455]: I0904 17:42:33.711520 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-cilium-config-path\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721271 kubelet[3455]: I0904 17:42:33.711574 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-bpf-maps\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721271 kubelet[3455]: I0904 17:42:33.711607 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-hostproc\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721271 kubelet[3455]: I0904 17:42:33.711643 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-cilium-run\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721271 kubelet[3455]: I0904 17:42:33.711690 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-host-proc-sys-kernel\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721271 kubelet[3455]: I0904 17:42:33.716868 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99drc\" (UniqueName: \"kubernetes.io/projected/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-kube-api-access-99drc\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721651 kubelet[3455]: I0904 17:42:33.716962 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-clustermesh-secrets\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721651 kubelet[3455]: I0904 17:42:33.716995 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-cilium-ipsec-secrets\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721651 kubelet[3455]: I0904 17:42:33.717026 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-host-proc-sys-net\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721651 kubelet[3455]: I0904 17:42:33.717060 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-etc-cni-netd\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721651 kubelet[3455]: I0904 17:42:33.717095 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-xtables-lock\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.721651 kubelet[3455]: I0904 17:42:33.717128 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-hubble-tls\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.722214 kubelet[3455]: I0904 17:42:33.717164 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-cilium-cgroup\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.722214 kubelet[3455]: I0904 17:42:33.717196 3455 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8f36ed50-36c7-41b5-9b6e-9e72cf5d8047-lib-modules\") pod \"cilium-hvd5t\" (UID: \"8f36ed50-36c7-41b5-9b6e-9e72cf5d8047\") " pod="kube-system/cilium-hvd5t" Sep 4 17:42:33.791814 sshd[5321]: Accepted publickey for core from 139.178.68.195 port 45398 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:33.800450 sshd[5321]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:33.833345 systemd-logind[2045]: New session 27 of user core. Sep 4 17:42:33.846961 systemd[1]: Started session-27.scope - Session 27 of User core. Sep 4 17:42:33.964014 containerd[2079]: time="2024-09-04T17:42:33.963666353Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-hvd5t,Uid:8f36ed50-36c7-41b5-9b6e-9e72cf5d8047,Namespace:kube-system,Attempt:0,}" Sep 4 17:42:34.008849 containerd[2079]: time="2024-09-04T17:42:34.008552655Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:42:34.009264 containerd[2079]: time="2024-09-04T17:42:34.008642302Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:42:34.009264 containerd[2079]: time="2024-09-04T17:42:34.009052460Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:34.009360 containerd[2079]: time="2024-09-04T17:42:34.009224732Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:42:34.063003 sshd[5321]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:34.070255 systemd-logind[2045]: Session 27 logged out. Waiting for processes to exit. Sep 4 17:42:34.071330 systemd[1]: sshd@26-172.31.20.143:22-139.178.68.195:45398.service: Deactivated successfully. Sep 4 17:42:34.101360 systemd[1]: Started sshd@27-172.31.20.143:22-139.178.68.195:45404.service - OpenSSH per-connection server daemon (139.178.68.195:45404). Sep 4 17:42:34.102910 systemd[1]: session-27.scope: Deactivated successfully. Sep 4 17:42:34.109032 containerd[2079]: time="2024-09-04T17:42:34.108899573Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-hvd5t,Uid:8f36ed50-36c7-41b5-9b6e-9e72cf5d8047,Namespace:kube-system,Attempt:0,} returns sandbox id \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\"" Sep 4 17:42:34.110765 systemd-logind[2045]: Removed session 27. Sep 4 17:42:34.121672 containerd[2079]: time="2024-09-04T17:42:34.121630809Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 4 17:42:34.145451 containerd[2079]: time="2024-09-04T17:42:34.145407502Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"5d329b31dc445ec2c17fe24e3f2c22daa7e0d33629204d0287d7a03f1a04ccaa\"" Sep 4 17:42:34.147339 containerd[2079]: time="2024-09-04T17:42:34.147302710Z" level=info msg="StartContainer for \"5d329b31dc445ec2c17fe24e3f2c22daa7e0d33629204d0287d7a03f1a04ccaa\"" Sep 4 17:42:34.228579 containerd[2079]: time="2024-09-04T17:42:34.228526436Z" level=info msg="StartContainer for \"5d329b31dc445ec2c17fe24e3f2c22daa7e0d33629204d0287d7a03f1a04ccaa\" returns successfully" Sep 4 17:42:34.290160 sshd[5369]: Accepted publickey for core from 139.178.68.195 port 45404 ssh2: RSA SHA256:7R68OPxBD1aKub0NQezDW73KPeSGi+cl3Ia6CweCJtQ Sep 4 17:42:34.292558 sshd[5369]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 4 17:42:34.299571 systemd-logind[2045]: New session 28 of user core. Sep 4 17:42:34.305572 systemd[1]: Started session-28.scope - Session 28 of User core. Sep 4 17:42:34.318078 containerd[2079]: time="2024-09-04T17:42:34.315726013Z" level=info msg="shim disconnected" id=5d329b31dc445ec2c17fe24e3f2c22daa7e0d33629204d0287d7a03f1a04ccaa namespace=k8s.io Sep 4 17:42:34.318078 containerd[2079]: time="2024-09-04T17:42:34.315793780Z" level=warning msg="cleaning up after shim disconnected" id=5d329b31dc445ec2c17fe24e3f2c22daa7e0d33629204d0287d7a03f1a04ccaa namespace=k8s.io Sep 4 17:42:34.318078 containerd[2079]: time="2024-09-04T17:42:34.315808276Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:34.380105 kubelet[3455]: E0904 17:42:34.380081 3455 kubelet.go:2855] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 4 17:42:34.790943 containerd[2079]: time="2024-09-04T17:42:34.790890050Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 4 17:42:34.816604 containerd[2079]: time="2024-09-04T17:42:34.813059656Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"84acb737bf4056974c22be5b824d3ed30b3d26a8dba3696c145fdc0dcd310f0d\"" Sep 4 17:42:34.816604 containerd[2079]: time="2024-09-04T17:42:34.815457338Z" level=info msg="StartContainer for \"84acb737bf4056974c22be5b824d3ed30b3d26a8dba3696c145fdc0dcd310f0d\"" Sep 4 17:42:34.910818 containerd[2079]: time="2024-09-04T17:42:34.910691268Z" level=info msg="StartContainer for \"84acb737bf4056974c22be5b824d3ed30b3d26a8dba3696c145fdc0dcd310f0d\" returns successfully" Sep 4 17:42:34.976889 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-84acb737bf4056974c22be5b824d3ed30b3d26a8dba3696c145fdc0dcd310f0d-rootfs.mount: Deactivated successfully. Sep 4 17:42:34.986447 containerd[2079]: time="2024-09-04T17:42:34.986378481Z" level=info msg="shim disconnected" id=84acb737bf4056974c22be5b824d3ed30b3d26a8dba3696c145fdc0dcd310f0d namespace=k8s.io Sep 4 17:42:34.986447 containerd[2079]: time="2024-09-04T17:42:34.986443330Z" level=warning msg="cleaning up after shim disconnected" id=84acb737bf4056974c22be5b824d3ed30b3d26a8dba3696c145fdc0dcd310f0d namespace=k8s.io Sep 4 17:42:34.986447 containerd[2079]: time="2024-09-04T17:42:34.986454744Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:35.044873 kubelet[3455]: E0904 17:42:35.044285 3455 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-5dd5756b68-bgp4z" podUID="e9ebf3ef-0547-4332-846a-0d5bf557b87d" Sep 4 17:42:35.812927 containerd[2079]: time="2024-09-04T17:42:35.812555770Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 4 17:42:35.873347 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2967456849.mount: Deactivated successfully. Sep 4 17:42:35.873796 containerd[2079]: time="2024-09-04T17:42:35.873406848Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7a05b64941dfba4b76a620cde38d0d6cd41beaf96f50f102af76acbac0711605\"" Sep 4 17:42:35.877111 containerd[2079]: time="2024-09-04T17:42:35.874435629Z" level=info msg="StartContainer for \"7a05b64941dfba4b76a620cde38d0d6cd41beaf96f50f102af76acbac0711605\"" Sep 4 17:42:35.975919 systemd[1]: run-containerd-runc-k8s.io-7a05b64941dfba4b76a620cde38d0d6cd41beaf96f50f102af76acbac0711605-runc.i71ZAA.mount: Deactivated successfully. Sep 4 17:42:36.024668 containerd[2079]: time="2024-09-04T17:42:36.024621877Z" level=info msg="StartContainer for \"7a05b64941dfba4b76a620cde38d0d6cd41beaf96f50f102af76acbac0711605\" returns successfully" Sep 4 17:42:36.087945 containerd[2079]: time="2024-09-04T17:42:36.087498816Z" level=info msg="shim disconnected" id=7a05b64941dfba4b76a620cde38d0d6cd41beaf96f50f102af76acbac0711605 namespace=k8s.io Sep 4 17:42:36.087945 containerd[2079]: time="2024-09-04T17:42:36.087859439Z" level=warning msg="cleaning up after shim disconnected" id=7a05b64941dfba4b76a620cde38d0d6cd41beaf96f50f102af76acbac0711605 namespace=k8s.io Sep 4 17:42:36.087945 containerd[2079]: time="2024-09-04T17:42:36.087876641Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:36.806856 containerd[2079]: time="2024-09-04T17:42:36.806781744Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 4 17:42:36.832965 containerd[2079]: time="2024-09-04T17:42:36.832760989Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"03622e91195863445ca89251a2c387d77c191d06d70ed2cd9b70963914c52e2c\"" Sep 4 17:42:36.835462 containerd[2079]: time="2024-09-04T17:42:36.835298136Z" level=info msg="StartContainer for \"03622e91195863445ca89251a2c387d77c191d06d70ed2cd9b70963914c52e2c\"" Sep 4 17:42:36.863186 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7a05b64941dfba4b76a620cde38d0d6cd41beaf96f50f102af76acbac0711605-rootfs.mount: Deactivated successfully. Sep 4 17:42:36.930339 containerd[2079]: time="2024-09-04T17:42:36.930087662Z" level=info msg="StartContainer for \"03622e91195863445ca89251a2c387d77c191d06d70ed2cd9b70963914c52e2c\" returns successfully" Sep 4 17:42:36.967463 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-03622e91195863445ca89251a2c387d77c191d06d70ed2cd9b70963914c52e2c-rootfs.mount: Deactivated successfully. Sep 4 17:42:36.977218 containerd[2079]: time="2024-09-04T17:42:36.977113910Z" level=info msg="shim disconnected" id=03622e91195863445ca89251a2c387d77c191d06d70ed2cd9b70963914c52e2c namespace=k8s.io Sep 4 17:42:36.977218 containerd[2079]: time="2024-09-04T17:42:36.977202202Z" level=warning msg="cleaning up after shim disconnected" id=03622e91195863445ca89251a2c387d77c191d06d70ed2cd9b70963914c52e2c namespace=k8s.io Sep 4 17:42:36.977218 containerd[2079]: time="2024-09-04T17:42:36.977217135Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:42:37.026063 kubelet[3455]: E0904 17:42:37.025607 3455 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-5dd5756b68-bgp4z" podUID="e9ebf3ef-0547-4332-846a-0d5bf557b87d" Sep 4 17:42:37.808306 containerd[2079]: time="2024-09-04T17:42:37.806404792Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 4 17:42:37.834026 containerd[2079]: time="2024-09-04T17:42:37.830633063Z" level=info msg="CreateContainer within sandbox \"70bfe707f0a431b550ad1d7e04508d717fe436c07c052c76094e48002f7ee77a\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"bd6058757089c61778e2cb8d8dfcdd545ceb675d7dd3fac73c1c282ec9d6bd5b\"" Sep 4 17:42:37.838217 containerd[2079]: time="2024-09-04T17:42:37.836540320Z" level=info msg="StartContainer for \"bd6058757089c61778e2cb8d8dfcdd545ceb675d7dd3fac73c1c282ec9d6bd5b\"" Sep 4 17:42:37.957307 containerd[2079]: time="2024-09-04T17:42:37.957261111Z" level=info msg="StartContainer for \"bd6058757089c61778e2cb8d8dfcdd545ceb675d7dd3fac73c1c282ec9d6bd5b\" returns successfully" Sep 4 17:42:38.996881 containerd[2079]: time="2024-09-04T17:42:38.996830951Z" level=info msg="StopPodSandbox for \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\"" Sep 4 17:42:38.997444 containerd[2079]: time="2024-09-04T17:42:38.997376475Z" level=info msg="TearDown network for sandbox \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" successfully" Sep 4 17:42:38.997444 containerd[2079]: time="2024-09-04T17:42:38.997402326Z" level=info msg="StopPodSandbox for \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" returns successfully" Sep 4 17:42:39.007368 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 4 17:42:39.017741 containerd[2079]: time="2024-09-04T17:42:39.017592725Z" level=info msg="RemovePodSandbox for \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\"" Sep 4 17:42:39.024345 containerd[2079]: time="2024-09-04T17:42:39.024297195Z" level=info msg="Forcibly stopping sandbox \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\"" Sep 4 17:42:39.032126 containerd[2079]: time="2024-09-04T17:42:39.031981538Z" level=info msg="TearDown network for sandbox \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" successfully" Sep 4 17:42:39.034342 kubelet[3455]: E0904 17:42:39.034023 3455 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-5dd5756b68-bgp4z" podUID="e9ebf3ef-0547-4332-846a-0d5bf557b87d" Sep 4 17:42:39.047406 containerd[2079]: time="2024-09-04T17:42:39.046790266Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:42:39.047554 containerd[2079]: time="2024-09-04T17:42:39.047440384Z" level=info msg="RemovePodSandbox \"5e8e2154d460013cefce0a518d7022cfb567394282846c6b0737e8151d56e5e4\" returns successfully" Sep 4 17:42:39.055150 containerd[2079]: time="2024-09-04T17:42:39.054388017Z" level=info msg="StopPodSandbox for \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\"" Sep 4 17:42:39.055150 containerd[2079]: time="2024-09-04T17:42:39.054608519Z" level=info msg="TearDown network for sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" successfully" Sep 4 17:42:39.055150 containerd[2079]: time="2024-09-04T17:42:39.054628947Z" level=info msg="StopPodSandbox for \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" returns successfully" Sep 4 17:42:39.060091 containerd[2079]: time="2024-09-04T17:42:39.058211784Z" level=info msg="RemovePodSandbox for \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\"" Sep 4 17:42:39.060091 containerd[2079]: time="2024-09-04T17:42:39.058937770Z" level=info msg="Forcibly stopping sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\"" Sep 4 17:42:39.060091 containerd[2079]: time="2024-09-04T17:42:39.059094183Z" level=info msg="TearDown network for sandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" successfully" Sep 4 17:42:39.069966 containerd[2079]: time="2024-09-04T17:42:39.069609284Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:42:39.069966 containerd[2079]: time="2024-09-04T17:42:39.069695502Z" level=info msg="RemovePodSandbox \"de198e6310b1942f13d342ff951ea0ba71a16a7a2cd75f0d8fdd6cca88a743af\" returns successfully" Sep 4 17:42:41.235163 systemd[1]: run-containerd-runc-k8s.io-bd6058757089c61778e2cb8d8dfcdd545ceb675d7dd3fac73c1c282ec9d6bd5b-runc.3bd8ci.mount: Deactivated successfully. Sep 4 17:42:42.950496 systemd-networkd[1648]: lxc_health: Link UP Sep 4 17:42:42.956141 systemd-networkd[1648]: lxc_health: Gained carrier Sep 4 17:42:42.973136 (udev-worker)[6178]: Network interface NamePolicy= disabled on kernel command line. Sep 4 17:42:43.573440 systemd[1]: run-containerd-runc-k8s.io-bd6058757089c61778e2cb8d8dfcdd545ceb675d7dd3fac73c1c282ec9d6bd5b-runc.1myn3i.mount: Deactivated successfully. Sep 4 17:42:44.053977 kubelet[3455]: I0904 17:42:44.053921 3455 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-hvd5t" podStartSLOduration=11.0419796 podCreationTimestamp="2024-09-04 17:42:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:42:38.896802259 +0000 UTC m=+120.546011030" watchObservedRunningTime="2024-09-04 17:42:44.0419796 +0000 UTC m=+125.691188372" Sep 4 17:42:44.229696 systemd-networkd[1648]: lxc_health: Gained IPv6LL Sep 4 17:42:46.046899 systemd[1]: run-containerd-runc-k8s.io-bd6058757089c61778e2cb8d8dfcdd545ceb675d7dd3fac73c1c282ec9d6bd5b-runc.8sGhb8.mount: Deactivated successfully. Sep 4 17:42:46.187089 kubelet[3455]: E0904 17:42:46.187027 3455 upgradeaware.go:425] Error proxying data from client to backend: readfrom tcp 127.0.0.1:46268->127.0.0.1:33585: write tcp 127.0.0.1:46268->127.0.0.1:33585: write: connection reset by peer Sep 4 17:42:47.059517 ntpd[2033]: Listen normally on 13 lxc_health [fe80::4412:15ff:fe09:f96a%14]:123 Sep 4 17:42:47.060364 ntpd[2033]: 4 Sep 17:42:47 ntpd[2033]: Listen normally on 13 lxc_health [fe80::4412:15ff:fe09:f96a%14]:123 Sep 4 17:42:48.532429 sshd[5369]: pam_unix(sshd:session): session closed for user core Sep 4 17:42:48.543777 systemd[1]: sshd@27-172.31.20.143:22-139.178.68.195:45404.service: Deactivated successfully. Sep 4 17:42:48.553565 systemd[1]: session-28.scope: Deactivated successfully. Sep 4 17:42:48.555989 systemd-logind[2045]: Session 28 logged out. Waiting for processes to exit. Sep 4 17:42:48.559775 systemd-logind[2045]: Removed session 28. Sep 4 17:43:03.350015 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bee494728ab0a9cd7334dce1badaac45e0d9da4c73409fa45d941ba0f971126a-rootfs.mount: Deactivated successfully. Sep 4 17:43:03.358705 containerd[2079]: time="2024-09-04T17:43:03.358465407Z" level=info msg="shim disconnected" id=bee494728ab0a9cd7334dce1badaac45e0d9da4c73409fa45d941ba0f971126a namespace=k8s.io Sep 4 17:43:03.358705 containerd[2079]: time="2024-09-04T17:43:03.358667479Z" level=warning msg="cleaning up after shim disconnected" id=bee494728ab0a9cd7334dce1badaac45e0d9da4c73409fa45d941ba0f971126a namespace=k8s.io Sep 4 17:43:03.363885 containerd[2079]: time="2024-09-04T17:43:03.358683382Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:43:03.923474 kubelet[3455]: I0904 17:43:03.923443 3455 scope.go:117] "RemoveContainer" containerID="bee494728ab0a9cd7334dce1badaac45e0d9da4c73409fa45d941ba0f971126a" Sep 4 17:43:03.926897 containerd[2079]: time="2024-09-04T17:43:03.926848764Z" level=info msg="CreateContainer within sandbox \"7ab0f5a3d699b3896b1cc70c93f3bff768f6532092bdd932c8cde4c766f2dad6\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Sep 4 17:43:03.950384 containerd[2079]: time="2024-09-04T17:43:03.950342499Z" level=info msg="CreateContainer within sandbox \"7ab0f5a3d699b3896b1cc70c93f3bff768f6532092bdd932c8cde4c766f2dad6\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"9fafd529f01226e081e10cf8bce8ed357ac3b68faef5d7a7a73b2fef38d57f25\"" Sep 4 17:43:03.950981 containerd[2079]: time="2024-09-04T17:43:03.950949817Z" level=info msg="StartContainer for \"9fafd529f01226e081e10cf8bce8ed357ac3b68faef5d7a7a73b2fef38d57f25\"" Sep 4 17:43:04.117878 containerd[2079]: time="2024-09-04T17:43:04.117833370Z" level=info msg="StartContainer for \"9fafd529f01226e081e10cf8bce8ed357ac3b68faef5d7a7a73b2fef38d57f25\" returns successfully" Sep 4 17:43:08.212914 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0e089d9cb36b6332e22ad835626d20962d57f0fa01363b032ef03a4af90aee42-rootfs.mount: Deactivated successfully. Sep 4 17:43:08.227102 containerd[2079]: time="2024-09-04T17:43:08.227001886Z" level=info msg="shim disconnected" id=0e089d9cb36b6332e22ad835626d20962d57f0fa01363b032ef03a4af90aee42 namespace=k8s.io Sep 4 17:43:08.227102 containerd[2079]: time="2024-09-04T17:43:08.227095952Z" level=warning msg="cleaning up after shim disconnected" id=0e089d9cb36b6332e22ad835626d20962d57f0fa01363b032ef03a4af90aee42 namespace=k8s.io Sep 4 17:43:08.227810 containerd[2079]: time="2024-09-04T17:43:08.227115156Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:43:08.940491 kubelet[3455]: I0904 17:43:08.940460 3455 scope.go:117] "RemoveContainer" containerID="0e089d9cb36b6332e22ad835626d20962d57f0fa01363b032ef03a4af90aee42" Sep 4 17:43:08.943562 containerd[2079]: time="2024-09-04T17:43:08.943406551Z" level=info msg="CreateContainer within sandbox \"6cac7bb63ccc218e1f530235f4cc5808c3fca5e0ccfcf89580de0053fb4c38c1\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Sep 4 17:43:08.969131 containerd[2079]: time="2024-09-04T17:43:08.968846461Z" level=info msg="CreateContainer within sandbox \"6cac7bb63ccc218e1f530235f4cc5808c3fca5e0ccfcf89580de0053fb4c38c1\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"9b78e2f2bbdee8bb680be57bee69bfbd323f7a62e87cd028cab4e7f5f82a1b40\"" Sep 4 17:43:08.972006 containerd[2079]: time="2024-09-04T17:43:08.971963410Z" level=info msg="StartContainer for \"9b78e2f2bbdee8bb680be57bee69bfbd323f7a62e87cd028cab4e7f5f82a1b40\"" Sep 4 17:43:09.073086 containerd[2079]: time="2024-09-04T17:43:09.073039602Z" level=info msg="StartContainer for \"9b78e2f2bbdee8bb680be57bee69bfbd323f7a62e87cd028cab4e7f5f82a1b40\" returns successfully" Sep 4 17:43:11.234009 kubelet[3455]: E0904 17:43:11.233956 3455 controller.go:193] "Failed to update lease" err="Put \"https://172.31.20.143:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-143?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Sep 4 17:43:21.236917 kubelet[3455]: E0904 17:43:21.235874 3455 controller.go:193] "Failed to update lease" err="Put \"https://172.31.20.143:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-143?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"