Sep 12 22:51:52.967783 kernel: Linux version 6.12.47-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Fri Sep 12 20:38:35 -00 2025 Sep 12 22:51:52.967810 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=8e60d6befc710e967d67e9a1d87ced7416895090c99a765b3a00e66a62f49e40 Sep 12 22:51:52.967820 kernel: BIOS-provided physical RAM map: Sep 12 22:51:52.967827 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 12 22:51:52.967834 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable Sep 12 22:51:52.967840 kernel: BIOS-e820: [mem 0x00000000786ce000-0x000000007894dfff] reserved Sep 12 22:51:52.967848 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Sep 12 22:51:52.967855 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Sep 12 22:51:52.967864 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable Sep 12 22:51:52.967871 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Sep 12 22:51:52.967878 kernel: NX (Execute Disable) protection: active Sep 12 22:51:52.967885 kernel: APIC: Static calls initialized Sep 12 22:51:52.967892 kernel: e820: update [mem 0x768c0018-0x768c8e57] usable ==> usable Sep 12 22:51:52.967899 kernel: extended physical RAM map: Sep 12 22:51:52.967911 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Sep 12 22:51:52.967919 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000768c0017] usable Sep 12 22:51:52.967926 kernel: reserve setup_data: [mem 0x00000000768c0018-0x00000000768c8e57] usable Sep 12 22:51:52.967934 kernel: reserve setup_data: [mem 0x00000000768c8e58-0x00000000786cdfff] usable Sep 12 22:51:52.967942 kernel: reserve setup_data: [mem 0x00000000786ce000-0x000000007894dfff] reserved Sep 12 22:51:52.967949 kernel: reserve setup_data: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Sep 12 22:51:52.967957 kernel: reserve setup_data: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Sep 12 22:51:52.967964 kernel: reserve setup_data: [mem 0x00000000789de000-0x000000007c97bfff] usable Sep 12 22:51:52.967972 kernel: reserve setup_data: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Sep 12 22:51:52.967979 kernel: efi: EFI v2.7 by EDK II Sep 12 22:51:52.967989 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77003518 Sep 12 22:51:52.967996 kernel: secureboot: Secure boot disabled Sep 12 22:51:52.968004 kernel: SMBIOS 2.7 present. Sep 12 22:51:52.968011 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Sep 12 22:51:52.968019 kernel: DMI: Memory slots populated: 1/1 Sep 12 22:51:52.968026 kernel: Hypervisor detected: KVM Sep 12 22:51:52.968033 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 12 22:51:52.968041 kernel: kvm-clock: using sched offset of 5033388069 cycles Sep 12 22:51:52.968049 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 12 22:51:52.968057 kernel: tsc: Detected 2499.998 MHz processor Sep 12 22:51:52.968065 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 12 22:51:52.968075 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 12 22:51:52.968083 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 Sep 12 22:51:52.968090 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Sep 12 22:51:52.968098 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 12 22:51:52.968106 kernel: Using GB pages for direct mapping Sep 12 22:51:52.968117 kernel: ACPI: Early table checksum verification disabled Sep 12 22:51:52.968127 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) Sep 12 22:51:52.968135 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) Sep 12 22:51:52.968143 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Sep 12 22:51:52.968152 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Sep 12 22:51:52.968160 kernel: ACPI: FACS 0x00000000789D0000 000040 Sep 12 22:51:52.968168 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Sep 12 22:51:52.968176 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Sep 12 22:51:52.968184 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Sep 12 22:51:52.968194 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Sep 12 22:51:52.968202 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Sep 12 22:51:52.968210 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Sep 12 22:51:52.968218 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Sep 12 22:51:52.968226 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) Sep 12 22:51:52.968235 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] Sep 12 22:51:52.968243 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] Sep 12 22:51:52.968251 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] Sep 12 22:51:52.968778 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] Sep 12 22:51:52.968797 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] Sep 12 22:51:52.968805 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] Sep 12 22:51:52.968813 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] Sep 12 22:51:52.968822 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] Sep 12 22:51:52.968830 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] Sep 12 22:51:52.968838 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] Sep 12 22:51:52.968846 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] Sep 12 22:51:52.968854 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Sep 12 22:51:52.968863 kernel: NUMA: Initialized distance table, cnt=1 Sep 12 22:51:52.968873 kernel: NODE_DATA(0) allocated [mem 0x7a8eddc0-0x7a8f4fff] Sep 12 22:51:52.968882 kernel: Zone ranges: Sep 12 22:51:52.968890 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 12 22:51:52.968898 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] Sep 12 22:51:52.968906 kernel: Normal empty Sep 12 22:51:52.968914 kernel: Device empty Sep 12 22:51:52.968922 kernel: Movable zone start for each node Sep 12 22:51:52.968931 kernel: Early memory node ranges Sep 12 22:51:52.968939 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Sep 12 22:51:52.968949 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] Sep 12 22:51:52.968957 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] Sep 12 22:51:52.968966 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] Sep 12 22:51:52.968974 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 12 22:51:52.968982 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Sep 12 22:51:52.968990 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Sep 12 22:51:52.968998 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges Sep 12 22:51:52.969007 kernel: ACPI: PM-Timer IO Port: 0xb008 Sep 12 22:51:52.969015 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 12 22:51:52.969026 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Sep 12 22:51:52.969034 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 12 22:51:52.969042 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 12 22:51:52.969050 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 12 22:51:52.969058 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 12 22:51:52.969066 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 12 22:51:52.969074 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Sep 12 22:51:52.969083 kernel: TSC deadline timer available Sep 12 22:51:52.969091 kernel: CPU topo: Max. logical packages: 1 Sep 12 22:51:52.969099 kernel: CPU topo: Max. logical dies: 1 Sep 12 22:51:52.969109 kernel: CPU topo: Max. dies per package: 1 Sep 12 22:51:52.969117 kernel: CPU topo: Max. threads per core: 2 Sep 12 22:51:52.969125 kernel: CPU topo: Num. cores per package: 1 Sep 12 22:51:52.969133 kernel: CPU topo: Num. threads per package: 2 Sep 12 22:51:52.969141 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Sep 12 22:51:52.969149 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Sep 12 22:51:52.969157 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices Sep 12 22:51:52.969166 kernel: Booting paravirtualized kernel on KVM Sep 12 22:51:52.969174 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 12 22:51:52.969184 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Sep 12 22:51:52.969193 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Sep 12 22:51:52.969201 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Sep 12 22:51:52.969209 kernel: pcpu-alloc: [0] 0 1 Sep 12 22:51:52.969217 kernel: kvm-guest: PV spinlocks enabled Sep 12 22:51:52.969226 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 12 22:51:52.969236 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=8e60d6befc710e967d67e9a1d87ced7416895090c99a765b3a00e66a62f49e40 Sep 12 22:51:52.969245 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 12 22:51:52.969333 kernel: random: crng init done Sep 12 22:51:52.969343 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 12 22:51:52.969351 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 12 22:51:52.969359 kernel: Fallback order for Node 0: 0 Sep 12 22:51:52.969368 kernel: Built 1 zonelists, mobility grouping on. Total pages: 509451 Sep 12 22:51:52.969376 kernel: Policy zone: DMA32 Sep 12 22:51:52.969393 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 12 22:51:52.969404 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 12 22:51:52.969413 kernel: Kernel/User page tables isolation: enabled Sep 12 22:51:52.969421 kernel: ftrace: allocating 40125 entries in 157 pages Sep 12 22:51:52.969430 kernel: ftrace: allocated 157 pages with 5 groups Sep 12 22:51:52.969441 kernel: Dynamic Preempt: voluntary Sep 12 22:51:52.969450 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 12 22:51:52.969459 kernel: rcu: RCU event tracing is enabled. Sep 12 22:51:52.969468 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 12 22:51:52.969477 kernel: Trampoline variant of Tasks RCU enabled. Sep 12 22:51:52.969486 kernel: Rude variant of Tasks RCU enabled. Sep 12 22:51:52.969497 kernel: Tracing variant of Tasks RCU enabled. Sep 12 22:51:52.969506 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 12 22:51:52.969514 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 12 22:51:52.969523 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 22:51:52.969532 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 22:51:52.969541 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Sep 12 22:51:52.969550 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Sep 12 22:51:52.969559 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 12 22:51:52.969570 kernel: Console: colour dummy device 80x25 Sep 12 22:51:52.969578 kernel: printk: legacy console [tty0] enabled Sep 12 22:51:52.969587 kernel: printk: legacy console [ttyS0] enabled Sep 12 22:51:52.969596 kernel: ACPI: Core revision 20240827 Sep 12 22:51:52.969605 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Sep 12 22:51:52.969614 kernel: APIC: Switch to symmetric I/O mode setup Sep 12 22:51:52.969622 kernel: x2apic enabled Sep 12 22:51:52.969631 kernel: APIC: Switched APIC routing to: physical x2apic Sep 12 22:51:52.969639 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Sep 12 22:51:52.969651 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499998) Sep 12 22:51:52.969659 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Sep 12 22:51:52.969668 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Sep 12 22:51:52.969677 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 12 22:51:52.969685 kernel: Spectre V2 : Mitigation: Retpolines Sep 12 22:51:52.969694 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 12 22:51:52.969703 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Sep 12 22:51:52.969711 kernel: RETBleed: Vulnerable Sep 12 22:51:52.969720 kernel: Speculative Store Bypass: Vulnerable Sep 12 22:51:52.969728 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Sep 12 22:51:52.969737 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 12 22:51:52.969747 kernel: GDS: Unknown: Dependent on hypervisor status Sep 12 22:51:52.969756 kernel: active return thunk: its_return_thunk Sep 12 22:51:52.969764 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 12 22:51:52.969773 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 12 22:51:52.969781 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 12 22:51:52.969790 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 12 22:51:52.969798 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Sep 12 22:51:52.969807 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Sep 12 22:51:52.969815 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Sep 12 22:51:52.969824 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Sep 12 22:51:52.969832 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Sep 12 22:51:52.969844 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Sep 12 22:51:52.969852 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 12 22:51:52.969860 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Sep 12 22:51:52.969869 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Sep 12 22:51:52.969877 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Sep 12 22:51:52.969886 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Sep 12 22:51:52.969894 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Sep 12 22:51:52.969903 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Sep 12 22:51:52.969912 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Sep 12 22:51:52.969920 kernel: Freeing SMP alternatives memory: 32K Sep 12 22:51:52.969928 kernel: pid_max: default: 32768 minimum: 301 Sep 12 22:51:52.969939 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Sep 12 22:51:52.969948 kernel: landlock: Up and running. Sep 12 22:51:52.969956 kernel: SELinux: Initializing. Sep 12 22:51:52.969965 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 12 22:51:52.969974 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Sep 12 22:51:52.969982 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8175M CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x4) Sep 12 22:51:52.969991 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Sep 12 22:51:52.970000 kernel: signal: max sigframe size: 3632 Sep 12 22:51:52.970008 kernel: rcu: Hierarchical SRCU implementation. Sep 12 22:51:52.970017 kernel: rcu: Max phase no-delay instances is 400. Sep 12 22:51:52.972299 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Sep 12 22:51:52.972320 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 12 22:51:52.972330 kernel: smp: Bringing up secondary CPUs ... Sep 12 22:51:52.972339 kernel: smpboot: x86: Booting SMP configuration: Sep 12 22:51:52.972348 kernel: .... node #0, CPUs: #1 Sep 12 22:51:52.972358 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Sep 12 22:51:52.972375 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 12 22:51:52.972384 kernel: smp: Brought up 1 node, 2 CPUs Sep 12 22:51:52.972393 kernel: smpboot: Total of 2 processors activated (9999.99 BogoMIPS) Sep 12 22:51:52.972405 kernel: Memory: 1908060K/2037804K available (14336K kernel code, 2432K rwdata, 9992K rodata, 54084K init, 2880K bss, 125188K reserved, 0K cma-reserved) Sep 12 22:51:52.972413 kernel: devtmpfs: initialized Sep 12 22:51:52.972423 kernel: x86/mm: Memory block size: 128MB Sep 12 22:51:52.972432 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) Sep 12 22:51:52.972441 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 12 22:51:52.972450 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 12 22:51:52.972459 kernel: pinctrl core: initialized pinctrl subsystem Sep 12 22:51:52.972467 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 12 22:51:52.972476 kernel: audit: initializing netlink subsys (disabled) Sep 12 22:51:52.972487 kernel: audit: type=2000 audit(1757717510.989:1): state=initialized audit_enabled=0 res=1 Sep 12 22:51:52.972496 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 12 22:51:52.972505 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 12 22:51:52.972514 kernel: cpuidle: using governor menu Sep 12 22:51:52.972522 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 12 22:51:52.972531 kernel: dca service started, version 1.12.1 Sep 12 22:51:52.972540 kernel: PCI: Using configuration type 1 for base access Sep 12 22:51:52.972549 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 12 22:51:52.972557 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 12 22:51:52.972569 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Sep 12 22:51:52.972578 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 12 22:51:52.972586 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Sep 12 22:51:52.972595 kernel: ACPI: Added _OSI(Module Device) Sep 12 22:51:52.972604 kernel: ACPI: Added _OSI(Processor Device) Sep 12 22:51:52.972612 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 12 22:51:52.972621 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Sep 12 22:51:52.972630 kernel: ACPI: Interpreter enabled Sep 12 22:51:52.972638 kernel: ACPI: PM: (supports S0 S5) Sep 12 22:51:52.972650 kernel: ACPI: Using IOAPIC for interrupt routing Sep 12 22:51:52.972659 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 12 22:51:52.972667 kernel: PCI: Using E820 reservations for host bridge windows Sep 12 22:51:52.972676 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Sep 12 22:51:52.972685 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 12 22:51:52.972851 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Sep 12 22:51:52.972947 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Sep 12 22:51:52.973040 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Sep 12 22:51:52.973051 kernel: acpiphp: Slot [3] registered Sep 12 22:51:52.973060 kernel: acpiphp: Slot [4] registered Sep 12 22:51:52.973076 kernel: acpiphp: Slot [5] registered Sep 12 22:51:52.973085 kernel: acpiphp: Slot [6] registered Sep 12 22:51:52.973093 kernel: acpiphp: Slot [7] registered Sep 12 22:51:52.973102 kernel: acpiphp: Slot [8] registered Sep 12 22:51:52.973111 kernel: acpiphp: Slot [9] registered Sep 12 22:51:52.973119 kernel: acpiphp: Slot [10] registered Sep 12 22:51:52.973134 kernel: acpiphp: Slot [11] registered Sep 12 22:51:52.973143 kernel: acpiphp: Slot [12] registered Sep 12 22:51:52.973152 kernel: acpiphp: Slot [13] registered Sep 12 22:51:52.973160 kernel: acpiphp: Slot [14] registered Sep 12 22:51:52.973169 kernel: acpiphp: Slot [15] registered Sep 12 22:51:52.973178 kernel: acpiphp: Slot [16] registered Sep 12 22:51:52.973187 kernel: acpiphp: Slot [17] registered Sep 12 22:51:52.973195 kernel: acpiphp: Slot [18] registered Sep 12 22:51:52.973204 kernel: acpiphp: Slot [19] registered Sep 12 22:51:52.973213 kernel: acpiphp: Slot [20] registered Sep 12 22:51:52.973224 kernel: acpiphp: Slot [21] registered Sep 12 22:51:52.973233 kernel: acpiphp: Slot [22] registered Sep 12 22:51:52.973242 kernel: acpiphp: Slot [23] registered Sep 12 22:51:52.973250 kernel: acpiphp: Slot [24] registered Sep 12 22:51:52.973273 kernel: acpiphp: Slot [25] registered Sep 12 22:51:52.973282 kernel: acpiphp: Slot [26] registered Sep 12 22:51:52.973291 kernel: acpiphp: Slot [27] registered Sep 12 22:51:52.973300 kernel: acpiphp: Slot [28] registered Sep 12 22:51:52.973308 kernel: acpiphp: Slot [29] registered Sep 12 22:51:52.973320 kernel: acpiphp: Slot [30] registered Sep 12 22:51:52.973329 kernel: acpiphp: Slot [31] registered Sep 12 22:51:52.973338 kernel: PCI host bridge to bus 0000:00 Sep 12 22:51:52.973436 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 12 22:51:52.973522 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 12 22:51:52.973604 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 12 22:51:52.973684 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Sep 12 22:51:52.973763 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] Sep 12 22:51:52.973846 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 12 22:51:52.973950 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint Sep 12 22:51:52.974055 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint Sep 12 22:51:52.974152 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 conventional PCI endpoint Sep 12 22:51:52.974240 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Sep 12 22:51:52.974346 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Sep 12 22:51:52.974438 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Sep 12 22:51:52.974525 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Sep 12 22:51:52.974613 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Sep 12 22:51:52.974701 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Sep 12 22:51:52.974788 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Sep 12 22:51:52.974883 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 conventional PCI endpoint Sep 12 22:51:52.974972 kernel: pci 0000:00:03.0: BAR 0 [mem 0x80000000-0x803fffff pref] Sep 12 22:51:52.975066 kernel: pci 0000:00:03.0: ROM [mem 0xffff0000-0xffffffff pref] Sep 12 22:51:52.975156 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Sep 12 22:51:52.975250 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 PCIe Endpoint Sep 12 22:51:52.975355 kernel: pci 0000:00:04.0: BAR 0 [mem 0x80404000-0x80407fff] Sep 12 22:51:52.975450 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 PCIe Endpoint Sep 12 22:51:52.975540 kernel: pci 0000:00:05.0: BAR 0 [mem 0x80400000-0x80403fff] Sep 12 22:51:52.975556 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 12 22:51:52.975565 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 12 22:51:52.975574 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 12 22:51:52.975583 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 12 22:51:52.975592 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Sep 12 22:51:52.975601 kernel: iommu: Default domain type: Translated Sep 12 22:51:52.975610 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 12 22:51:52.975619 kernel: efivars: Registered efivars operations Sep 12 22:51:52.975628 kernel: PCI: Using ACPI for IRQ routing Sep 12 22:51:52.975640 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 12 22:51:52.975649 kernel: e820: reserve RAM buffer [mem 0x768c0018-0x77ffffff] Sep 12 22:51:52.975658 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] Sep 12 22:51:52.975666 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] Sep 12 22:51:52.975757 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Sep 12 22:51:52.975847 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Sep 12 22:51:52.975935 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Sep 12 22:51:52.975947 kernel: vgaarb: loaded Sep 12 22:51:52.975960 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Sep 12 22:51:52.975969 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Sep 12 22:51:52.975978 kernel: clocksource: Switched to clocksource kvm-clock Sep 12 22:51:52.975987 kernel: VFS: Disk quotas dquot_6.6.0 Sep 12 22:51:52.975996 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 12 22:51:52.976004 kernel: pnp: PnP ACPI init Sep 12 22:51:52.976013 kernel: pnp: PnP ACPI: found 5 devices Sep 12 22:51:52.976022 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 12 22:51:52.976031 kernel: NET: Registered PF_INET protocol family Sep 12 22:51:52.976043 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 12 22:51:52.976052 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Sep 12 22:51:52.976061 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 12 22:51:52.976070 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 12 22:51:52.976079 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Sep 12 22:51:52.976087 kernel: TCP: Hash tables configured (established 16384 bind 16384) Sep 12 22:51:52.976096 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 12 22:51:52.976105 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Sep 12 22:51:52.976114 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 12 22:51:52.976126 kernel: NET: Registered PF_XDP protocol family Sep 12 22:51:52.976210 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 12 22:51:52.977386 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 12 22:51:52.977485 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 12 22:51:52.977566 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Sep 12 22:51:52.977645 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] Sep 12 22:51:52.977742 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Sep 12 22:51:52.977754 kernel: PCI: CLS 0 bytes, default 64 Sep 12 22:51:52.977771 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 12 22:51:52.977780 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Sep 12 22:51:52.977789 kernel: clocksource: Switched to clocksource tsc Sep 12 22:51:52.977798 kernel: Initialise system trusted keyrings Sep 12 22:51:52.977807 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Sep 12 22:51:52.977816 kernel: Key type asymmetric registered Sep 12 22:51:52.977826 kernel: Asymmetric key parser 'x509' registered Sep 12 22:51:52.977835 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Sep 12 22:51:52.977844 kernel: io scheduler mq-deadline registered Sep 12 22:51:52.977855 kernel: io scheduler kyber registered Sep 12 22:51:52.977864 kernel: io scheduler bfq registered Sep 12 22:51:52.977873 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 12 22:51:52.977882 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 12 22:51:52.977891 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 12 22:51:52.977899 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 12 22:51:52.977909 kernel: i8042: Warning: Keylock active Sep 12 22:51:52.977917 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 12 22:51:52.977926 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 12 22:51:52.978029 kernel: rtc_cmos 00:00: RTC can wake from S4 Sep 12 22:51:52.978115 kernel: rtc_cmos 00:00: registered as rtc0 Sep 12 22:51:52.978199 kernel: rtc_cmos 00:00: setting system clock to 2025-09-12T22:51:52 UTC (1757717512) Sep 12 22:51:52.978330 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Sep 12 22:51:52.978363 kernel: intel_pstate: CPU model not supported Sep 12 22:51:52.978374 kernel: efifb: probing for efifb Sep 12 22:51:52.978384 kernel: efifb: framebuffer at 0x80000000, using 1876k, total 1875k Sep 12 22:51:52.978393 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Sep 12 22:51:52.978405 kernel: efifb: scrolling: redraw Sep 12 22:51:52.978415 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Sep 12 22:51:52.978424 kernel: Console: switching to colour frame buffer device 100x37 Sep 12 22:51:52.978433 kernel: fb0: EFI VGA frame buffer device Sep 12 22:51:52.978442 kernel: pstore: Using crash dump compression: deflate Sep 12 22:51:52.978452 kernel: pstore: Registered efi_pstore as persistent store backend Sep 12 22:51:52.978461 kernel: NET: Registered PF_INET6 protocol family Sep 12 22:51:52.978470 kernel: Segment Routing with IPv6 Sep 12 22:51:52.978480 kernel: In-situ OAM (IOAM) with IPv6 Sep 12 22:51:52.978491 kernel: NET: Registered PF_PACKET protocol family Sep 12 22:51:52.978501 kernel: Key type dns_resolver registered Sep 12 22:51:52.978510 kernel: IPI shorthand broadcast: enabled Sep 12 22:51:52.978520 kernel: sched_clock: Marking stable (2657002062, 193327424)->(2982594936, -132265450) Sep 12 22:51:52.978529 kernel: registered taskstats version 1 Sep 12 22:51:52.978538 kernel: Loading compiled-in X.509 certificates Sep 12 22:51:52.978547 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.47-flatcar: c3297a5801573420030c321362a802da1fd49c4e' Sep 12 22:51:52.978557 kernel: Demotion targets for Node 0: null Sep 12 22:51:52.978566 kernel: Key type .fscrypt registered Sep 12 22:51:52.978577 kernel: Key type fscrypt-provisioning registered Sep 12 22:51:52.978587 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 12 22:51:52.978596 kernel: ima: Allocated hash algorithm: sha1 Sep 12 22:51:52.978606 kernel: ima: No architecture policies found Sep 12 22:51:52.978615 kernel: clk: Disabling unused clocks Sep 12 22:51:52.978624 kernel: Warning: unable to open an initial console. Sep 12 22:51:52.978633 kernel: Freeing unused kernel image (initmem) memory: 54084K Sep 12 22:51:52.978643 kernel: Write protecting the kernel read-only data: 24576k Sep 12 22:51:52.978653 kernel: Freeing unused kernel image (rodata/data gap) memory: 248K Sep 12 22:51:52.978664 kernel: Run /init as init process Sep 12 22:51:52.978673 kernel: with arguments: Sep 12 22:51:52.978683 kernel: /init Sep 12 22:51:52.978692 kernel: with environment: Sep 12 22:51:52.978701 kernel: HOME=/ Sep 12 22:51:52.978712 kernel: TERM=linux Sep 12 22:51:52.978722 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 12 22:51:52.978733 systemd[1]: Successfully made /usr/ read-only. Sep 12 22:51:52.978748 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 22:51:52.978759 systemd[1]: Detected virtualization amazon. Sep 12 22:51:52.978768 systemd[1]: Detected architecture x86-64. Sep 12 22:51:52.978778 systemd[1]: Running in initrd. Sep 12 22:51:52.978790 systemd[1]: No hostname configured, using default hostname. Sep 12 22:51:52.978800 systemd[1]: Hostname set to . Sep 12 22:51:52.978810 systemd[1]: Initializing machine ID from VM UUID. Sep 12 22:51:52.978819 systemd[1]: Queued start job for default target initrd.target. Sep 12 22:51:52.978829 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 22:51:52.978839 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 22:51:52.978853 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 12 22:51:52.978863 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 22:51:52.978876 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 12 22:51:52.978887 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 12 22:51:52.978897 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 12 22:51:52.978907 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 12 22:51:52.978917 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 22:51:52.978927 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 22:51:52.978937 systemd[1]: Reached target paths.target - Path Units. Sep 12 22:51:52.978949 systemd[1]: Reached target slices.target - Slice Units. Sep 12 22:51:52.978959 systemd[1]: Reached target swap.target - Swaps. Sep 12 22:51:52.978969 systemd[1]: Reached target timers.target - Timer Units. Sep 12 22:51:52.978978 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 22:51:52.978988 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 22:51:52.978998 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 12 22:51:52.979008 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Sep 12 22:51:52.979017 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 22:51:52.979027 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 22:51:52.979040 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 22:51:52.979049 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 22:51:52.979059 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 12 22:51:52.979069 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 22:51:52.979079 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 12 22:51:52.979089 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Sep 12 22:51:52.979099 systemd[1]: Starting systemd-fsck-usr.service... Sep 12 22:51:52.979109 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 22:51:52.979122 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 22:51:52.979131 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:51:52.979141 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 12 22:51:52.979151 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 22:51:52.979185 systemd-journald[207]: Collecting audit messages is disabled. Sep 12 22:51:52.979211 systemd[1]: Finished systemd-fsck-usr.service. Sep 12 22:51:52.979221 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 12 22:51:52.979233 systemd-journald[207]: Journal started Sep 12 22:51:52.979269 systemd-journald[207]: Runtime Journal (/run/log/journal/ec2159a5766735cfc288a4bd9c268fcc) is 4.8M, max 38.4M, 33.6M free. Sep 12 22:51:52.968726 systemd-modules-load[209]: Inserted module 'overlay' Sep 12 22:51:52.984287 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 22:51:52.989104 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 22:51:52.996124 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:51:53.002395 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 12 22:51:53.008605 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 12 22:51:53.005134 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 12 22:51:53.018221 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 12 22:51:53.018248 kernel: Bridge firewalling registered Sep 12 22:51:53.017126 systemd-modules-load[209]: Inserted module 'br_netfilter' Sep 12 22:51:53.018315 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 22:51:53.019173 systemd-tmpfiles[222]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Sep 12 22:51:53.022582 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 22:51:53.029393 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 22:51:53.033323 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 22:51:53.039418 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 22:51:53.043650 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 22:51:53.048559 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 22:51:53.050675 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 22:51:53.052943 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 12 22:51:53.070980 dracut-cmdline[248]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=8e60d6befc710e967d67e9a1d87ced7416895090c99a765b3a00e66a62f49e40 Sep 12 22:51:53.083734 systemd-resolved[242]: Positive Trust Anchors: Sep 12 22:51:53.084409 systemd-resolved[242]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 22:51:53.084448 systemd-resolved[242]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 22:51:53.089943 systemd-resolved[242]: Defaulting to hostname 'linux'. Sep 12 22:51:53.090916 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 22:51:53.091488 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 22:51:53.165318 kernel: SCSI subsystem initialized Sep 12 22:51:53.175298 kernel: Loading iSCSI transport class v2.0-870. Sep 12 22:51:53.188297 kernel: iscsi: registered transport (tcp) Sep 12 22:51:53.210826 kernel: iscsi: registered transport (qla4xxx) Sep 12 22:51:53.210909 kernel: QLogic iSCSI HBA Driver Sep 12 22:51:53.231619 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 12 22:51:53.247892 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 22:51:53.248853 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 12 22:51:53.300981 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 12 22:51:53.302985 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 12 22:51:53.368315 kernel: raid6: avx512x4 gen() 17583 MB/s Sep 12 22:51:53.386300 kernel: raid6: avx512x2 gen() 17696 MB/s Sep 12 22:51:53.404311 kernel: raid6: avx512x1 gen() 17625 MB/s Sep 12 22:51:53.422303 kernel: raid6: avx2x4 gen() 17595 MB/s Sep 12 22:51:53.440299 kernel: raid6: avx2x2 gen() 17528 MB/s Sep 12 22:51:53.459711 kernel: raid6: avx2x1 gen() 13445 MB/s Sep 12 22:51:53.459781 kernel: raid6: using algorithm avx512x2 gen() 17696 MB/s Sep 12 22:51:53.479555 kernel: raid6: .... xor() 23721 MB/s, rmw enabled Sep 12 22:51:53.479642 kernel: raid6: using avx512x2 recovery algorithm Sep 12 22:51:53.502303 kernel: xor: automatically using best checksumming function avx Sep 12 22:51:53.676302 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 12 22:51:53.683859 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 12 22:51:53.686393 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 22:51:53.713790 systemd-udevd[457]: Using default interface naming scheme 'v255'. Sep 12 22:51:53.720600 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 22:51:53.724069 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 12 22:51:53.749180 dracut-pre-trigger[463]: rd.md=0: removing MD RAID activation Sep 12 22:51:53.779591 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 22:51:53.781846 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 22:51:53.841894 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 22:51:53.847902 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 12 22:51:53.934393 kernel: ena 0000:00:05.0: ENA device version: 0.10 Sep 12 22:51:53.934686 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Sep 12 22:51:53.938297 kernel: cryptd: max_cpu_qlen set to 1000 Sep 12 22:51:53.941288 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Sep 12 22:51:53.960307 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:37:25:44:ea:41 Sep 12 22:51:53.982542 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 22:51:53.982639 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:51:53.984529 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:51:53.987523 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:51:53.989298 kernel: AES CTR mode by8 optimization enabled Sep 12 22:51:53.999666 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 22:51:54.000612 (udev-worker)[518]: Network interface NamePolicy= disabled on kernel command line. Sep 12 22:51:54.017322 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input2 Sep 12 22:51:54.017630 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 22:51:54.019130 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:51:54.033569 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:51:54.051785 kernel: nvme nvme0: pci function 0000:00:04.0 Sep 12 22:51:54.055199 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Sep 12 22:51:54.066313 kernel: nvme nvme0: 2/0/0 default/read/poll queues Sep 12 22:51:54.081411 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 12 22:51:54.081488 kernel: GPT:9289727 != 16777215 Sep 12 22:51:54.081509 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 12 22:51:54.081528 kernel: GPT:9289727 != 16777215 Sep 12 22:51:54.081546 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 12 22:51:54.081574 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 12 22:51:54.091693 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:51:54.108302 kernel: nvme nvme0: using unchecked data buffer Sep 12 22:51:54.189330 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Sep 12 22:51:54.238688 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Sep 12 22:51:54.250380 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 12 22:51:54.268730 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Sep 12 22:51:54.269442 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Sep 12 22:51:54.281379 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Sep 12 22:51:54.282071 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 22:51:54.283352 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 22:51:54.284598 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 22:51:54.286378 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 12 22:51:54.289591 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 12 22:51:54.308249 disk-uuid[693]: Primary Header is updated. Sep 12 22:51:54.308249 disk-uuid[693]: Secondary Entries is updated. Sep 12 22:51:54.308249 disk-uuid[693]: Secondary Header is updated. Sep 12 22:51:54.316334 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 12 22:51:54.322541 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 12 22:51:55.334282 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Sep 12 22:51:55.334357 disk-uuid[695]: The operation has completed successfully. Sep 12 22:51:55.482976 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 12 22:51:55.483112 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 12 22:51:55.520212 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 12 22:51:55.541171 sh[961]: Success Sep 12 22:51:55.572854 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 12 22:51:55.572937 kernel: device-mapper: uevent: version 1.0.3 Sep 12 22:51:55.575756 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Sep 12 22:51:55.592289 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" Sep 12 22:51:55.677546 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 12 22:51:55.682399 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 12 22:51:55.695073 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 12 22:51:55.714295 kernel: BTRFS: device fsid 5d2ab445-1154-4e47-9d7e-ff4b81d84474 devid 1 transid 37 /dev/mapper/usr (254:0) scanned by mount (984) Sep 12 22:51:55.719620 kernel: BTRFS info (device dm-0): first mount of filesystem 5d2ab445-1154-4e47-9d7e-ff4b81d84474 Sep 12 22:51:55.719695 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Sep 12 22:51:55.734589 kernel: BTRFS info (device dm-0): enabling ssd optimizations Sep 12 22:51:55.734674 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 12 22:51:55.734689 kernel: BTRFS info (device dm-0): enabling free space tree Sep 12 22:51:55.744188 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 12 22:51:55.745173 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Sep 12 22:51:55.745699 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 12 22:51:55.746481 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 12 22:51:55.747897 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 12 22:51:55.778318 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1017) Sep 12 22:51:55.785410 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem fd5cdc72-255e-4ed2-8d25-c5e581a08827 Sep 12 22:51:55.785494 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 12 22:51:55.801642 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 12 22:51:55.803103 kernel: BTRFS info (device nvme0n1p6): enabling free space tree Sep 12 22:51:55.812571 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem fd5cdc72-255e-4ed2-8d25-c5e581a08827 Sep 12 22:51:55.814080 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 12 22:51:55.817127 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 12 22:51:55.874481 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 22:51:55.877296 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 22:51:55.929342 systemd-networkd[1153]: lo: Link UP Sep 12 22:51:55.929354 systemd-networkd[1153]: lo: Gained carrier Sep 12 22:51:55.933965 systemd-networkd[1153]: Enumeration completed Sep 12 22:51:55.934231 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 22:51:55.935136 systemd-networkd[1153]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:51:55.935142 systemd-networkd[1153]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 12 22:51:55.936806 systemd[1]: Reached target network.target - Network. Sep 12 22:51:55.938736 systemd-networkd[1153]: eth0: Link UP Sep 12 22:51:55.938741 systemd-networkd[1153]: eth0: Gained carrier Sep 12 22:51:55.938760 systemd-networkd[1153]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:51:55.953368 systemd-networkd[1153]: eth0: DHCPv4 address 172.31.29.105/20, gateway 172.31.16.1 acquired from 172.31.16.1 Sep 12 22:51:56.256074 ignition[1086]: Ignition 2.22.0 Sep 12 22:51:56.256097 ignition[1086]: Stage: fetch-offline Sep 12 22:51:56.256283 ignition[1086]: no configs at "/usr/lib/ignition/base.d" Sep 12 22:51:56.256295 ignition[1086]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 12 22:51:56.256612 ignition[1086]: Ignition finished successfully Sep 12 22:51:56.258140 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 22:51:56.260180 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Sep 12 22:51:56.295224 ignition[1164]: Ignition 2.22.0 Sep 12 22:51:56.295249 ignition[1164]: Stage: fetch Sep 12 22:51:56.295669 ignition[1164]: no configs at "/usr/lib/ignition/base.d" Sep 12 22:51:56.295682 ignition[1164]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 12 22:51:56.295801 ignition[1164]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 12 22:51:56.303660 ignition[1164]: PUT result: OK Sep 12 22:51:56.305332 ignition[1164]: parsed url from cmdline: "" Sep 12 22:51:56.305343 ignition[1164]: no config URL provided Sep 12 22:51:56.305351 ignition[1164]: reading system config file "/usr/lib/ignition/user.ign" Sep 12 22:51:56.305361 ignition[1164]: no config at "/usr/lib/ignition/user.ign" Sep 12 22:51:56.305377 ignition[1164]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 12 22:51:56.306001 ignition[1164]: PUT result: OK Sep 12 22:51:56.306056 ignition[1164]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Sep 12 22:51:56.306706 ignition[1164]: GET result: OK Sep 12 22:51:56.306775 ignition[1164]: parsing config with SHA512: 6eb2f6d3716132b758fa5e4422b4472ec6684f71f1c9ee5c08d26d9c6acd53863d52bb8efa4b7356f1cd7edbce82401bb6a77be3dad54c01534935f42dd4cd44 Sep 12 22:51:56.310890 unknown[1164]: fetched base config from "system" Sep 12 22:51:56.311462 unknown[1164]: fetched base config from "system" Sep 12 22:51:56.312163 ignition[1164]: fetch: fetch complete Sep 12 22:51:56.311472 unknown[1164]: fetched user config from "aws" Sep 12 22:51:56.312168 ignition[1164]: fetch: fetch passed Sep 12 22:51:56.312237 ignition[1164]: Ignition finished successfully Sep 12 22:51:56.316879 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Sep 12 22:51:56.319071 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 12 22:51:56.352965 ignition[1171]: Ignition 2.22.0 Sep 12 22:51:56.352990 ignition[1171]: Stage: kargs Sep 12 22:51:56.353418 ignition[1171]: no configs at "/usr/lib/ignition/base.d" Sep 12 22:51:56.353431 ignition[1171]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 12 22:51:56.353533 ignition[1171]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 12 22:51:56.361610 ignition[1171]: PUT result: OK Sep 12 22:51:56.369862 ignition[1171]: kargs: kargs passed Sep 12 22:51:56.369950 ignition[1171]: Ignition finished successfully Sep 12 22:51:56.372208 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 12 22:51:56.374122 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 12 22:51:56.408699 ignition[1177]: Ignition 2.22.0 Sep 12 22:51:56.408724 ignition[1177]: Stage: disks Sep 12 22:51:56.409105 ignition[1177]: no configs at "/usr/lib/ignition/base.d" Sep 12 22:51:56.409118 ignition[1177]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 12 22:51:56.409228 ignition[1177]: PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 12 22:51:56.410583 ignition[1177]: PUT result: OK Sep 12 22:51:56.412977 ignition[1177]: disks: disks passed Sep 12 22:51:56.413060 ignition[1177]: Ignition finished successfully Sep 12 22:51:56.415153 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 12 22:51:56.415819 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 12 22:51:56.416177 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 12 22:51:56.416872 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 22:51:56.417448 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 22:51:56.417994 systemd[1]: Reached target basic.target - Basic System. Sep 12 22:51:56.419750 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 12 22:51:56.462743 systemd-fsck[1185]: ROOT: clean, 15/553520 files, 52789/553472 blocks Sep 12 22:51:56.466477 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 12 22:51:56.468405 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 12 22:51:56.619286 kernel: EXT4-fs (nvme0n1p9): mounted filesystem d027afc5-396a-49bf-a5be-60ddd42cb089 r/w with ordered data mode. Quota mode: none. Sep 12 22:51:56.620107 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 12 22:51:56.621126 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 12 22:51:56.623174 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 22:51:56.626362 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 12 22:51:56.627106 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 12 22:51:56.627162 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 12 22:51:56.627191 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 22:51:56.644431 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 12 22:51:56.646834 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 12 22:51:56.660299 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1204) Sep 12 22:51:56.666095 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem fd5cdc72-255e-4ed2-8d25-c5e581a08827 Sep 12 22:51:56.666832 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 12 22:51:56.675971 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 12 22:51:56.676051 kernel: BTRFS info (device nvme0n1p6): enabling free space tree Sep 12 22:51:56.678408 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 22:51:56.911250 initrd-setup-root[1228]: cut: /sysroot/etc/passwd: No such file or directory Sep 12 22:51:56.918284 initrd-setup-root[1235]: cut: /sysroot/etc/group: No such file or directory Sep 12 22:51:56.926542 initrd-setup-root[1242]: cut: /sysroot/etc/shadow: No such file or directory Sep 12 22:51:56.932954 initrd-setup-root[1249]: cut: /sysroot/etc/gshadow: No such file or directory Sep 12 22:51:57.082371 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 12 22:51:57.085377 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 12 22:51:57.089497 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 12 22:51:57.109550 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 12 22:51:57.113748 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem fd5cdc72-255e-4ed2-8d25-c5e581a08827 Sep 12 22:51:57.139182 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 12 22:51:57.152228 ignition[1317]: INFO : Ignition 2.22.0 Sep 12 22:51:57.152228 ignition[1317]: INFO : Stage: mount Sep 12 22:51:57.153954 ignition[1317]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 22:51:57.153954 ignition[1317]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 12 22:51:57.153954 ignition[1317]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 12 22:51:57.155416 ignition[1317]: INFO : PUT result: OK Sep 12 22:51:57.157847 ignition[1317]: INFO : mount: mount passed Sep 12 22:51:57.159289 ignition[1317]: INFO : Ignition finished successfully Sep 12 22:51:57.157899 systemd-networkd[1153]: eth0: Gained IPv6LL Sep 12 22:51:57.162034 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 12 22:51:57.163678 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 12 22:51:57.185958 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 12 22:51:57.214401 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1328) Sep 12 22:51:57.218826 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem fd5cdc72-255e-4ed2-8d25-c5e581a08827 Sep 12 22:51:57.220510 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Sep 12 22:51:57.229117 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Sep 12 22:51:57.229200 kernel: BTRFS info (device nvme0n1p6): enabling free space tree Sep 12 22:51:57.231510 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 12 22:51:57.267724 ignition[1344]: INFO : Ignition 2.22.0 Sep 12 22:51:57.267724 ignition[1344]: INFO : Stage: files Sep 12 22:51:57.269460 ignition[1344]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 22:51:57.269460 ignition[1344]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 12 22:51:57.269460 ignition[1344]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 12 22:51:57.269460 ignition[1344]: INFO : PUT result: OK Sep 12 22:51:57.272027 ignition[1344]: DEBUG : files: compiled without relabeling support, skipping Sep 12 22:51:57.273373 ignition[1344]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 12 22:51:57.273373 ignition[1344]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 12 22:51:57.276522 ignition[1344]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 12 22:51:57.277646 ignition[1344]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 12 22:51:57.278823 ignition[1344]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 12 22:51:57.278116 unknown[1344]: wrote ssh authorized keys file for user: core Sep 12 22:51:57.281619 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 12 22:51:57.282535 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Sep 12 22:51:57.353335 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 12 22:51:57.770969 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Sep 12 22:51:57.770969 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 12 22:51:57.772873 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 12 22:51:57.772873 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 12 22:51:57.772873 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 12 22:51:57.772873 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 22:51:57.772873 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 12 22:51:57.772873 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 22:51:57.772873 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 12 22:51:57.777891 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 22:51:57.777891 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 12 22:51:57.777891 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 12 22:51:57.781007 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 12 22:51:57.781007 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 12 22:51:57.781007 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Sep 12 22:51:58.264080 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 12 22:51:59.218098 ignition[1344]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Sep 12 22:51:59.218098 ignition[1344]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 12 22:51:59.221082 ignition[1344]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 22:51:59.224958 ignition[1344]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 12 22:51:59.224958 ignition[1344]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 12 22:51:59.224958 ignition[1344]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Sep 12 22:51:59.227432 ignition[1344]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Sep 12 22:51:59.227432 ignition[1344]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 12 22:51:59.227432 ignition[1344]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 12 22:51:59.227432 ignition[1344]: INFO : files: files passed Sep 12 22:51:59.227432 ignition[1344]: INFO : Ignition finished successfully Sep 12 22:51:59.226804 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 12 22:51:59.230396 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 12 22:51:59.233722 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 12 22:51:59.245880 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 12 22:51:59.246911 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 12 22:51:59.257804 initrd-setup-root-after-ignition[1375]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 22:51:59.257804 initrd-setup-root-after-ignition[1375]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 12 22:51:59.260491 initrd-setup-root-after-ignition[1379]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 12 22:51:59.260842 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 22:51:59.262591 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 12 22:51:59.264771 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 12 22:51:59.316017 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 12 22:51:59.316162 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 12 22:51:59.317595 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 12 22:51:59.318788 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 12 22:51:59.319721 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 12 22:51:59.320986 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 12 22:51:59.360029 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 22:51:59.362656 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 12 22:51:59.383654 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 12 22:51:59.384879 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 22:51:59.385412 systemd[1]: Stopped target timers.target - Timer Units. Sep 12 22:51:59.385868 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 12 22:51:59.386001 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 12 22:51:59.386894 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 12 22:51:59.387766 systemd[1]: Stopped target basic.target - Basic System. Sep 12 22:51:59.388578 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 12 22:51:59.389433 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 12 22:51:59.390096 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 12 22:51:59.391067 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Sep 12 22:51:59.391885 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 12 22:51:59.392807 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 12 22:51:59.393700 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 12 22:51:59.394465 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 12 22:51:59.395321 systemd[1]: Stopped target swap.target - Swaps. Sep 12 22:51:59.396652 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 12 22:51:59.396912 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 12 22:51:59.397896 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 12 22:51:59.398765 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 22:51:59.399440 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 12 22:51:59.399585 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 22:51:59.400291 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 12 22:51:59.400581 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 12 22:51:59.401685 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 12 22:51:59.401938 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 12 22:51:59.402719 systemd[1]: ignition-files.service: Deactivated successfully. Sep 12 22:51:59.402915 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 12 22:51:59.405368 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 12 22:51:59.406376 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 12 22:51:59.406562 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 22:51:59.412883 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 12 22:51:59.415810 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 12 22:51:59.416100 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 22:51:59.418048 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 12 22:51:59.418278 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 12 22:51:59.428365 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 12 22:51:59.428494 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 12 22:51:59.452037 ignition[1399]: INFO : Ignition 2.22.0 Sep 12 22:51:59.452037 ignition[1399]: INFO : Stage: umount Sep 12 22:51:59.452037 ignition[1399]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 12 22:51:59.452037 ignition[1399]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Sep 12 22:51:59.452037 ignition[1399]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Sep 12 22:51:59.455845 ignition[1399]: INFO : PUT result: OK Sep 12 22:51:59.453904 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 12 22:51:59.462945 ignition[1399]: INFO : umount: umount passed Sep 12 22:51:59.462945 ignition[1399]: INFO : Ignition finished successfully Sep 12 22:51:59.466074 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 12 22:51:59.466220 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 12 22:51:59.467367 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 12 22:51:59.467440 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 12 22:51:59.467987 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 12 22:51:59.468049 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 12 22:51:59.468863 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 12 22:51:59.468922 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Sep 12 22:51:59.469682 systemd[1]: Stopped target network.target - Network. Sep 12 22:51:59.470302 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 12 22:51:59.470365 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 12 22:51:59.470988 systemd[1]: Stopped target paths.target - Path Units. Sep 12 22:51:59.471681 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 12 22:51:59.474415 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 22:51:59.475599 systemd[1]: Stopped target slices.target - Slice Units. Sep 12 22:51:59.476054 systemd[1]: Stopped target sockets.target - Socket Units. Sep 12 22:51:59.477040 systemd[1]: iscsid.socket: Deactivated successfully. Sep 12 22:51:59.477104 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 12 22:51:59.477776 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 12 22:51:59.477826 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 12 22:51:59.479875 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 12 22:51:59.479980 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 12 22:51:59.480819 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 12 22:51:59.480884 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 12 22:51:59.481708 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 12 22:51:59.482757 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 12 22:51:59.486353 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 12 22:51:59.486515 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 12 22:51:59.490017 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Sep 12 22:51:59.490489 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 12 22:51:59.490557 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 22:51:59.494639 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Sep 12 22:51:59.495037 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 12 22:51:59.495186 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 12 22:51:59.497527 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Sep 12 22:51:59.498634 systemd[1]: Stopped target network-pre.target - Preparation for Network. Sep 12 22:51:59.499155 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 12 22:51:59.499207 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 12 22:51:59.501147 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 12 22:51:59.501938 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 12 22:51:59.502009 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 12 22:51:59.502645 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 12 22:51:59.502707 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 12 22:51:59.506453 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 12 22:51:59.506545 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 12 22:51:59.507576 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 22:51:59.514102 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 12 22:51:59.524694 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 12 22:51:59.524910 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 22:51:59.527027 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 12 22:51:59.527123 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 12 22:51:59.529113 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 12 22:51:59.529177 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 22:51:59.530002 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 12 22:51:59.530073 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 12 22:51:59.531236 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 12 22:51:59.533127 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 12 22:51:59.534403 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 12 22:51:59.534479 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 12 22:51:59.538171 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 12 22:51:59.538737 systemd[1]: systemd-network-generator.service: Deactivated successfully. Sep 12 22:51:59.538821 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 22:51:59.541057 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 12 22:51:59.541130 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 22:51:59.543615 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 22:51:59.543695 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:51:59.547396 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 12 22:51:59.547560 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 12 22:51:59.566222 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 12 22:51:59.566639 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 12 22:51:59.677036 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 12 22:51:59.677182 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 12 22:51:59.685718 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 12 22:51:59.686398 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 12 22:51:59.686513 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 12 22:51:59.688384 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 12 22:51:59.719462 systemd[1]: Switching root. Sep 12 22:51:59.761038 systemd-journald[207]: Journal stopped Sep 12 22:52:01.818888 systemd-journald[207]: Received SIGTERM from PID 1 (systemd). Sep 12 22:52:01.818969 kernel: SELinux: policy capability network_peer_controls=1 Sep 12 22:52:01.818999 kernel: SELinux: policy capability open_perms=1 Sep 12 22:52:01.819025 kernel: SELinux: policy capability extended_socket_class=1 Sep 12 22:52:01.819045 kernel: SELinux: policy capability always_check_network=0 Sep 12 22:52:01.819064 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 12 22:52:01.819084 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 12 22:52:01.819109 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 12 22:52:01.819128 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 12 22:52:01.819147 kernel: SELinux: policy capability userspace_initial_context=0 Sep 12 22:52:01.819167 kernel: audit: type=1403 audit(1757717520.005:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 12 22:52:01.819189 systemd[1]: Successfully loaded SELinux policy in 90.615ms. Sep 12 22:52:01.819226 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 20.092ms. Sep 12 22:52:01.819249 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Sep 12 22:52:01.839752 systemd[1]: Detected virtualization amazon. Sep 12 22:52:01.839785 systemd[1]: Detected architecture x86-64. Sep 12 22:52:01.839813 systemd[1]: Detected first boot. Sep 12 22:52:01.839835 systemd[1]: Initializing machine ID from VM UUID. Sep 12 22:52:01.839861 zram_generator::config[1442]: No configuration found. Sep 12 22:52:01.839888 kernel: Guest personality initialized and is inactive Sep 12 22:52:01.839910 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Sep 12 22:52:01.839932 kernel: Initialized host personality Sep 12 22:52:01.839951 kernel: NET: Registered PF_VSOCK protocol family Sep 12 22:52:01.839972 systemd[1]: Populated /etc with preset unit settings. Sep 12 22:52:01.839997 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Sep 12 22:52:01.840019 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 12 22:52:01.840041 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 12 22:52:01.840062 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 12 22:52:01.840083 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 12 22:52:01.840103 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 12 22:52:01.840124 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 12 22:52:01.840145 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 12 22:52:01.840165 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 12 22:52:01.840189 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 12 22:52:01.840210 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 12 22:52:01.840233 systemd[1]: Created slice user.slice - User and Session Slice. Sep 12 22:52:01.840254 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 12 22:52:01.840302 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 12 22:52:01.840324 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 12 22:52:01.840344 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 12 22:52:01.840367 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 12 22:52:01.840391 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 12 22:52:01.840412 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Sep 12 22:52:01.840433 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 12 22:52:01.840453 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 12 22:52:01.840475 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 12 22:52:01.840495 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 12 22:52:01.840516 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 12 22:52:01.840537 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 12 22:52:01.840560 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 12 22:52:01.840580 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 12 22:52:01.840601 systemd[1]: Reached target slices.target - Slice Units. Sep 12 22:52:01.840622 systemd[1]: Reached target swap.target - Swaps. Sep 12 22:52:01.840643 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 12 22:52:01.840663 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 12 22:52:01.840684 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Sep 12 22:52:01.840706 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 12 22:52:01.840727 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 12 22:52:01.840748 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 12 22:52:01.840771 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 12 22:52:01.840793 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 12 22:52:01.840813 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 12 22:52:01.840834 systemd[1]: Mounting media.mount - External Media Directory... Sep 12 22:52:01.840856 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:01.840877 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 12 22:52:01.840898 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 12 22:52:01.840920 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 12 22:52:01.840944 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 12 22:52:01.840965 systemd[1]: Reached target machines.target - Containers. Sep 12 22:52:01.840986 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 12 22:52:01.841007 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:52:01.841028 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 12 22:52:01.841048 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 12 22:52:01.841069 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 22:52:01.841090 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 22:52:01.841111 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 22:52:01.841135 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 12 22:52:01.841158 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 22:52:01.841179 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 12 22:52:01.841201 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 12 22:52:01.841221 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 12 22:52:01.841242 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 12 22:52:01.847330 systemd[1]: Stopped systemd-fsck-usr.service. Sep 12 22:52:01.847379 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:52:01.847410 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 12 22:52:01.847431 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 12 22:52:01.847452 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 12 22:52:01.847476 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 12 22:52:01.847496 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Sep 12 22:52:01.847516 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 12 22:52:01.847540 systemd[1]: verity-setup.service: Deactivated successfully. Sep 12 22:52:01.847560 systemd[1]: Stopped verity-setup.service. Sep 12 22:52:01.847582 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:01.847606 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 12 22:52:01.847629 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 12 22:52:01.847649 systemd[1]: Mounted media.mount - External Media Directory. Sep 12 22:52:01.847669 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 12 22:52:01.847689 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 12 22:52:01.847709 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 12 22:52:01.847729 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 12 22:52:01.847750 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 12 22:52:01.847769 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 12 22:52:01.847789 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 22:52:01.847812 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 22:52:01.847835 kernel: loop: module loaded Sep 12 22:52:01.859668 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 22:52:01.859715 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 22:52:01.859736 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 22:52:01.859755 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 22:52:01.859774 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 12 22:52:01.859795 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 12 22:52:01.859814 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 12 22:52:01.859841 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 12 22:52:01.859860 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 12 22:52:01.859880 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 12 22:52:01.859899 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 12 22:52:01.859918 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Sep 12 22:52:01.859938 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 12 22:52:01.859958 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:52:01.859980 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 12 22:52:01.860000 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 22:52:01.860062 systemd-journald[1524]: Collecting audit messages is disabled. Sep 12 22:52:01.860109 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 12 22:52:01.860129 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 22:52:01.860151 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 12 22:52:01.860171 kernel: fuse: init (API version 7.41) Sep 12 22:52:01.860191 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 12 22:52:01.860210 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 12 22:52:01.860230 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 12 22:52:01.860249 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Sep 12 22:52:01.860353 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 12 22:52:01.860380 systemd-journald[1524]: Journal started Sep 12 22:52:01.860424 systemd-journald[1524]: Runtime Journal (/run/log/journal/ec2159a5766735cfc288a4bd9c268fcc) is 4.8M, max 38.4M, 33.6M free. Sep 12 22:52:01.904802 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 12 22:52:01.904887 kernel: ACPI: bus type drm_connector registered Sep 12 22:52:01.297121 systemd[1]: Queued start job for default target multi-user.target. Sep 12 22:52:01.314507 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Sep 12 22:52:01.315009 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 12 22:52:01.911581 systemd[1]: Started systemd-journald.service - Journal Service. Sep 12 22:52:01.916988 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 22:52:01.917778 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 22:52:01.921579 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 12 22:52:01.969482 kernel: loop0: detected capacity change from 0 to 110984 Sep 12 22:52:01.952819 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 12 22:52:01.989468 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 12 22:52:01.992132 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 12 22:52:02.003080 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 12 22:52:02.011298 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 12 22:52:02.017322 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Sep 12 22:52:02.024458 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 12 22:52:02.054379 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 12 22:52:02.070950 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 12 22:52:02.081335 systemd-journald[1524]: Time spent on flushing to /var/log/journal/ec2159a5766735cfc288a4bd9c268fcc is 80.372ms for 1023 entries. Sep 12 22:52:02.081335 systemd-journald[1524]: System Journal (/var/log/journal/ec2159a5766735cfc288a4bd9c268fcc) is 8M, max 195.6M, 187.6M free. Sep 12 22:52:02.169967 systemd-journald[1524]: Received client request to flush runtime journal. Sep 12 22:52:02.170032 kernel: loop1: detected capacity change from 0 to 229808 Sep 12 22:52:02.155733 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 12 22:52:02.160668 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 12 22:52:02.173899 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 12 22:52:02.199440 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Sep 12 22:52:02.215501 systemd-tmpfiles[1591]: ACLs are not supported, ignoring. Sep 12 22:52:02.215533 systemd-tmpfiles[1591]: ACLs are not supported, ignoring. Sep 12 22:52:02.221650 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 12 22:52:02.317928 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 12 22:52:02.384412 kernel: loop2: detected capacity change from 0 to 72368 Sep 12 22:52:02.514316 kernel: loop3: detected capacity change from 0 to 128016 Sep 12 22:52:02.579344 kernel: loop4: detected capacity change from 0 to 110984 Sep 12 22:52:02.625300 kernel: loop5: detected capacity change from 0 to 229808 Sep 12 22:52:02.658310 kernel: loop6: detected capacity change from 0 to 72368 Sep 12 22:52:02.676369 kernel: loop7: detected capacity change from 0 to 128016 Sep 12 22:52:02.701925 (sd-merge)[1602]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Sep 12 22:52:02.703029 (sd-merge)[1602]: Merged extensions into '/usr'. Sep 12 22:52:02.711450 systemd[1]: Reload requested from client PID 1549 ('systemd-sysext') (unit systemd-sysext.service)... Sep 12 22:52:02.711636 systemd[1]: Reloading... Sep 12 22:52:02.798174 ldconfig[1542]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 12 22:52:02.866314 zram_generator::config[1637]: No configuration found. Sep 12 22:52:03.144751 systemd[1]: Reloading finished in 432 ms. Sep 12 22:52:03.169916 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 12 22:52:03.170804 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 12 22:52:03.171655 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 12 22:52:03.183619 systemd[1]: Starting ensure-sysext.service... Sep 12 22:52:03.185173 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Sep 12 22:52:03.188576 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 12 22:52:03.209520 systemd[1]: Reload requested from client PID 1681 ('systemctl') (unit ensure-sysext.service)... Sep 12 22:52:03.209690 systemd[1]: Reloading... Sep 12 22:52:03.219359 systemd-tmpfiles[1682]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Sep 12 22:52:03.219784 systemd-tmpfiles[1682]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Sep 12 22:52:03.220188 systemd-tmpfiles[1682]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 12 22:52:03.221228 systemd-tmpfiles[1682]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 12 22:52:03.226444 systemd-tmpfiles[1682]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 12 22:52:03.226921 systemd-tmpfiles[1682]: ACLs are not supported, ignoring. Sep 12 22:52:03.227010 systemd-tmpfiles[1682]: ACLs are not supported, ignoring. Sep 12 22:52:03.237563 systemd-tmpfiles[1682]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 22:52:03.237584 systemd-tmpfiles[1682]: Skipping /boot Sep 12 22:52:03.255356 systemd-tmpfiles[1682]: Detected autofs mount point /boot during canonicalization of boot. Sep 12 22:52:03.255373 systemd-tmpfiles[1682]: Skipping /boot Sep 12 22:52:03.265129 systemd-udevd[1683]: Using default interface naming scheme 'v255'. Sep 12 22:52:03.319292 zram_generator::config[1709]: No configuration found. Sep 12 22:52:03.550067 (udev-worker)[1745]: Network interface NamePolicy= disabled on kernel command line. Sep 12 22:52:03.696283 kernel: mousedev: PS/2 mouse device common for all mice Sep 12 22:52:03.711297 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Sep 12 22:52:03.723295 kernel: ACPI: button: Power Button [PWRF] Sep 12 22:52:03.725404 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input4 Sep 12 22:52:03.727296 kernel: ACPI: button: Sleep Button [SLPF] Sep 12 22:52:03.854294 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Sep 12 22:52:03.932945 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Sep 12 22:52:03.933811 systemd[1]: Reloading finished in 723 ms. Sep 12 22:52:03.946370 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 12 22:52:03.947552 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Sep 12 22:52:04.015191 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:04.019582 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 22:52:04.025578 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 12 22:52:04.027599 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:52:04.033397 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 22:52:04.039904 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 22:52:04.044428 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 12 22:52:04.045217 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:52:04.045513 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:52:04.050663 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 12 22:52:04.059838 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 12 22:52:04.067186 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 12 22:52:04.083728 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 12 22:52:04.084434 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:04.102943 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 22:52:04.103181 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 22:52:04.105025 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 22:52:04.105296 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 22:52:04.117127 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:04.117463 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:52:04.121164 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 12 22:52:04.125716 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 12 22:52:04.127416 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:52:04.128442 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:52:04.128605 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:04.140531 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 12 22:52:04.142165 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 12 22:52:04.146326 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:04.146793 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 12 22:52:04.151779 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 12 22:52:04.152878 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 12 22:52:04.153240 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Sep 12 22:52:04.153548 systemd[1]: Reached target time-set.target - System Time Set. Sep 12 22:52:04.155503 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 12 22:52:04.174473 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 12 22:52:04.176549 systemd[1]: Finished ensure-sysext.service. Sep 12 22:52:04.194140 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 12 22:52:04.196999 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 12 22:52:04.200871 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 12 22:52:04.201116 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 12 22:52:04.202016 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 12 22:52:04.209136 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 12 22:52:04.212897 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 12 22:52:04.224506 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 12 22:52:04.230184 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 12 22:52:04.232795 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 12 22:52:04.235471 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 12 22:52:04.254346 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Sep 12 22:52:04.257947 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 12 22:52:04.270356 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 12 22:52:04.271390 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 12 22:52:04.289412 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 12 22:52:04.300192 augenrules[1935]: No rules Sep 12 22:52:04.303723 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 22:52:04.304472 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 22:52:04.315347 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 12 22:52:04.328767 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 12 22:52:04.334654 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:52:04.410939 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 12 22:52:04.411223 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:52:04.415081 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Sep 12 22:52:04.420603 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 12 22:52:04.519768 systemd-networkd[1897]: lo: Link UP Sep 12 22:52:04.519786 systemd-networkd[1897]: lo: Gained carrier Sep 12 22:52:04.521646 systemd-networkd[1897]: Enumeration completed Sep 12 22:52:04.521787 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 12 22:52:04.523769 systemd-networkd[1897]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:52:04.523775 systemd-networkd[1897]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 12 22:52:04.526865 systemd-networkd[1897]: eth0: Link UP Sep 12 22:52:04.527073 systemd-networkd[1897]: eth0: Gained carrier Sep 12 22:52:04.527114 systemd-networkd[1897]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 12 22:52:04.527508 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Sep 12 22:52:04.533210 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 12 22:52:04.536342 systemd-networkd[1897]: eth0: DHCPv4 address 172.31.29.105/20, gateway 172.31.16.1 acquired from 172.31.16.1 Sep 12 22:52:04.549704 systemd-resolved[1898]: Positive Trust Anchors: Sep 12 22:52:04.549729 systemd-resolved[1898]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 12 22:52:04.549775 systemd-resolved[1898]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Sep 12 22:52:04.555341 systemd-resolved[1898]: Defaulting to hostname 'linux'. Sep 12 22:52:04.557584 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 12 22:52:04.558401 systemd[1]: Reached target network.target - Network. Sep 12 22:52:04.559087 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 12 22:52:04.572740 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Sep 12 22:52:04.575491 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 12 22:52:04.576201 systemd[1]: Reached target sysinit.target - System Initialization. Sep 12 22:52:04.576825 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 12 22:52:04.577394 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 12 22:52:04.577817 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Sep 12 22:52:04.578718 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 12 22:52:04.579200 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 12 22:52:04.579629 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 12 22:52:04.580010 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 12 22:52:04.580060 systemd[1]: Reached target paths.target - Path Units. Sep 12 22:52:04.580495 systemd[1]: Reached target timers.target - Timer Units. Sep 12 22:52:04.582000 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 12 22:52:04.584009 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 12 22:52:04.587114 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Sep 12 22:52:04.587849 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Sep 12 22:52:04.588587 systemd[1]: Reached target ssh-access.target - SSH Access Available. Sep 12 22:52:04.591296 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 12 22:52:04.592210 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Sep 12 22:52:04.593573 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 12 22:52:04.594908 systemd[1]: Reached target sockets.target - Socket Units. Sep 12 22:52:04.595377 systemd[1]: Reached target basic.target - Basic System. Sep 12 22:52:04.595827 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 12 22:52:04.595883 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 12 22:52:04.597039 systemd[1]: Starting containerd.service - containerd container runtime... Sep 12 22:52:04.601441 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Sep 12 22:52:04.604422 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 12 22:52:04.608722 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 12 22:52:04.616192 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 12 22:52:04.626591 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 12 22:52:04.628417 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 12 22:52:04.630533 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Sep 12 22:52:04.638594 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 12 22:52:04.644563 systemd[1]: Started ntpd.service - Network Time Service. Sep 12 22:52:04.657743 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 12 22:52:04.666403 systemd[1]: Starting setup-oem.service - Setup OEM... Sep 12 22:52:04.670306 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 12 22:52:04.678717 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 12 22:52:04.694191 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 12 22:52:04.697424 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 12 22:52:04.698172 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 12 22:52:04.707299 google_oslogin_nss_cache[1974]: oslogin_cache_refresh[1974]: Refreshing passwd entry cache Sep 12 22:52:04.704589 systemd[1]: Starting update-engine.service - Update Engine... Sep 12 22:52:04.701075 oslogin_cache_refresh[1974]: Refreshing passwd entry cache Sep 12 22:52:04.709148 jq[1972]: false Sep 12 22:52:04.710445 oslogin_cache_refresh[1974]: Failure getting users, quitting Sep 12 22:52:04.711380 google_oslogin_nss_cache[1974]: oslogin_cache_refresh[1974]: Failure getting users, quitting Sep 12 22:52:04.711380 google_oslogin_nss_cache[1974]: oslogin_cache_refresh[1974]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 12 22:52:04.711380 google_oslogin_nss_cache[1974]: oslogin_cache_refresh[1974]: Refreshing group entry cache Sep 12 22:52:04.710467 oslogin_cache_refresh[1974]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Sep 12 22:52:04.710523 oslogin_cache_refresh[1974]: Refreshing group entry cache Sep 12 22:52:04.712095 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 12 22:52:04.715945 oslogin_cache_refresh[1974]: Failure getting groups, quitting Sep 12 22:52:04.737461 google_oslogin_nss_cache[1974]: oslogin_cache_refresh[1974]: Failure getting groups, quitting Sep 12 22:52:04.737461 google_oslogin_nss_cache[1974]: oslogin_cache_refresh[1974]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 12 22:52:04.717351 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 12 22:52:04.715962 oslogin_cache_refresh[1974]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Sep 12 22:52:04.718783 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 12 22:52:04.720520 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 12 22:52:04.720905 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Sep 12 22:52:04.721431 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Sep 12 22:52:04.767608 coreos-metadata[1969]: Sep 12 22:52:04.763 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Sep 12 22:52:04.767608 coreos-metadata[1969]: Sep 12 22:52:04.765 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Sep 12 22:52:04.767608 coreos-metadata[1969]: Sep 12 22:52:04.765 INFO Fetch successful Sep 12 22:52:04.767608 coreos-metadata[1969]: Sep 12 22:52:04.765 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Sep 12 22:52:04.767608 coreos-metadata[1969]: Sep 12 22:52:04.767 INFO Fetch successful Sep 12 22:52:04.767608 coreos-metadata[1969]: Sep 12 22:52:04.767 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Sep 12 22:52:04.767282 systemd[1]: motdgen.service: Deactivated successfully. Sep 12 22:52:04.768322 coreos-metadata[1969]: Sep 12 22:52:04.768 INFO Fetch successful Sep 12 22:52:04.768322 coreos-metadata[1969]: Sep 12 22:52:04.768 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.768 INFO Fetch successful Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.768 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.769 INFO Fetch failed with 404: resource not found Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.769 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.770 INFO Fetch successful Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.770 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.771 INFO Fetch successful Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.771 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.771 INFO Fetch successful Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.771 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.772 INFO Fetch successful Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.772 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Sep 12 22:52:04.775286 coreos-metadata[1969]: Sep 12 22:52:04.773 INFO Fetch successful Sep 12 22:52:04.769390 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 12 22:52:04.798529 extend-filesystems[1973]: Found /dev/nvme0n1p6 Sep 12 22:52:04.822124 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 12 22:52:04.822793 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 12 22:52:04.830541 extend-filesystems[1973]: Found /dev/nvme0n1p9 Sep 12 22:52:04.832373 jq[1985]: true Sep 12 22:52:04.850774 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Sep 12 22:52:04.853761 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 12 22:52:04.859458 update_engine[1984]: I20250912 22:52:04.856225 1984 main.cc:92] Flatcar Update Engine starting Sep 12 22:52:04.865427 ntpd[1976]: ntpd 4.2.8p18@1.4062-o Fri Sep 12 20:09:42 UTC 2025 (1): Starting Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: ntpd 4.2.8p18@1.4062-o Fri Sep 12 20:09:42 UTC 2025 (1): Starting Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: ---------------------------------------------------- Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: ntp-4 is maintained by Network Time Foundation, Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: corporation. Support and training for ntp-4 are Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: available at https://www.nwtime.org/support Sep 12 22:52:04.866923 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: ---------------------------------------------------- Sep 12 22:52:04.865506 ntpd[1976]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Sep 12 22:52:04.865518 ntpd[1976]: ---------------------------------------------------- Sep 12 22:52:04.865527 ntpd[1976]: ntp-4 is maintained by Network Time Foundation, Sep 12 22:52:04.865536 ntpd[1976]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Sep 12 22:52:04.865544 ntpd[1976]: corporation. Support and training for ntp-4 are Sep 12 22:52:04.865553 ntpd[1976]: available at https://www.nwtime.org/support Sep 12 22:52:04.865564 ntpd[1976]: ---------------------------------------------------- Sep 12 22:52:04.869944 extend-filesystems[1973]: Checking size of /dev/nvme0n1p9 Sep 12 22:52:04.878579 ntpd[1976]: proto: precision = 0.065 usec (-24) Sep 12 22:52:04.878897 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: proto: precision = 0.065 usec (-24) Sep 12 22:52:04.880355 tar[1994]: linux-amd64/LICENSE Sep 12 22:52:04.880355 tar[1994]: linux-amd64/helm Sep 12 22:52:04.881140 (ntainerd)[2017]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 12 22:52:04.927975 kernel: ntpd[1976]: segfault at 24 ip 0000555634accaeb sp 00007ffd40541940 error 4 in ntpd[68aeb,555634a6a000+80000] likely on CPU 1 (core 0, socket 0) Sep 12 22:52:04.928046 kernel: Code: 0f 1e fa 41 56 41 55 41 54 55 53 48 89 fb e8 8c eb f9 ff 44 8b 28 49 89 c4 e8 51 6b ff ff 48 89 c5 48 85 db 0f 84 a5 00 00 00 <0f> b7 0b 66 83 f9 02 0f 84 c0 00 00 00 66 83 f9 0a 74 32 66 85 c9 Sep 12 22:52:04.925059 systemd-coredump[2026]: Process 1976 (ntpd) of user 0 terminated abnormally with signal 11/SEGV, processing... Sep 12 22:52:04.904074 ntpd[1976]: basedate set to 2025-08-31 Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: basedate set to 2025-08-31 Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: gps base set to 2025-08-31 (week 2382) Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: Listen and drop on 0 v6wildcard [::]:123 Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: Listen normally on 2 lo 127.0.0.1:123 Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: Listen normally on 3 eth0 172.31.29.105:123 Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: Listen normally on 4 lo [::1]:123 Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: bind(21) AF_INET6 [fe80::437:25ff:fe44:ea41%2]:123 flags 0x811 failed: Cannot assign requested address Sep 12 22:52:04.928324 ntpd[1976]: 12 Sep 22:52:04 ntpd[1976]: unable to create socket on eth0 (5) for [fe80::437:25ff:fe44:ea41%2]:123 Sep 12 22:52:04.904104 ntpd[1976]: gps base set to 2025-08-31 (week 2382) Sep 12 22:52:04.904303 ntpd[1976]: Listen and drop on 0 v6wildcard [::]:123 Sep 12 22:52:04.904336 ntpd[1976]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Sep 12 22:52:04.904598 ntpd[1976]: Listen normally on 2 lo 127.0.0.1:123 Sep 12 22:52:04.904628 ntpd[1976]: Listen normally on 3 eth0 172.31.29.105:123 Sep 12 22:52:04.904656 ntpd[1976]: Listen normally on 4 lo [::1]:123 Sep 12 22:52:04.904685 ntpd[1976]: bind(21) AF_INET6 [fe80::437:25ff:fe44:ea41%2]:123 flags 0x811 failed: Cannot assign requested address Sep 12 22:52:04.904703 ntpd[1976]: unable to create socket on eth0 (5) for [fe80::437:25ff:fe44:ea41%2]:123 Sep 12 22:52:04.929472 systemd[1]: Created slice system-systemd\x2dcoredump.slice - Slice /system/systemd-coredump. Sep 12 22:52:04.936438 systemd[1]: Started systemd-coredump@0-2026-0.service - Process Core Dump (PID 2026/UID 0). Sep 12 22:52:04.945292 jq[2018]: true Sep 12 22:52:04.962830 systemd[1]: Finished setup-oem.service - Setup OEM. Sep 12 22:52:04.971292 dbus-daemon[1970]: [system] SELinux support is enabled Sep 12 22:52:04.971541 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 12 22:52:04.977745 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 12 22:52:04.980721 dbus-daemon[1970]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1897 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Sep 12 22:52:04.981624 extend-filesystems[1973]: Resized partition /dev/nvme0n1p9 Sep 12 22:52:04.977789 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 12 22:52:04.979526 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 12 22:52:04.979551 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 12 22:52:04.986982 dbus-daemon[1970]: [system] Successfully activated service 'org.freedesktop.systemd1' Sep 12 22:52:05.003021 extend-filesystems[2037]: resize2fs 1.47.3 (8-Jul-2025) Sep 12 22:52:05.003899 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Sep 12 22:52:05.011809 systemd[1]: Started update-engine.service - Update Engine. Sep 12 22:52:05.012616 update_engine[1984]: I20250912 22:52:05.012550 1984 update_check_scheduler.cc:74] Next update check in 7m41s Sep 12 22:52:05.020298 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Sep 12 22:52:05.061537 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 12 22:52:05.090854 systemd-logind[1983]: Watching system buttons on /dev/input/event2 (Power Button) Sep 12 22:52:05.090883 systemd-logind[1983]: Watching system buttons on /dev/input/event3 (Sleep Button) Sep 12 22:52:05.090909 systemd-logind[1983]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 12 22:52:05.091206 systemd-logind[1983]: New seat seat0. Sep 12 22:52:05.092756 systemd[1]: Started systemd-logind.service - User Login Management. Sep 12 22:52:05.184299 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Sep 12 22:52:05.207876 extend-filesystems[2037]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Sep 12 22:52:05.207876 extend-filesystems[2037]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 12 22:52:05.207876 extend-filesystems[2037]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Sep 12 22:52:05.241584 extend-filesystems[1973]: Resized filesystem in /dev/nvme0n1p9 Sep 12 22:52:05.211793 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 12 22:52:05.250986 bash[2059]: Updated "/home/core/.ssh/authorized_keys" Sep 12 22:52:05.212077 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 12 22:52:05.219556 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 12 22:52:05.233703 systemd[1]: Starting sshkeys.service... Sep 12 22:52:05.297484 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Sep 12 22:52:05.302464 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Sep 12 22:52:05.568568 systemd-coredump[2030]: Process 1976 (ntpd) of user 0 dumped core. Module libnss_usrfiles.so.2 without build-id. Module libgcc_s.so.1 without build-id. Module ld-linux-x86-64.so.2 without build-id. Module libc.so.6 without build-id. Module libcrypto.so.3 without build-id. Module libm.so.6 without build-id. Module libcap.so.2 without build-id. Module ntpd without build-id. Stack trace of thread 1976: #0 0x0000555634accaeb n/a (ntpd + 0x68aeb) #1 0x0000555634a75cdf n/a (ntpd + 0x11cdf) #2 0x0000555634a76575 n/a (ntpd + 0x12575) #3 0x0000555634a71d8a n/a (ntpd + 0xdd8a) #4 0x0000555634a735d3 n/a (ntpd + 0xf5d3) #5 0x0000555634a7bfd1 n/a (ntpd + 0x17fd1) #6 0x0000555634a6cc2d n/a (ntpd + 0x8c2d) #7 0x00007efdec76316c n/a (libc.so.6 + 0x2716c) #8 0x00007efdec763229 __libc_start_main (libc.so.6 + 0x27229) #9 0x0000555634a6cc55 n/a (ntpd + 0x8c55) ELF object binary architecture: AMD x86-64 Sep 12 22:52:05.580570 systemd[1]: ntpd.service: Main process exited, code=dumped, status=11/SEGV Sep 12 22:52:05.580772 systemd[1]: ntpd.service: Failed with result 'core-dump'. Sep 12 22:52:05.590905 systemd[1]: systemd-coredump@0-2026-0.service: Deactivated successfully. Sep 12 22:52:05.636503 coreos-metadata[2096]: Sep 12 22:52:05.636 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Sep 12 22:52:05.645330 coreos-metadata[2096]: Sep 12 22:52:05.640 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Sep 12 22:52:05.645330 coreos-metadata[2096]: Sep 12 22:52:05.643 INFO Fetch successful Sep 12 22:52:05.645330 coreos-metadata[2096]: Sep 12 22:52:05.643 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Sep 12 22:52:05.645330 coreos-metadata[2096]: Sep 12 22:52:05.644 INFO Fetch successful Sep 12 22:52:05.655391 unknown[2096]: wrote ssh authorized keys file for user: core Sep 12 22:52:05.673193 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Sep 12 22:52:05.683677 dbus-daemon[1970]: [system] Successfully activated service 'org.freedesktop.hostname1' Sep 12 22:52:05.691193 systemd[1]: ntpd.service: Scheduled restart job, restart counter is at 1. Sep 12 22:52:05.696573 systemd[1]: Started ntpd.service - Network Time Service. Sep 12 22:52:05.722662 dbus-daemon[1970]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=2038 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Sep 12 22:52:05.731814 systemd[1]: Starting polkit.service - Authorization Manager... Sep 12 22:52:05.733500 systemd-networkd[1897]: eth0: Gained IPv6LL Sep 12 22:52:05.742862 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 12 22:52:05.745203 systemd[1]: Reached target network-online.target - Network is Online. Sep 12 22:52:05.756059 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Sep 12 22:52:05.762128 update-ssh-keys[2164]: Updated "/home/core/.ssh/authorized_keys" Sep 12 22:52:05.765731 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:52:05.772781 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 12 22:52:05.775723 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Sep 12 22:52:05.782030 systemd[1]: Finished sshkeys.service. Sep 12 22:52:05.890476 containerd[2017]: time="2025-09-12T22:52:05Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Sep 12 22:52:05.901841 containerd[2017]: time="2025-09-12T22:52:05.901785874Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Sep 12 22:52:05.906753 ntpd[2166]: ntpd 4.2.8p18@1.4062-o Fri Sep 12 20:09:42 UTC 2025 (1): Starting Sep 12 22:52:05.907080 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: ntpd 4.2.8p18@1.4062-o Fri Sep 12 20:09:42 UTC 2025 (1): Starting Sep 12 22:52:05.913308 ntpd[2166]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: ---------------------------------------------------- Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: ntp-4 is maintained by Network Time Foundation, Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: corporation. Support and training for ntp-4 are Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: available at https://www.nwtime.org/support Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: ---------------------------------------------------- Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: proto: precision = 0.075 usec (-24) Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: basedate set to 2025-08-31 Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: gps base set to 2025-08-31 (week 2382) Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Listen and drop on 0 v6wildcard [::]:123 Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Listen normally on 2 lo 127.0.0.1:123 Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Listen normally on 3 eth0 172.31.29.105:123 Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Listen normally on 4 lo [::1]:123 Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Listen normally on 5 eth0 [fe80::437:25ff:fe44:ea41%2]:123 Sep 12 22:52:05.915596 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: Listening on routing socket on fd #22 for interface updates Sep 12 22:52:05.913347 ntpd[2166]: ---------------------------------------------------- Sep 12 22:52:05.913358 ntpd[2166]: ntp-4 is maintained by Network Time Foundation, Sep 12 22:52:05.913368 ntpd[2166]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Sep 12 22:52:05.913377 ntpd[2166]: corporation. Support and training for ntp-4 are Sep 12 22:52:05.913386 ntpd[2166]: available at https://www.nwtime.org/support Sep 12 22:52:05.913395 ntpd[2166]: ---------------------------------------------------- Sep 12 22:52:05.914200 ntpd[2166]: proto: precision = 0.075 usec (-24) Sep 12 22:52:05.914474 ntpd[2166]: basedate set to 2025-08-31 Sep 12 22:52:05.914487 ntpd[2166]: gps base set to 2025-08-31 (week 2382) Sep 12 22:52:05.914579 ntpd[2166]: Listen and drop on 0 v6wildcard [::]:123 Sep 12 22:52:05.914607 ntpd[2166]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Sep 12 22:52:05.914785 ntpd[2166]: Listen normally on 2 lo 127.0.0.1:123 Sep 12 22:52:05.914812 ntpd[2166]: Listen normally on 3 eth0 172.31.29.105:123 Sep 12 22:52:05.914839 ntpd[2166]: Listen normally on 4 lo [::1]:123 Sep 12 22:52:05.914866 ntpd[2166]: Listen normally on 5 eth0 [fe80::437:25ff:fe44:ea41%2]:123 Sep 12 22:52:05.914890 ntpd[2166]: Listening on routing socket on fd #22 for interface updates Sep 12 22:52:05.928884 ntpd[2166]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 12 22:52:05.928934 ntpd[2166]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 12 22:52:05.929107 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 12 22:52:05.929107 ntpd[2166]: 12 Sep 22:52:05 ntpd[2166]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Sep 12 22:52:05.929826 locksmithd[2040]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 12 22:52:05.998987 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 12 22:52:06.008648 amazon-ssm-agent[2174]: Initializing new seelog logger Sep 12 22:52:06.019392 amazon-ssm-agent[2174]: New Seelog Logger Creation Complete Sep 12 22:52:06.019500 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.019500 amazon-ssm-agent[2174]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.020056 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 processing appconfig overrides Sep 12 22:52:06.024530 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.024530 amazon-ssm-agent[2174]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.024530 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 processing appconfig overrides Sep 12 22:52:06.024894 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.025996 amazon-ssm-agent[2174]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.025996 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 processing appconfig overrides Sep 12 22:52:06.025996 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.0204 INFO Proxy environment variables: Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037418189Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="12.377µs" Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037461770Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037486634Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037692910Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037715743Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037748215Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037826177Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Sep 12 22:52:06.037958 containerd[2017]: time="2025-09-12T22:52:06.037841952Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 12 22:52:06.038333 containerd[2017]: time="2025-09-12T22:52:06.038118650Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Sep 12 22:52:06.038333 containerd[2017]: time="2025-09-12T22:52:06.038142225Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 12 22:52:06.038333 containerd[2017]: time="2025-09-12T22:52:06.038158933Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Sep 12 22:52:06.038333 containerd[2017]: time="2025-09-12T22:52:06.038171711Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Sep 12 22:52:06.041700 containerd[2017]: time="2025-09-12T22:52:06.040747363Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Sep 12 22:52:06.041700 containerd[2017]: time="2025-09-12T22:52:06.041064302Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 12 22:52:06.041700 containerd[2017]: time="2025-09-12T22:52:06.041107451Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Sep 12 22:52:06.041700 containerd[2017]: time="2025-09-12T22:52:06.041121773Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Sep 12 22:52:06.042734 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.042734 amazon-ssm-agent[2174]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:06.042734 amazon-ssm-agent[2174]: 2025/09/12 22:52:06 processing appconfig overrides Sep 12 22:52:06.043633 containerd[2017]: time="2025-09-12T22:52:06.043553316Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Sep 12 22:52:06.046676 containerd[2017]: time="2025-09-12T22:52:06.045756120Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Sep 12 22:52:06.046676 containerd[2017]: time="2025-09-12T22:52:06.045890733Z" level=info msg="metadata content store policy set" policy=shared Sep 12 22:52:06.079122 containerd[2017]: time="2025-09-12T22:52:06.079021338Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079301001Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079408492Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079431692Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079455813Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079475497Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079491786Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079509538Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079527743Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079552454Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079569663Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079587211Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079764873Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079792555Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Sep 12 22:52:06.081818 containerd[2017]: time="2025-09-12T22:52:06.079819252Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079836697Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079853036Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079870041Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079887787Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079902643Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079919643Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079940567Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.079954959Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.080043364Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.080060609Z" level=info msg="Start snapshots syncer" Sep 12 22:52:06.082968 containerd[2017]: time="2025-09-12T22:52:06.081723779Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Sep 12 22:52:06.083382 containerd[2017]: time="2025-09-12T22:52:06.082635576Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Sep 12 22:52:06.083382 containerd[2017]: time="2025-09-12T22:52:06.082709963Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099095713Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099340755Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099392247Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099410635Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099430640Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099452574Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099469449Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099486715Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099526169Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099543979Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099560473Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099634087Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 12 22:52:06.099943 containerd[2017]: time="2025-09-12T22:52:06.099656450Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100373448Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100407357Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100421981Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100439138Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100456785Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100480354Z" level=info msg="runtime interface created" Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100488408Z" level=info msg="created NRI interface" Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100501317Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100524376Z" level=info msg="Connect containerd service" Sep 12 22:52:06.101055 containerd[2017]: time="2025-09-12T22:52:06.100575437Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 12 22:52:06.111400 containerd[2017]: time="2025-09-12T22:52:06.110342634Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 12 22:52:06.126857 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.0232 INFO https_proxy: Sep 12 22:52:06.158797 polkitd[2168]: Started polkitd version 126 Sep 12 22:52:06.167775 polkitd[2168]: Loading rules from directory /etc/polkit-1/rules.d Sep 12 22:52:06.170680 polkitd[2168]: Loading rules from directory /run/polkit-1/rules.d Sep 12 22:52:06.170753 polkitd[2168]: Error opening rules directory: Error opening directory “/run/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Sep 12 22:52:06.172119 polkitd[2168]: Loading rules from directory /usr/local/share/polkit-1/rules.d Sep 12 22:52:06.172173 polkitd[2168]: Error opening rules directory: Error opening directory “/usr/local/share/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Sep 12 22:52:06.172243 polkitd[2168]: Loading rules from directory /usr/share/polkit-1/rules.d Sep 12 22:52:06.174825 polkitd[2168]: Finished loading, compiling and executing 2 rules Sep 12 22:52:06.175169 systemd[1]: Started polkit.service - Authorization Manager. Sep 12 22:52:06.177733 dbus-daemon[1970]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Sep 12 22:52:06.189325 polkitd[2168]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Sep 12 22:52:06.246655 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.0232 INFO http_proxy: Sep 12 22:52:06.284778 systemd-hostnamed[2038]: Hostname set to (transient) Sep 12 22:52:06.285781 systemd-resolved[1898]: System hostname changed to 'ip-172-31-29-105'. Sep 12 22:52:06.348045 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.0232 INFO no_proxy: Sep 12 22:52:06.408406 sshd_keygen[2024]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 12 22:52:06.408661 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 12 22:52:06.443956 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 12 22:52:06.446097 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.0234 INFO Checking if agent identity type OnPrem can be assumed Sep 12 22:52:06.452937 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 12 22:52:06.459388 systemd[1]: Started sshd@0-172.31.29.105:22-139.178.89.65:39598.service - OpenSSH per-connection server daemon (139.178.89.65:39598). Sep 12 22:52:06.470113 containerd[2017]: time="2025-09-12T22:52:06.470053527Z" level=info msg="Start subscribing containerd event" Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471326001Z" level=info msg="Start recovering state" Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471488922Z" level=info msg="Start event monitor" Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471507493Z" level=info msg="Start cni network conf syncer for default" Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471517147Z" level=info msg="Start streaming server" Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471529977Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471539809Z" level=info msg="runtime interface starting up..." Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471547969Z" level=info msg="starting plugins..." Sep 12 22:52:06.471661 containerd[2017]: time="2025-09-12T22:52:06.471561363Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Sep 12 22:52:06.476938 containerd[2017]: time="2025-09-12T22:52:06.474743225Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 12 22:52:06.476938 containerd[2017]: time="2025-09-12T22:52:06.474819342Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 12 22:52:06.479357 systemd[1]: Started containerd.service - containerd container runtime. Sep 12 22:52:06.483571 containerd[2017]: time="2025-09-12T22:52:06.481846284Z" level=info msg="containerd successfully booted in 0.592199s" Sep 12 22:52:06.517758 systemd[1]: issuegen.service: Deactivated successfully. Sep 12 22:52:06.518053 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 12 22:52:06.520921 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 12 22:52:06.544645 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.0247 INFO Checking if agent identity type EC2 can be assumed Sep 12 22:52:06.576874 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 12 22:52:06.582041 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 12 22:52:06.588492 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Sep 12 22:52:06.590613 systemd[1]: Reached target getty.target - Login Prompts. Sep 12 22:52:06.644046 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.2750 INFO Agent will take identity from EC2 Sep 12 22:52:06.699386 tar[1994]: linux-amd64/README.md Sep 12 22:52:06.731815 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 12 22:52:06.737300 sshd[2234]: Accepted publickey for core from 139.178.89.65 port 39598 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:06.738390 sshd-session[2234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:06.743390 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.2837 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.3.0.0 Sep 12 22:52:06.749624 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 12 22:52:06.753859 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 12 22:52:06.774983 systemd-logind[1983]: New session 1 of user core. Sep 12 22:52:06.788856 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 12 22:52:06.799667 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 12 22:52:06.815340 (systemd)[2250]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 12 22:52:06.819629 systemd-logind[1983]: New session c1 of user core. Sep 12 22:52:06.842953 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.2837 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Sep 12 22:52:06.944281 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.2837 INFO [amazon-ssm-agent] Starting Core Agent Sep 12 22:52:07.044064 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.2838 INFO [amazon-ssm-agent] Registrar detected. Attempting registration Sep 12 22:52:07.081128 systemd[2250]: Queued start job for default target default.target. Sep 12 22:52:07.087374 systemd[2250]: Created slice app.slice - User Application Slice. Sep 12 22:52:07.087521 systemd[2250]: Reached target paths.target - Paths. Sep 12 22:52:07.087579 systemd[2250]: Reached target timers.target - Timers. Sep 12 22:52:07.090390 systemd[2250]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 12 22:52:07.111008 systemd[2250]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 12 22:52:07.111165 systemd[2250]: Reached target sockets.target - Sockets. Sep 12 22:52:07.111319 systemd[2250]: Reached target basic.target - Basic System. Sep 12 22:52:07.111385 systemd[2250]: Reached target default.target - Main User Target. Sep 12 22:52:07.111428 systemd[2250]: Startup finished in 278ms. Sep 12 22:52:07.111546 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 12 22:52:07.124552 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 12 22:52:07.144793 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.2838 INFO [Registrar] Starting registrar module Sep 12 22:52:07.246380 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.3015 INFO [EC2Identity] Checking disk for registration info Sep 12 22:52:07.287750 systemd[1]: Started sshd@1-172.31.29.105:22-139.178.89.65:39614.service - OpenSSH per-connection server daemon (139.178.89.65:39614). Sep 12 22:52:07.347351 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.3016 INFO [EC2Identity] No registration info found for ec2 instance, attempting registration Sep 12 22:52:07.447527 amazon-ssm-agent[2174]: 2025-09-12 22:52:06.3016 INFO [EC2Identity] Generating registration keypair Sep 12 22:52:07.488872 sshd[2261]: Accepted publickey for core from 139.178.89.65 port 39614 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:07.490642 sshd-session[2261]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:07.498200 systemd-logind[1983]: New session 2 of user core. Sep 12 22:52:07.507417 amazon-ssm-agent[2174]: 2025/09/12 22:52:07 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:07.507417 amazon-ssm-agent[2174]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Sep 12 22:52:07.507555 amazon-ssm-agent[2174]: 2025/09/12 22:52:07 processing appconfig overrides Sep 12 22:52:07.508684 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 12 22:52:07.539959 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.4618 INFO [EC2Identity] Checking write access before registering Sep 12 22:52:07.539959 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.4621 INFO [EC2Identity] Registering EC2 instance with Systems Manager Sep 12 22:52:07.539959 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.5071 INFO [EC2Identity] EC2 registration was successful. Sep 12 22:52:07.539959 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.5072 INFO [amazon-ssm-agent] Registration attempted. Resuming core agent startup. Sep 12 22:52:07.539959 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.5073 INFO [CredentialRefresher] credentialRefresher has started Sep 12 22:52:07.539959 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.5073 INFO [CredentialRefresher] Starting credentials refresher loop Sep 12 22:52:07.539959 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.5396 INFO EC2RoleProvider Successfully connected with instance profile role credentials Sep 12 22:52:07.540405 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.5398 INFO [CredentialRefresher] Credentials ready Sep 12 22:52:07.549015 amazon-ssm-agent[2174]: 2025-09-12 22:52:07.5399 INFO [CredentialRefresher] Next credential rotation will be in 29.999994978166665 minutes Sep 12 22:52:07.634065 sshd[2264]: Connection closed by 139.178.89.65 port 39614 Sep 12 22:52:07.635488 sshd-session[2261]: pam_unix(sshd:session): session closed for user core Sep 12 22:52:07.639973 systemd[1]: sshd@1-172.31.29.105:22-139.178.89.65:39614.service: Deactivated successfully. Sep 12 22:52:07.642374 systemd[1]: session-2.scope: Deactivated successfully. Sep 12 22:52:07.643862 systemd-logind[1983]: Session 2 logged out. Waiting for processes to exit. Sep 12 22:52:07.645958 systemd-logind[1983]: Removed session 2. Sep 12 22:52:07.670108 systemd[1]: Started sshd@2-172.31.29.105:22-139.178.89.65:39622.service - OpenSSH per-connection server daemon (139.178.89.65:39622). Sep 12 22:52:07.849025 sshd[2270]: Accepted publickey for core from 139.178.89.65 port 39622 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:07.850457 sshd-session[2270]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:07.856901 systemd-logind[1983]: New session 3 of user core. Sep 12 22:52:07.862548 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 12 22:52:07.983347 sshd[2273]: Connection closed by 139.178.89.65 port 39622 Sep 12 22:52:07.984143 sshd-session[2270]: pam_unix(sshd:session): session closed for user core Sep 12 22:52:07.988594 systemd[1]: sshd@2-172.31.29.105:22-139.178.89.65:39622.service: Deactivated successfully. Sep 12 22:52:07.991703 systemd[1]: session-3.scope: Deactivated successfully. Sep 12 22:52:07.994106 systemd-logind[1983]: Session 3 logged out. Waiting for processes to exit. Sep 12 22:52:07.995507 systemd-logind[1983]: Removed session 3. Sep 12 22:52:08.556186 amazon-ssm-agent[2174]: 2025-09-12 22:52:08.5535 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Sep 12 22:52:08.655201 amazon-ssm-agent[2174]: 2025-09-12 22:52:08.5568 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2280) started Sep 12 22:52:08.756365 amazon-ssm-agent[2174]: 2025-09-12 22:52:08.5569 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Sep 12 22:52:09.405680 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:09.406919 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 12 22:52:09.409083 systemd[1]: Startup finished in 2.734s (kernel) + 7.318s (initrd) + 9.490s (userspace) = 19.543s. Sep 12 22:52:09.416678 (kubelet)[2297]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 22:52:11.297994 kubelet[2297]: E0912 22:52:11.297906 2297 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 22:52:11.300878 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 22:52:11.301023 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 22:52:11.301563 systemd[1]: kubelet.service: Consumed 1.077s CPU time, 266.8M memory peak. Sep 12 22:52:14.310482 systemd-resolved[1898]: Clock change detected. Flushing caches. Sep 12 22:52:19.418610 systemd[1]: Started sshd@3-172.31.29.105:22-139.178.89.65:57688.service - OpenSSH per-connection server daemon (139.178.89.65:57688). Sep 12 22:52:19.584478 sshd[2310]: Accepted publickey for core from 139.178.89.65 port 57688 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:19.585807 sshd-session[2310]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:19.591479 systemd-logind[1983]: New session 4 of user core. Sep 12 22:52:19.593644 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 12 22:52:19.718826 sshd[2313]: Connection closed by 139.178.89.65 port 57688 Sep 12 22:52:19.719372 sshd-session[2310]: pam_unix(sshd:session): session closed for user core Sep 12 22:52:19.722665 systemd[1]: sshd@3-172.31.29.105:22-139.178.89.65:57688.service: Deactivated successfully. Sep 12 22:52:19.724852 systemd[1]: session-4.scope: Deactivated successfully. Sep 12 22:52:19.729109 systemd-logind[1983]: Session 4 logged out. Waiting for processes to exit. Sep 12 22:52:19.733254 systemd-logind[1983]: Removed session 4. Sep 12 22:52:19.756246 systemd[1]: Started sshd@4-172.31.29.105:22-139.178.89.65:57702.service - OpenSSH per-connection server daemon (139.178.89.65:57702). Sep 12 22:52:19.939029 sshd[2319]: Accepted publickey for core from 139.178.89.65 port 57702 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:19.940576 sshd-session[2319]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:19.946746 systemd-logind[1983]: New session 5 of user core. Sep 12 22:52:19.952727 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 12 22:52:20.070077 sshd[2322]: Connection closed by 139.178.89.65 port 57702 Sep 12 22:52:20.070668 sshd-session[2319]: pam_unix(sshd:session): session closed for user core Sep 12 22:52:20.074255 systemd[1]: sshd@4-172.31.29.105:22-139.178.89.65:57702.service: Deactivated successfully. Sep 12 22:52:20.076768 systemd[1]: session-5.scope: Deactivated successfully. Sep 12 22:52:20.078432 systemd-logind[1983]: Session 5 logged out. Waiting for processes to exit. Sep 12 22:52:20.079942 systemd-logind[1983]: Removed session 5. Sep 12 22:52:20.101177 systemd[1]: Started sshd@5-172.31.29.105:22-139.178.89.65:39552.service - OpenSSH per-connection server daemon (139.178.89.65:39552). Sep 12 22:52:20.275007 sshd[2328]: Accepted publickey for core from 139.178.89.65 port 39552 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:20.277927 sshd-session[2328]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:20.284223 systemd-logind[1983]: New session 6 of user core. Sep 12 22:52:20.290762 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 12 22:52:20.408008 sshd[2331]: Connection closed by 139.178.89.65 port 39552 Sep 12 22:52:20.408731 sshd-session[2328]: pam_unix(sshd:session): session closed for user core Sep 12 22:52:20.412003 systemd[1]: sshd@5-172.31.29.105:22-139.178.89.65:39552.service: Deactivated successfully. Sep 12 22:52:20.413697 systemd[1]: session-6.scope: Deactivated successfully. Sep 12 22:52:20.414821 systemd-logind[1983]: Session 6 logged out. Waiting for processes to exit. Sep 12 22:52:20.416243 systemd-logind[1983]: Removed session 6. Sep 12 22:52:20.444703 systemd[1]: Started sshd@6-172.31.29.105:22-139.178.89.65:39566.service - OpenSSH per-connection server daemon (139.178.89.65:39566). Sep 12 22:52:20.628714 sshd[2337]: Accepted publickey for core from 139.178.89.65 port 39566 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:20.630062 sshd-session[2337]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:20.635760 systemd-logind[1983]: New session 7 of user core. Sep 12 22:52:20.644706 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 12 22:52:20.751271 sudo[2341]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 12 22:52:20.751616 sudo[2341]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:52:20.768691 sudo[2341]: pam_unix(sudo:session): session closed for user root Sep 12 22:52:20.790384 sshd[2340]: Connection closed by 139.178.89.65 port 39566 Sep 12 22:52:20.791109 sshd-session[2337]: pam_unix(sshd:session): session closed for user core Sep 12 22:52:20.795226 systemd[1]: sshd@6-172.31.29.105:22-139.178.89.65:39566.service: Deactivated successfully. Sep 12 22:52:20.796946 systemd[1]: session-7.scope: Deactivated successfully. Sep 12 22:52:20.797661 systemd-logind[1983]: Session 7 logged out. Waiting for processes to exit. Sep 12 22:52:20.798881 systemd-logind[1983]: Removed session 7. Sep 12 22:52:20.825301 systemd[1]: Started sshd@7-172.31.29.105:22-139.178.89.65:39582.service - OpenSSH per-connection server daemon (139.178.89.65:39582). Sep 12 22:52:20.996576 sshd[2347]: Accepted publickey for core from 139.178.89.65 port 39582 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:20.998321 sshd-session[2347]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:21.004918 systemd-logind[1983]: New session 8 of user core. Sep 12 22:52:21.010809 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 12 22:52:21.109081 sudo[2352]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 12 22:52:21.109350 sudo[2352]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:52:21.115831 sudo[2352]: pam_unix(sudo:session): session closed for user root Sep 12 22:52:21.121958 sudo[2351]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Sep 12 22:52:21.122227 sudo[2351]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:52:21.133138 systemd[1]: Starting audit-rules.service - Load Audit Rules... Sep 12 22:52:21.174032 augenrules[2374]: No rules Sep 12 22:52:21.175208 systemd[1]: audit-rules.service: Deactivated successfully. Sep 12 22:52:21.175572 systemd[1]: Finished audit-rules.service - Load Audit Rules. Sep 12 22:52:21.177727 sudo[2351]: pam_unix(sudo:session): session closed for user root Sep 12 22:52:21.200626 sshd[2350]: Connection closed by 139.178.89.65 port 39582 Sep 12 22:52:21.201181 sshd-session[2347]: pam_unix(sshd:session): session closed for user core Sep 12 22:52:21.206023 systemd[1]: sshd@7-172.31.29.105:22-139.178.89.65:39582.service: Deactivated successfully. Sep 12 22:52:21.207988 systemd[1]: session-8.scope: Deactivated successfully. Sep 12 22:52:21.209307 systemd-logind[1983]: Session 8 logged out. Waiting for processes to exit. Sep 12 22:52:21.211132 systemd-logind[1983]: Removed session 8. Sep 12 22:52:21.237446 systemd[1]: Started sshd@8-172.31.29.105:22-139.178.89.65:39590.service - OpenSSH per-connection server daemon (139.178.89.65:39590). Sep 12 22:52:21.409106 sshd[2383]: Accepted publickey for core from 139.178.89.65 port 39590 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:52:21.410366 sshd-session[2383]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:52:21.415842 systemd-logind[1983]: New session 9 of user core. Sep 12 22:52:21.422685 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 12 22:52:21.520776 sudo[2387]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 12 22:52:21.521161 sudo[2387]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Sep 12 22:52:21.924911 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 12 22:52:21.941943 (dockerd)[2405]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 12 22:52:22.244398 dockerd[2405]: time="2025-09-12T22:52:22.243994439Z" level=info msg="Starting up" Sep 12 22:52:22.248881 dockerd[2405]: time="2025-09-12T22:52:22.248759662Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Sep 12 22:52:22.262498 dockerd[2405]: time="2025-09-12T22:52:22.262367239Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Sep 12 22:52:22.335478 dockerd[2405]: time="2025-09-12T22:52:22.335411603Z" level=info msg="Loading containers: start." Sep 12 22:52:22.348534 kernel: Initializing XFRM netlink socket Sep 12 22:52:22.655014 (udev-worker)[2426]: Network interface NamePolicy= disabled on kernel command line. Sep 12 22:52:22.707391 systemd-networkd[1897]: docker0: Link UP Sep 12 22:52:22.718405 dockerd[2405]: time="2025-09-12T22:52:22.718340566Z" level=info msg="Loading containers: done." Sep 12 22:52:22.721356 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 12 22:52:22.723859 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:52:22.738589 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4587218-merged.mount: Deactivated successfully. Sep 12 22:52:22.752933 dockerd[2405]: time="2025-09-12T22:52:22.752030381Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 12 22:52:22.752933 dockerd[2405]: time="2025-09-12T22:52:22.752118810Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Sep 12 22:52:22.752933 dockerd[2405]: time="2025-09-12T22:52:22.752209970Z" level=info msg="Initializing buildkit" Sep 12 22:52:22.813941 dockerd[2405]: time="2025-09-12T22:52:22.813901832Z" level=info msg="Completed buildkit initialization" Sep 12 22:52:22.824489 dockerd[2405]: time="2025-09-12T22:52:22.822330125Z" level=info msg="Daemon has completed initialization" Sep 12 22:52:22.824489 dockerd[2405]: time="2025-09-12T22:52:22.822527479Z" level=info msg="API listen on /run/docker.sock" Sep 12 22:52:22.822620 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 12 22:52:23.035344 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:23.051950 (kubelet)[2620]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 22:52:23.106401 kubelet[2620]: E0912 22:52:23.106324 2620 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 22:52:23.110863 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 22:52:23.111050 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 22:52:23.111560 systemd[1]: kubelet.service: Consumed 194ms CPU time, 110.5M memory peak. Sep 12 22:52:24.765133 containerd[2017]: time="2025-09-12T22:52:24.765089602Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\"" Sep 12 22:52:25.367274 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2154227702.mount: Deactivated successfully. Sep 12 22:52:27.176814 containerd[2017]: time="2025-09-12T22:52:27.176749663Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:27.177881 containerd[2017]: time="2025-09-12T22:52:27.177721940Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.5: active requests=0, bytes read=30114893" Sep 12 22:52:27.178898 containerd[2017]: time="2025-09-12T22:52:27.178860279Z" level=info msg="ImageCreate event name:\"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:27.181543 containerd[2017]: time="2025-09-12T22:52:27.181488930Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:27.183030 containerd[2017]: time="2025-09-12T22:52:27.182333211Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.5\" with image id \"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:1b9c6c00bc1fe86860e72efb8e4148f9e436a132eba4ca636ca4f48d61d6dfb4\", size \"30111492\" in 2.417202079s" Sep 12 22:52:27.183030 containerd[2017]: time="2025-09-12T22:52:27.182371455Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.5\" returns image reference \"sha256:b7335a56022aba291f5df653c01b7ab98d64fb5cab221378617f4a1236e06a62\"" Sep 12 22:52:27.183334 containerd[2017]: time="2025-09-12T22:52:27.183299404Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\"" Sep 12 22:52:29.363987 containerd[2017]: time="2025-09-12T22:52:29.363929288Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:29.366184 containerd[2017]: time="2025-09-12T22:52:29.366128775Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.5: active requests=0, bytes read=26020844" Sep 12 22:52:29.368860 containerd[2017]: time="2025-09-12T22:52:29.368799383Z" level=info msg="ImageCreate event name:\"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:29.372764 containerd[2017]: time="2025-09-12T22:52:29.372695072Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:29.373788 containerd[2017]: time="2025-09-12T22:52:29.373669915Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.5\" with image id \"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:1082a6ab67fb46397314dd36b36cb197ba4a4c5365033e9ad22bc7edaaaabd5c\", size \"27681301\" in 2.190337356s" Sep 12 22:52:29.373788 containerd[2017]: time="2025-09-12T22:52:29.373704928Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.5\" returns image reference \"sha256:8bb43160a0df4d7d34c89d9edbc48735bc2f830771e4b501937338221be0f668\"" Sep 12 22:52:29.374579 containerd[2017]: time="2025-09-12T22:52:29.374520225Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\"" Sep 12 22:52:30.792404 containerd[2017]: time="2025-09-12T22:52:30.792351493Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:30.794513 containerd[2017]: time="2025-09-12T22:52:30.794475895Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.5: active requests=0, bytes read=20155568" Sep 12 22:52:30.797133 containerd[2017]: time="2025-09-12T22:52:30.797091182Z" level=info msg="ImageCreate event name:\"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:30.801291 containerd[2017]: time="2025-09-12T22:52:30.800953383Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:30.801814 containerd[2017]: time="2025-09-12T22:52:30.801786022Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.5\" with image id \"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:3e7b57c9d9f06b77f0064e5be7f3df61e0151101160acd5fdecce911df28a189\", size \"21816043\" in 1.427232414s" Sep 12 22:52:30.801871 containerd[2017]: time="2025-09-12T22:52:30.801819218Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.5\" returns image reference \"sha256:33b680aadf474b7e5e73957fc00c6af86dd0484c699c8461ba33ee656d1823bf\"" Sep 12 22:52:30.802314 containerd[2017]: time="2025-09-12T22:52:30.802293967Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\"" Sep 12 22:52:32.019022 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3202602165.mount: Deactivated successfully. Sep 12 22:52:32.685180 containerd[2017]: time="2025-09-12T22:52:32.685121245Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:32.687123 containerd[2017]: time="2025-09-12T22:52:32.687074470Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.5: active requests=0, bytes read=31929469" Sep 12 22:52:32.689559 containerd[2017]: time="2025-09-12T22:52:32.689496228Z" level=info msg="ImageCreate event name:\"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:32.692791 containerd[2017]: time="2025-09-12T22:52:32.692727270Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:32.693477 containerd[2017]: time="2025-09-12T22:52:32.693314193Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.5\" with image id \"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\", repo tag \"registry.k8s.io/kube-proxy:v1.33.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:71445ec84ad98bd52a7784865a9d31b1b50b56092d3f7699edc39eefd71befe1\", size \"31928488\" in 1.890993328s" Sep 12 22:52:32.693477 containerd[2017]: time="2025-09-12T22:52:32.693353055Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.5\" returns image reference \"sha256:2844ee7bb56c2c194e1f4adafb9e7b60b9ed16aa4d07ab8ad1f019362e2efab3\"" Sep 12 22:52:32.694333 containerd[2017]: time="2025-09-12T22:52:32.694141475Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Sep 12 22:52:33.208392 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 12 22:52:33.209881 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:52:33.243060 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount127710799.mount: Deactivated successfully. Sep 12 22:52:33.495614 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:33.507301 (kubelet)[2727]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 12 22:52:33.561445 kubelet[2727]: E0912 22:52:33.561355 2727 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 12 22:52:33.567447 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 12 22:52:33.567606 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 12 22:52:33.567910 systemd[1]: kubelet.service: Consumed 178ms CPU time, 110.7M memory peak. Sep 12 22:52:34.680720 containerd[2017]: time="2025-09-12T22:52:34.680657308Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:34.683423 containerd[2017]: time="2025-09-12T22:52:34.683138429Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Sep 12 22:52:34.685692 containerd[2017]: time="2025-09-12T22:52:34.685649927Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:34.689603 containerd[2017]: time="2025-09-12T22:52:34.689564237Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:34.691059 containerd[2017]: time="2025-09-12T22:52:34.690641364Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.996473861s" Sep 12 22:52:34.691059 containerd[2017]: time="2025-09-12T22:52:34.690677927Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Sep 12 22:52:34.691217 containerd[2017]: time="2025-09-12T22:52:34.691085048Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 12 22:52:35.166386 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3220213632.mount: Deactivated successfully. Sep 12 22:52:35.178892 containerd[2017]: time="2025-09-12T22:52:35.178802615Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 22:52:35.180848 containerd[2017]: time="2025-09-12T22:52:35.180802579Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Sep 12 22:52:35.183363 containerd[2017]: time="2025-09-12T22:52:35.183289790Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 22:52:35.186387 containerd[2017]: time="2025-09-12T22:52:35.186324078Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 12 22:52:35.186926 containerd[2017]: time="2025-09-12T22:52:35.186897545Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 495.790151ms" Sep 12 22:52:35.186926 containerd[2017]: time="2025-09-12T22:52:35.186927190Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 12 22:52:35.187921 containerd[2017]: time="2025-09-12T22:52:35.187889256Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Sep 12 22:52:35.694982 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3791572980.mount: Deactivated successfully. Sep 12 22:52:37.702679 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Sep 12 22:52:37.875405 containerd[2017]: time="2025-09-12T22:52:37.875307119Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:37.877677 containerd[2017]: time="2025-09-12T22:52:37.877433123Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58378433" Sep 12 22:52:37.879950 containerd[2017]: time="2025-09-12T22:52:37.879910438Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:37.884122 containerd[2017]: time="2025-09-12T22:52:37.884025415Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:37.885521 containerd[2017]: time="2025-09-12T22:52:37.885292483Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 2.69736886s" Sep 12 22:52:37.885521 containerd[2017]: time="2025-09-12T22:52:37.885332045Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Sep 12 22:52:40.835358 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:40.835644 systemd[1]: kubelet.service: Consumed 178ms CPU time, 110.7M memory peak. Sep 12 22:52:40.838896 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:52:40.880542 systemd[1]: Reload requested from client PID 2865 ('systemctl') (unit session-9.scope)... Sep 12 22:52:40.880571 systemd[1]: Reloading... Sep 12 22:52:41.020503 zram_generator::config[2908]: No configuration found. Sep 12 22:52:41.321475 systemd[1]: Reloading finished in 440 ms. Sep 12 22:52:41.380155 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 12 22:52:41.380284 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 12 22:52:41.380805 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:41.380869 systemd[1]: kubelet.service: Consumed 146ms CPU time, 98M memory peak. Sep 12 22:52:41.383282 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:52:41.612203 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:41.624901 (kubelet)[2972]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 22:52:41.675031 kubelet[2972]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:52:41.675031 kubelet[2972]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 12 22:52:41.675031 kubelet[2972]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:52:41.676928 kubelet[2972]: I0912 22:52:41.676872 2972 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 22:52:42.489483 kubelet[2972]: I0912 22:52:42.488826 2972 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 12 22:52:42.489483 kubelet[2972]: I0912 22:52:42.488864 2972 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 22:52:42.489483 kubelet[2972]: I0912 22:52:42.489367 2972 server.go:956] "Client rotation is on, will bootstrap in background" Sep 12 22:52:42.536107 kubelet[2972]: I0912 22:52:42.535982 2972 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 22:52:42.537302 kubelet[2972]: E0912 22:52:42.537222 2972 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.31.29.105:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Sep 12 22:52:42.560511 kubelet[2972]: I0912 22:52:42.560446 2972 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 12 22:52:42.569026 kubelet[2972]: I0912 22:52:42.568975 2972 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 22:52:42.575279 kubelet[2972]: I0912 22:52:42.575191 2972 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 22:52:42.578970 kubelet[2972]: I0912 22:52:42.575262 2972 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-29-105","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 22:52:42.578970 kubelet[2972]: I0912 22:52:42.578960 2972 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 22:52:42.578970 kubelet[2972]: I0912 22:52:42.578975 2972 container_manager_linux.go:303] "Creating device plugin manager" Sep 12 22:52:42.579220 kubelet[2972]: I0912 22:52:42.579116 2972 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:52:42.586537 kubelet[2972]: I0912 22:52:42.586482 2972 kubelet.go:480] "Attempting to sync node with API server" Sep 12 22:52:42.586727 kubelet[2972]: I0912 22:52:42.586553 2972 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 22:52:42.589537 kubelet[2972]: I0912 22:52:42.589508 2972 kubelet.go:386] "Adding apiserver pod source" Sep 12 22:52:42.591675 kubelet[2972]: I0912 22:52:42.591483 2972 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 22:52:42.598497 kubelet[2972]: E0912 22:52:42.589509 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.29.105:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-29-105&limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 12 22:52:42.600695 kubelet[2972]: E0912 22:52:42.600492 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.29.105:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 12 22:52:42.600967 kubelet[2972]: I0912 22:52:42.600943 2972 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 12 22:52:42.602104 kubelet[2972]: I0912 22:52:42.601439 2972 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 12 22:52:42.602435 kubelet[2972]: W0912 22:52:42.602389 2972 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 12 22:52:42.605880 kubelet[2972]: I0912 22:52:42.605828 2972 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 12 22:52:42.605880 kubelet[2972]: I0912 22:52:42.605887 2972 server.go:1289] "Started kubelet" Sep 12 22:52:42.609212 kubelet[2972]: I0912 22:52:42.608804 2972 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 22:52:42.615672 kubelet[2972]: I0912 22:52:42.615589 2972 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 22:52:42.616206 kubelet[2972]: I0912 22:52:42.616183 2972 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 22:52:42.618829 kubelet[2972]: I0912 22:52:42.618798 2972 server.go:317] "Adding debug handlers to kubelet server" Sep 12 22:52:42.627814 kubelet[2972]: E0912 22:52:42.625799 2972 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.29.105:6443/api/v1/namespaces/default/events\": dial tcp 172.31.29.105:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-29-105.1864aacf76965bc3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-29-105,UID:ip-172-31-29-105,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-29-105,},FirstTimestamp:2025-09-12 22:52:42.605853635 +0000 UTC m=+0.976597634,LastTimestamp:2025-09-12 22:52:42.605853635 +0000 UTC m=+0.976597634,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-29-105,}" Sep 12 22:52:42.631332 kubelet[2972]: I0912 22:52:42.631289 2972 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 22:52:42.632265 kubelet[2972]: I0912 22:52:42.631628 2972 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 22:52:42.633114 kubelet[2972]: E0912 22:52:42.632918 2972 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-29-105\" not found" Sep 12 22:52:42.633368 kubelet[2972]: I0912 22:52:42.633332 2972 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 12 22:52:42.634011 kubelet[2972]: I0912 22:52:42.633917 2972 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 12 22:52:42.634011 kubelet[2972]: I0912 22:52:42.633980 2972 reconciler.go:26] "Reconciler: start to sync state" Sep 12 22:52:42.636544 kubelet[2972]: E0912 22:52:42.635431 2972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.105:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-105?timeout=10s\": dial tcp 172.31.29.105:6443: connect: connection refused" interval="200ms" Sep 12 22:52:42.636544 kubelet[2972]: E0912 22:52:42.636276 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.29.105:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 12 22:52:42.637826 kubelet[2972]: I0912 22:52:42.637784 2972 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 22:52:42.640998 kubelet[2972]: E0912 22:52:42.640879 2972 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 22:52:42.640998 kubelet[2972]: I0912 22:52:42.640938 2972 factory.go:223] Registration of the containerd container factory successfully Sep 12 22:52:42.640998 kubelet[2972]: I0912 22:52:42.640954 2972 factory.go:223] Registration of the systemd container factory successfully Sep 12 22:52:42.663420 kubelet[2972]: I0912 22:52:42.663381 2972 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 12 22:52:42.663420 kubelet[2972]: I0912 22:52:42.663404 2972 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 12 22:52:42.663420 kubelet[2972]: I0912 22:52:42.663429 2972 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:52:42.670121 kubelet[2972]: I0912 22:52:42.669627 2972 policy_none.go:49] "None policy: Start" Sep 12 22:52:42.670121 kubelet[2972]: I0912 22:52:42.669657 2972 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 12 22:52:42.670121 kubelet[2972]: I0912 22:52:42.669672 2972 state_mem.go:35] "Initializing new in-memory state store" Sep 12 22:52:42.684294 kubelet[2972]: I0912 22:52:42.684227 2972 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 12 22:52:42.685000 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 12 22:52:42.689518 kubelet[2972]: I0912 22:52:42.689473 2972 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 12 22:52:42.689518 kubelet[2972]: I0912 22:52:42.689512 2972 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 12 22:52:42.689702 kubelet[2972]: I0912 22:52:42.689537 2972 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 12 22:52:42.689702 kubelet[2972]: I0912 22:52:42.689546 2972 kubelet.go:2436] "Starting kubelet main sync loop" Sep 12 22:52:42.689702 kubelet[2972]: E0912 22:52:42.689609 2972 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 22:52:42.691435 kubelet[2972]: E0912 22:52:42.690941 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.29.105:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 12 22:52:42.699607 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 12 22:52:42.711194 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 12 22:52:42.713841 kubelet[2972]: E0912 22:52:42.713799 2972 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 12 22:52:42.715919 kubelet[2972]: I0912 22:52:42.715447 2972 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 22:52:42.715919 kubelet[2972]: I0912 22:52:42.715526 2972 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 22:52:42.717003 kubelet[2972]: I0912 22:52:42.716596 2972 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 22:52:42.717401 kubelet[2972]: E0912 22:52:42.717384 2972 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 12 22:52:42.717593 kubelet[2972]: E0912 22:52:42.717579 2972 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-29-105\" not found" Sep 12 22:52:42.804845 systemd[1]: Created slice kubepods-burstable-pod7aeb5d2d42de14134cd0aac272f7daee.slice - libcontainer container kubepods-burstable-pod7aeb5d2d42de14134cd0aac272f7daee.slice. Sep 12 22:52:42.815896 kubelet[2972]: E0912 22:52:42.815860 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:42.817826 kubelet[2972]: I0912 22:52:42.817795 2972 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-29-105" Sep 12 22:52:42.818156 kubelet[2972]: E0912 22:52:42.818114 2972 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.29.105:6443/api/v1/nodes\": dial tcp 172.31.29.105:6443: connect: connection refused" node="ip-172-31-29-105" Sep 12 22:52:42.821848 systemd[1]: Created slice kubepods-burstable-pode1a23cc86b4cec69151cec61c3c9fb5a.slice - libcontainer container kubepods-burstable-pode1a23cc86b4cec69151cec61c3c9fb5a.slice. Sep 12 22:52:42.830114 kubelet[2972]: E0912 22:52:42.830084 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:42.833605 systemd[1]: Created slice kubepods-burstable-pod22a7d3bcb43acb469294dcca7bbcbe4e.slice - libcontainer container kubepods-burstable-pod22a7d3bcb43acb469294dcca7bbcbe4e.slice. Sep 12 22:52:42.838494 kubelet[2972]: E0912 22:52:42.836581 2972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.105:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-105?timeout=10s\": dial tcp 172.31.29.105:6443: connect: connection refused" interval="400ms" Sep 12 22:52:42.838494 kubelet[2972]: I0912 22:52:42.836761 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:42.838494 kubelet[2972]: I0912 22:52:42.836798 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7aeb5d2d42de14134cd0aac272f7daee-kubeconfig\") pod \"kube-scheduler-ip-172-31-29-105\" (UID: \"7aeb5d2d42de14134cd0aac272f7daee\") " pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:42.838494 kubelet[2972]: I0912 22:52:42.836826 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e1a23cc86b4cec69151cec61c3c9fb5a-ca-certs\") pod \"kube-apiserver-ip-172-31-29-105\" (UID: \"e1a23cc86b4cec69151cec61c3c9fb5a\") " pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:42.838494 kubelet[2972]: I0912 22:52:42.836901 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e1a23cc86b4cec69151cec61c3c9fb5a-k8s-certs\") pod \"kube-apiserver-ip-172-31-29-105\" (UID: \"e1a23cc86b4cec69151cec61c3c9fb5a\") " pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:42.838833 kubelet[2972]: I0912 22:52:42.836937 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-ca-certs\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:42.838833 kubelet[2972]: I0912 22:52:42.836970 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:42.838833 kubelet[2972]: I0912 22:52:42.837022 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-k8s-certs\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:42.838833 kubelet[2972]: I0912 22:52:42.837066 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-kubeconfig\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:42.840489 kubelet[2972]: E0912 22:52:42.840420 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:42.937938 kubelet[2972]: I0912 22:52:42.937702 2972 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e1a23cc86b4cec69151cec61c3c9fb5a-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-29-105\" (UID: \"e1a23cc86b4cec69151cec61c3c9fb5a\") " pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:43.020721 kubelet[2972]: I0912 22:52:43.020688 2972 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-29-105" Sep 12 22:52:43.021203 kubelet[2972]: E0912 22:52:43.021149 2972 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.29.105:6443/api/v1/nodes\": dial tcp 172.31.29.105:6443: connect: connection refused" node="ip-172-31-29-105" Sep 12 22:52:43.119180 containerd[2017]: time="2025-09-12T22:52:43.119045675Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-29-105,Uid:7aeb5d2d42de14134cd0aac272f7daee,Namespace:kube-system,Attempt:0,}" Sep 12 22:52:43.133163 containerd[2017]: time="2025-09-12T22:52:43.133109580Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-29-105,Uid:e1a23cc86b4cec69151cec61c3c9fb5a,Namespace:kube-system,Attempt:0,}" Sep 12 22:52:43.141735 containerd[2017]: time="2025-09-12T22:52:43.141679243Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-29-105,Uid:22a7d3bcb43acb469294dcca7bbcbe4e,Namespace:kube-system,Attempt:0,}" Sep 12 22:52:43.243666 kubelet[2972]: E0912 22:52:43.243619 2972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.105:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-105?timeout=10s\": dial tcp 172.31.29.105:6443: connect: connection refused" interval="800ms" Sep 12 22:52:43.280534 containerd[2017]: time="2025-09-12T22:52:43.280438061Z" level=info msg="connecting to shim ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714" address="unix:///run/containerd/s/077aeb1a6486decd148f8a1d701e244608fdb91e969d1e782287aee03568179f" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:52:43.281979 containerd[2017]: time="2025-09-12T22:52:43.281870583Z" level=info msg="connecting to shim 8a98106922830bd1cd82a61dc266a1303de93b06b4bdb861cf607db6a74fcec4" address="unix:///run/containerd/s/53553cee3a8ff680e4f749886554072a488a3842d58cecb271da727ed7166def" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:52:43.286483 containerd[2017]: time="2025-09-12T22:52:43.286375719Z" level=info msg="connecting to shim 9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f" address="unix:///run/containerd/s/c988aa1ce1fcf3f05a530c7d8c38cd9a08fd6181e931f806eda982958cdf9055" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:52:43.411192 systemd[1]: Started cri-containerd-8a98106922830bd1cd82a61dc266a1303de93b06b4bdb861cf607db6a74fcec4.scope - libcontainer container 8a98106922830bd1cd82a61dc266a1303de93b06b4bdb861cf607db6a74fcec4. Sep 12 22:52:43.414641 systemd[1]: Started cri-containerd-9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f.scope - libcontainer container 9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f. Sep 12 22:52:43.417364 systemd[1]: Started cri-containerd-ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714.scope - libcontainer container ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714. Sep 12 22:52:43.429831 kubelet[2972]: I0912 22:52:43.429369 2972 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-29-105" Sep 12 22:52:43.431658 kubelet[2972]: E0912 22:52:43.431498 2972 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.29.105:6443/api/v1/nodes\": dial tcp 172.31.29.105:6443: connect: connection refused" node="ip-172-31-29-105" Sep 12 22:52:43.450997 kubelet[2972]: E0912 22:52:43.450910 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.29.105:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-29-105&limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Sep 12 22:52:43.554003 containerd[2017]: time="2025-09-12T22:52:43.553565555Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-29-105,Uid:e1a23cc86b4cec69151cec61c3c9fb5a,Namespace:kube-system,Attempt:0,} returns sandbox id \"8a98106922830bd1cd82a61dc266a1303de93b06b4bdb861cf607db6a74fcec4\"" Sep 12 22:52:43.568518 containerd[2017]: time="2025-09-12T22:52:43.567954745Z" level=info msg="CreateContainer within sandbox \"8a98106922830bd1cd82a61dc266a1303de93b06b4bdb861cf607db6a74fcec4\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 12 22:52:43.568518 containerd[2017]: time="2025-09-12T22:52:43.568396006Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-29-105,Uid:7aeb5d2d42de14134cd0aac272f7daee,Namespace:kube-system,Attempt:0,} returns sandbox id \"9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f\"" Sep 12 22:52:43.587789 containerd[2017]: time="2025-09-12T22:52:43.587621593Z" level=info msg="Container a3dd446d661f89b92b593456115888e88b8408256590705e22b13acacb6d1135: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:52:43.588720 containerd[2017]: time="2025-09-12T22:52:43.588682034Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-29-105,Uid:22a7d3bcb43acb469294dcca7bbcbe4e,Namespace:kube-system,Attempt:0,} returns sandbox id \"ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714\"" Sep 12 22:52:43.588999 containerd[2017]: time="2025-09-12T22:52:43.588954340Z" level=info msg="CreateContainer within sandbox \"9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 12 22:52:43.598179 containerd[2017]: time="2025-09-12T22:52:43.598103599Z" level=info msg="CreateContainer within sandbox \"ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 12 22:52:43.611187 containerd[2017]: time="2025-09-12T22:52:43.611127527Z" level=info msg="Container 2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:52:43.619639 containerd[2017]: time="2025-09-12T22:52:43.619509195Z" level=info msg="CreateContainer within sandbox \"8a98106922830bd1cd82a61dc266a1303de93b06b4bdb861cf607db6a74fcec4\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"a3dd446d661f89b92b593456115888e88b8408256590705e22b13acacb6d1135\"" Sep 12 22:52:43.621532 containerd[2017]: time="2025-09-12T22:52:43.621135887Z" level=info msg="StartContainer for \"a3dd446d661f89b92b593456115888e88b8408256590705e22b13acacb6d1135\"" Sep 12 22:52:43.624371 containerd[2017]: time="2025-09-12T22:52:43.624252603Z" level=info msg="connecting to shim a3dd446d661f89b92b593456115888e88b8408256590705e22b13acacb6d1135" address="unix:///run/containerd/s/53553cee3a8ff680e4f749886554072a488a3842d58cecb271da727ed7166def" protocol=ttrpc version=3 Sep 12 22:52:43.636559 containerd[2017]: time="2025-09-12T22:52:43.636302534Z" level=info msg="Container 2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:52:43.638314 kubelet[2972]: E0912 22:52:43.638244 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.29.105:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Sep 12 22:52:43.639051 containerd[2017]: time="2025-09-12T22:52:43.639011716Z" level=info msg="CreateContainer within sandbox \"9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3\"" Sep 12 22:52:43.646588 containerd[2017]: time="2025-09-12T22:52:43.646546916Z" level=info msg="StartContainer for \"2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3\"" Sep 12 22:52:43.652895 containerd[2017]: time="2025-09-12T22:52:43.652730853Z" level=info msg="connecting to shim 2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3" address="unix:///run/containerd/s/c988aa1ce1fcf3f05a530c7d8c38cd9a08fd6181e931f806eda982958cdf9055" protocol=ttrpc version=3 Sep 12 22:52:43.661033 containerd[2017]: time="2025-09-12T22:52:43.660876393Z" level=info msg="CreateContainer within sandbox \"ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a\"" Sep 12 22:52:43.661886 systemd[1]: Started cri-containerd-a3dd446d661f89b92b593456115888e88b8408256590705e22b13acacb6d1135.scope - libcontainer container a3dd446d661f89b92b593456115888e88b8408256590705e22b13acacb6d1135. Sep 12 22:52:43.665910 containerd[2017]: time="2025-09-12T22:52:43.665043604Z" level=info msg="StartContainer for \"2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a\"" Sep 12 22:52:43.670731 containerd[2017]: time="2025-09-12T22:52:43.670677535Z" level=info msg="connecting to shim 2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a" address="unix:///run/containerd/s/077aeb1a6486decd148f8a1d701e244608fdb91e969d1e782287aee03568179f" protocol=ttrpc version=3 Sep 12 22:52:43.693907 systemd[1]: Started cri-containerd-2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3.scope - libcontainer container 2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3. Sep 12 22:52:43.729720 systemd[1]: Started cri-containerd-2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a.scope - libcontainer container 2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a. Sep 12 22:52:43.808797 containerd[2017]: time="2025-09-12T22:52:43.808604359Z" level=info msg="StartContainer for \"a3dd446d661f89b92b593456115888e88b8408256590705e22b13acacb6d1135\" returns successfully" Sep 12 22:52:43.843657 containerd[2017]: time="2025-09-12T22:52:43.843617707Z" level=info msg="StartContainer for \"2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3\" returns successfully" Sep 12 22:52:43.855126 kubelet[2972]: E0912 22:52:43.854965 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.29.105:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Sep 12 22:52:43.869613 containerd[2017]: time="2025-09-12T22:52:43.869534722Z" level=info msg="StartContainer for \"2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a\" returns successfully" Sep 12 22:52:44.046942 kubelet[2972]: E0912 22:52:44.046890 2972 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.29.105:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-105?timeout=10s\": dial tcp 172.31.29.105:6443: connect: connection refused" interval="1.6s" Sep 12 22:52:44.192094 kubelet[2972]: E0912 22:52:44.192046 2972 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.29.105:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.29.105:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Sep 12 22:52:44.235641 kubelet[2972]: I0912 22:52:44.235034 2972 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-29-105" Sep 12 22:52:44.235641 kubelet[2972]: E0912 22:52:44.235381 2972 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.29.105:6443/api/v1/nodes\": dial tcp 172.31.29.105:6443: connect: connection refused" node="ip-172-31-29-105" Sep 12 22:52:44.731229 kubelet[2972]: E0912 22:52:44.731194 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:44.740485 kubelet[2972]: E0912 22:52:44.738747 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:44.744370 kubelet[2972]: E0912 22:52:44.744328 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:45.749147 kubelet[2972]: E0912 22:52:45.749094 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:45.750272 kubelet[2972]: E0912 22:52:45.750240 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:45.750866 kubelet[2972]: E0912 22:52:45.750840 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:45.838205 kubelet[2972]: I0912 22:52:45.838175 2972 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-29-105" Sep 12 22:52:46.751265 kubelet[2972]: E0912 22:52:46.751228 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:46.752905 kubelet[2972]: E0912 22:52:46.752878 2972 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:47.234744 kubelet[2972]: E0912 22:52:47.234706 2972 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-29-105\" not found" node="ip-172-31-29-105" Sep 12 22:52:47.302061 kubelet[2972]: I0912 22:52:47.301802 2972 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-29-105" Sep 12 22:52:47.302061 kubelet[2972]: E0912 22:52:47.301854 2972 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ip-172-31-29-105\": node \"ip-172-31-29-105\" not found" Sep 12 22:52:47.335941 kubelet[2972]: I0912 22:52:47.335894 2972 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:47.395963 kubelet[2972]: E0912 22:52:47.395821 2972 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-29-105\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:47.395963 kubelet[2972]: I0912 22:52:47.395852 2972 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:47.397839 kubelet[2972]: E0912 22:52:47.397804 2972 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-29-105\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:47.397839 kubelet[2972]: I0912 22:52:47.397834 2972 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:47.400957 kubelet[2972]: E0912 22:52:47.399339 2972 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-29-105\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:47.601897 kubelet[2972]: I0912 22:52:47.601382 2972 apiserver.go:52] "Watching apiserver" Sep 12 22:52:47.634641 kubelet[2972]: I0912 22:52:47.634596 2972 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 12 22:52:48.693314 kubelet[2972]: I0912 22:52:48.693288 2972 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:49.429870 kubelet[2972]: I0912 22:52:49.429817 2972 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:49.585309 systemd[1]: Reload requested from client PID 3248 ('systemctl') (unit session-9.scope)... Sep 12 22:52:49.585328 systemd[1]: Reloading... Sep 12 22:52:49.712510 zram_generator::config[3298]: No configuration found. Sep 12 22:52:50.068628 systemd[1]: Reloading finished in 482 ms. Sep 12 22:52:50.103834 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:52:50.118644 systemd[1]: kubelet.service: Deactivated successfully. Sep 12 22:52:50.118903 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:50.118967 systemd[1]: kubelet.service: Consumed 1.474s CPU time, 130.1M memory peak. Sep 12 22:52:50.121134 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 12 22:52:50.407774 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 12 22:52:50.420040 (kubelet)[3352]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 12 22:52:50.488738 kubelet[3352]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:52:50.488738 kubelet[3352]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Sep 12 22:52:50.488738 kubelet[3352]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 12 22:52:50.489203 kubelet[3352]: I0912 22:52:50.488804 3352 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 12 22:52:50.496916 kubelet[3352]: I0912 22:52:50.496873 3352 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Sep 12 22:52:50.496916 kubelet[3352]: I0912 22:52:50.496906 3352 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 12 22:52:50.497222 kubelet[3352]: I0912 22:52:50.497198 3352 server.go:956] "Client rotation is on, will bootstrap in background" Sep 12 22:52:50.499340 kubelet[3352]: I0912 22:52:50.498535 3352 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Sep 12 22:52:50.501579 kubelet[3352]: I0912 22:52:50.501519 3352 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 12 22:52:50.507118 kubelet[3352]: I0912 22:52:50.507084 3352 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 12 22:52:50.517878 kubelet[3352]: I0912 22:52:50.517851 3352 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 12 22:52:50.518370 kubelet[3352]: I0912 22:52:50.518341 3352 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 12 22:52:50.518637 kubelet[3352]: I0912 22:52:50.518444 3352 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-29-105","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 12 22:52:50.518785 kubelet[3352]: I0912 22:52:50.518775 3352 topology_manager.go:138] "Creating topology manager with none policy" Sep 12 22:52:50.518833 kubelet[3352]: I0912 22:52:50.518828 3352 container_manager_linux.go:303] "Creating device plugin manager" Sep 12 22:52:50.518907 kubelet[3352]: I0912 22:52:50.518901 3352 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:52:50.519161 kubelet[3352]: I0912 22:52:50.519151 3352 kubelet.go:480] "Attempting to sync node with API server" Sep 12 22:52:50.519255 kubelet[3352]: I0912 22:52:50.519245 3352 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 12 22:52:50.519336 kubelet[3352]: I0912 22:52:50.519328 3352 kubelet.go:386] "Adding apiserver pod source" Sep 12 22:52:50.519417 kubelet[3352]: I0912 22:52:50.519406 3352 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 12 22:52:50.523619 kubelet[3352]: I0912 22:52:50.523585 3352 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Sep 12 22:52:50.524350 kubelet[3352]: I0912 22:52:50.524328 3352 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Sep 12 22:52:50.531008 kubelet[3352]: I0912 22:52:50.530981 3352 watchdog_linux.go:99] "Systemd watchdog is not enabled" Sep 12 22:52:50.532481 kubelet[3352]: I0912 22:52:50.531043 3352 server.go:1289] "Started kubelet" Sep 12 22:52:50.536648 kubelet[3352]: E0912 22:52:50.536620 3352 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 12 22:52:50.536998 kubelet[3352]: I0912 22:52:50.536951 3352 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 12 22:52:50.537357 kubelet[3352]: I0912 22:52:50.537345 3352 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 12 22:52:50.537499 kubelet[3352]: I0912 22:52:50.537480 3352 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Sep 12 22:52:50.538026 kubelet[3352]: I0912 22:52:50.538001 3352 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 12 22:52:50.539407 kubelet[3352]: I0912 22:52:50.539366 3352 server.go:317] "Adding debug handlers to kubelet server" Sep 12 22:52:50.547856 kubelet[3352]: I0912 22:52:50.547816 3352 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 12 22:52:50.551712 kubelet[3352]: I0912 22:52:50.551613 3352 volume_manager.go:297] "Starting Kubelet Volume Manager" Sep 12 22:52:50.552168 kubelet[3352]: E0912 22:52:50.552096 3352 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-29-105\" not found" Sep 12 22:52:50.553594 kubelet[3352]: I0912 22:52:50.552915 3352 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Sep 12 22:52:50.553594 kubelet[3352]: I0912 22:52:50.553044 3352 reconciler.go:26] "Reconciler: start to sync state" Sep 12 22:52:50.557716 kubelet[3352]: I0912 22:52:50.557679 3352 factory.go:223] Registration of the systemd container factory successfully Sep 12 22:52:50.557864 kubelet[3352]: I0912 22:52:50.557848 3352 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 12 22:52:50.564075 kubelet[3352]: I0912 22:52:50.563601 3352 factory.go:223] Registration of the containerd container factory successfully Sep 12 22:52:50.603846 kubelet[3352]: I0912 22:52:50.603780 3352 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Sep 12 22:52:50.608683 kubelet[3352]: I0912 22:52:50.608644 3352 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Sep 12 22:52:50.608859 kubelet[3352]: I0912 22:52:50.608849 3352 status_manager.go:230] "Starting to sync pod status with apiserver" Sep 12 22:52:50.610107 kubelet[3352]: I0912 22:52:50.610091 3352 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Sep 12 22:52:50.610744 kubelet[3352]: I0912 22:52:50.610731 3352 kubelet.go:2436] "Starting kubelet main sync loop" Sep 12 22:52:50.613562 kubelet[3352]: E0912 22:52:50.613502 3352 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671140 3352 cpu_manager.go:221] "Starting CPU manager" policy="none" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671189 3352 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671210 3352 state_mem.go:36] "Initialized new in-memory state store" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671389 3352 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671402 3352 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671425 3352 policy_none.go:49] "None policy: Start" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671437 3352 memory_manager.go:186] "Starting memorymanager" policy="None" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671470 3352 state_mem.go:35] "Initializing new in-memory state store" Sep 12 22:52:50.673300 kubelet[3352]: I0912 22:52:50.671599 3352 state_mem.go:75] "Updated machine memory state" Sep 12 22:52:50.681589 kubelet[3352]: E0912 22:52:50.681409 3352 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Sep 12 22:52:50.684788 kubelet[3352]: I0912 22:52:50.684654 3352 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 12 22:52:50.686498 kubelet[3352]: I0912 22:52:50.685896 3352 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 12 22:52:50.686718 kubelet[3352]: I0912 22:52:50.686683 3352 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 12 22:52:50.691117 kubelet[3352]: E0912 22:52:50.691089 3352 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Sep 12 22:52:50.714848 kubelet[3352]: I0912 22:52:50.714805 3352 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:50.716998 kubelet[3352]: I0912 22:52:50.716728 3352 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:50.718487 kubelet[3352]: I0912 22:52:50.718439 3352 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:50.742221 kubelet[3352]: E0912 22:52:50.741586 3352 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-29-105\" already exists" pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:50.742221 kubelet[3352]: E0912 22:52:50.741746 3352 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-29-105\" already exists" pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:50.807479 kubelet[3352]: I0912 22:52:50.807430 3352 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-29-105" Sep 12 22:52:50.816770 kubelet[3352]: I0912 22:52:50.816253 3352 kubelet_node_status.go:124] "Node was previously registered" node="ip-172-31-29-105" Sep 12 22:52:50.816770 kubelet[3352]: I0912 22:52:50.816338 3352 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-29-105" Sep 12 22:52:50.855034 kubelet[3352]: I0912 22:52:50.854925 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7aeb5d2d42de14134cd0aac272f7daee-kubeconfig\") pod \"kube-scheduler-ip-172-31-29-105\" (UID: \"7aeb5d2d42de14134cd0aac272f7daee\") " pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:50.855174 kubelet[3352]: I0912 22:52:50.855042 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/e1a23cc86b4cec69151cec61c3c9fb5a-ca-certs\") pod \"kube-apiserver-ip-172-31-29-105\" (UID: \"e1a23cc86b4cec69151cec61c3c9fb5a\") " pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:50.855174 kubelet[3352]: I0912 22:52:50.855072 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/e1a23cc86b4cec69151cec61c3c9fb5a-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-29-105\" (UID: \"e1a23cc86b4cec69151cec61c3c9fb5a\") " pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:50.855174 kubelet[3352]: I0912 22:52:50.855100 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:50.855174 kubelet[3352]: I0912 22:52:50.855122 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-k8s-certs\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:50.855174 kubelet[3352]: I0912 22:52:50.855144 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-kubeconfig\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:50.855406 kubelet[3352]: I0912 22:52:50.855166 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:50.855406 kubelet[3352]: I0912 22:52:50.855189 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/e1a23cc86b4cec69151cec61c3c9fb5a-k8s-certs\") pod \"kube-apiserver-ip-172-31-29-105\" (UID: \"e1a23cc86b4cec69151cec61c3c9fb5a\") " pod="kube-system/kube-apiserver-ip-172-31-29-105" Sep 12 22:52:50.855406 kubelet[3352]: I0912 22:52:50.855229 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/22a7d3bcb43acb469294dcca7bbcbe4e-ca-certs\") pod \"kube-controller-manager-ip-172-31-29-105\" (UID: \"22a7d3bcb43acb469294dcca7bbcbe4e\") " pod="kube-system/kube-controller-manager-ip-172-31-29-105" Sep 12 22:52:51.261551 update_engine[1984]: I20250912 22:52:51.261500 1984 update_attempter.cc:509] Updating boot flags... Sep 12 22:52:51.523747 kubelet[3352]: I0912 22:52:51.523545 3352 apiserver.go:52] "Watching apiserver" Sep 12 22:52:51.554583 kubelet[3352]: I0912 22:52:51.553518 3352 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Sep 12 22:52:51.657600 kubelet[3352]: I0912 22:52:51.657572 3352 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:51.673928 kubelet[3352]: E0912 22:52:51.673760 3352 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-29-105\" already exists" pod="kube-system/kube-scheduler-ip-172-31-29-105" Sep 12 22:52:51.772921 kubelet[3352]: I0912 22:52:51.772677 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-29-105" podStartSLOduration=2.772657813 podStartE2EDuration="2.772657813s" podCreationTimestamp="2025-09-12 22:52:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:52:51.772296755 +0000 UTC m=+1.342046922" watchObservedRunningTime="2025-09-12 22:52:51.772657813 +0000 UTC m=+1.342407978" Sep 12 22:52:51.775480 kubelet[3352]: I0912 22:52:51.774554 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-29-105" podStartSLOduration=3.774533722 podStartE2EDuration="3.774533722s" podCreationTimestamp="2025-09-12 22:52:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:52:51.739010769 +0000 UTC m=+1.308760937" watchObservedRunningTime="2025-09-12 22:52:51.774533722 +0000 UTC m=+1.344283889" Sep 12 22:52:51.842574 kubelet[3352]: I0912 22:52:51.841031 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-29-105" podStartSLOduration=1.8410111489999998 podStartE2EDuration="1.841011149s" podCreationTimestamp="2025-09-12 22:52:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:52:51.840796944 +0000 UTC m=+1.410547111" watchObservedRunningTime="2025-09-12 22:52:51.841011149 +0000 UTC m=+1.410761317" Sep 12 22:52:54.319043 kubelet[3352]: I0912 22:52:54.318988 3352 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 12 22:52:54.319504 containerd[2017]: time="2025-09-12T22:52:54.319320527Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 12 22:52:54.320393 kubelet[3352]: I0912 22:52:54.319496 3352 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 12 22:52:55.372210 systemd[1]: Created slice kubepods-besteffort-podd481bbee_25c8_435f_b8a3_634db32cda6a.slice - libcontainer container kubepods-besteffort-podd481bbee_25c8_435f_b8a3_634db32cda6a.slice. Sep 12 22:52:55.390943 kubelet[3352]: I0912 22:52:55.390889 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/d481bbee-25c8-435f-b8a3-634db32cda6a-kube-proxy\") pod \"kube-proxy-tp8pb\" (UID: \"d481bbee-25c8-435f-b8a3-634db32cda6a\") " pod="kube-system/kube-proxy-tp8pb" Sep 12 22:52:55.390943 kubelet[3352]: I0912 22:52:55.390946 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d481bbee-25c8-435f-b8a3-634db32cda6a-lib-modules\") pod \"kube-proxy-tp8pb\" (UID: \"d481bbee-25c8-435f-b8a3-634db32cda6a\") " pod="kube-system/kube-proxy-tp8pb" Sep 12 22:52:55.392623 kubelet[3352]: I0912 22:52:55.390970 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhqmq\" (UniqueName: \"kubernetes.io/projected/d481bbee-25c8-435f-b8a3-634db32cda6a-kube-api-access-vhqmq\") pod \"kube-proxy-tp8pb\" (UID: \"d481bbee-25c8-435f-b8a3-634db32cda6a\") " pod="kube-system/kube-proxy-tp8pb" Sep 12 22:52:55.392623 kubelet[3352]: I0912 22:52:55.390995 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d481bbee-25c8-435f-b8a3-634db32cda6a-xtables-lock\") pod \"kube-proxy-tp8pb\" (UID: \"d481bbee-25c8-435f-b8a3-634db32cda6a\") " pod="kube-system/kube-proxy-tp8pb" Sep 12 22:52:55.543971 systemd[1]: Created slice kubepods-besteffort-pod8fa588a9_f073_4493_9c89_a0a42ba66c3f.slice - libcontainer container kubepods-besteffort-pod8fa588a9_f073_4493_9c89_a0a42ba66c3f.slice. Sep 12 22:52:55.592022 kubelet[3352]: I0912 22:52:55.591844 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v97wm\" (UniqueName: \"kubernetes.io/projected/8fa588a9-f073-4493-9c89-a0a42ba66c3f-kube-api-access-v97wm\") pod \"tigera-operator-755d956888-jj6hp\" (UID: \"8fa588a9-f073-4493-9c89-a0a42ba66c3f\") " pod="tigera-operator/tigera-operator-755d956888-jj6hp" Sep 12 22:52:55.592022 kubelet[3352]: I0912 22:52:55.591955 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/8fa588a9-f073-4493-9c89-a0a42ba66c3f-var-lib-calico\") pod \"tigera-operator-755d956888-jj6hp\" (UID: \"8fa588a9-f073-4493-9c89-a0a42ba66c3f\") " pod="tigera-operator/tigera-operator-755d956888-jj6hp" Sep 12 22:52:55.682414 containerd[2017]: time="2025-09-12T22:52:55.682091537Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-tp8pb,Uid:d481bbee-25c8-435f-b8a3-634db32cda6a,Namespace:kube-system,Attempt:0,}" Sep 12 22:52:55.725983 containerd[2017]: time="2025-09-12T22:52:55.725642985Z" level=info msg="connecting to shim f0a70762722d43f2c6885460c93d26feb97987d9993d0cac57973bd2e9ff79fb" address="unix:///run/containerd/s/5eaed06979e1b6ff7e5c863f7d515da7445d7e9c1dfb93621eb983c7ed4a4663" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:52:55.767708 systemd[1]: Started cri-containerd-f0a70762722d43f2c6885460c93d26feb97987d9993d0cac57973bd2e9ff79fb.scope - libcontainer container f0a70762722d43f2c6885460c93d26feb97987d9993d0cac57973bd2e9ff79fb. Sep 12 22:52:55.803168 containerd[2017]: time="2025-09-12T22:52:55.803113324Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-tp8pb,Uid:d481bbee-25c8-435f-b8a3-634db32cda6a,Namespace:kube-system,Attempt:0,} returns sandbox id \"f0a70762722d43f2c6885460c93d26feb97987d9993d0cac57973bd2e9ff79fb\"" Sep 12 22:52:55.813199 containerd[2017]: time="2025-09-12T22:52:55.813099737Z" level=info msg="CreateContainer within sandbox \"f0a70762722d43f2c6885460c93d26feb97987d9993d0cac57973bd2e9ff79fb\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 12 22:52:55.834177 containerd[2017]: time="2025-09-12T22:52:55.834057379Z" level=info msg="Container c04c67d7b9afbae7809792a9dbe05e3ebe4bea923a3ee0e9aacc0c2cd9cc4fbf: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:52:55.848596 containerd[2017]: time="2025-09-12T22:52:55.848521993Z" level=info msg="CreateContainer within sandbox \"f0a70762722d43f2c6885460c93d26feb97987d9993d0cac57973bd2e9ff79fb\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"c04c67d7b9afbae7809792a9dbe05e3ebe4bea923a3ee0e9aacc0c2cd9cc4fbf\"" Sep 12 22:52:55.848976 containerd[2017]: time="2025-09-12T22:52:55.848940449Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-jj6hp,Uid:8fa588a9-f073-4493-9c89-a0a42ba66c3f,Namespace:tigera-operator,Attempt:0,}" Sep 12 22:52:55.850495 containerd[2017]: time="2025-09-12T22:52:55.849774460Z" level=info msg="StartContainer for \"c04c67d7b9afbae7809792a9dbe05e3ebe4bea923a3ee0e9aacc0c2cd9cc4fbf\"" Sep 12 22:52:55.852434 containerd[2017]: time="2025-09-12T22:52:55.852387633Z" level=info msg="connecting to shim c04c67d7b9afbae7809792a9dbe05e3ebe4bea923a3ee0e9aacc0c2cd9cc4fbf" address="unix:///run/containerd/s/5eaed06979e1b6ff7e5c863f7d515da7445d7e9c1dfb93621eb983c7ed4a4663" protocol=ttrpc version=3 Sep 12 22:52:55.880756 systemd[1]: Started cri-containerd-c04c67d7b9afbae7809792a9dbe05e3ebe4bea923a3ee0e9aacc0c2cd9cc4fbf.scope - libcontainer container c04c67d7b9afbae7809792a9dbe05e3ebe4bea923a3ee0e9aacc0c2cd9cc4fbf. Sep 12 22:52:55.902757 containerd[2017]: time="2025-09-12T22:52:55.902704179Z" level=info msg="connecting to shim aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046" address="unix:///run/containerd/s/6e4ad363abbdf3e512a4e72be044a5595373f9ec95405b2e2b35491f25276544" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:52:55.943782 systemd[1]: Started cri-containerd-aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046.scope - libcontainer container aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046. Sep 12 22:52:55.975093 containerd[2017]: time="2025-09-12T22:52:55.975049046Z" level=info msg="StartContainer for \"c04c67d7b9afbae7809792a9dbe05e3ebe4bea923a3ee0e9aacc0c2cd9cc4fbf\" returns successfully" Sep 12 22:52:56.026199 containerd[2017]: time="2025-09-12T22:52:56.026156167Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-jj6hp,Uid:8fa588a9-f073-4493-9c89-a0a42ba66c3f,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046\"" Sep 12 22:52:56.028947 containerd[2017]: time="2025-09-12T22:52:56.028787163Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Sep 12 22:52:56.507727 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount983184256.mount: Deactivated successfully. Sep 12 22:52:56.675577 kubelet[3352]: I0912 22:52:56.674889 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-tp8pb" podStartSLOduration=1.6748715280000002 podStartE2EDuration="1.674871528s" podCreationTimestamp="2025-09-12 22:52:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:52:56.674713286 +0000 UTC m=+6.244463451" watchObservedRunningTime="2025-09-12 22:52:56.674871528 +0000 UTC m=+6.244621694" Sep 12 22:52:57.451130 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3845534394.mount: Deactivated successfully. Sep 12 22:52:58.268203 containerd[2017]: time="2025-09-12T22:52:58.268100553Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:58.270490 containerd[2017]: time="2025-09-12T22:52:58.270404316Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=25062609" Sep 12 22:52:58.273656 containerd[2017]: time="2025-09-12T22:52:58.273422909Z" level=info msg="ImageCreate event name:\"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:58.277534 containerd[2017]: time="2025-09-12T22:52:58.277489303Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:52:58.277995 containerd[2017]: time="2025-09-12T22:52:58.277960895Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"25058604\" in 2.2490178s" Sep 12 22:52:58.277995 containerd[2017]: time="2025-09-12T22:52:58.277997729Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\"" Sep 12 22:52:58.284581 containerd[2017]: time="2025-09-12T22:52:58.284542560Z" level=info msg="CreateContainer within sandbox \"aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 12 22:52:58.302107 containerd[2017]: time="2025-09-12T22:52:58.299051672Z" level=info msg="Container f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:52:58.309939 containerd[2017]: time="2025-09-12T22:52:58.309867885Z" level=info msg="CreateContainer within sandbox \"aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\"" Sep 12 22:52:58.310546 containerd[2017]: time="2025-09-12T22:52:58.310485424Z" level=info msg="StartContainer for \"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\"" Sep 12 22:52:58.311712 containerd[2017]: time="2025-09-12T22:52:58.311668435Z" level=info msg="connecting to shim f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f" address="unix:///run/containerd/s/6e4ad363abbdf3e512a4e72be044a5595373f9ec95405b2e2b35491f25276544" protocol=ttrpc version=3 Sep 12 22:52:58.334702 systemd[1]: Started cri-containerd-f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f.scope - libcontainer container f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f. Sep 12 22:52:58.378094 containerd[2017]: time="2025-09-12T22:52:58.378046502Z" level=info msg="StartContainer for \"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\" returns successfully" Sep 12 22:52:59.668080 kubelet[3352]: I0912 22:52:59.668013 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-jj6hp" podStartSLOduration=2.416936197 podStartE2EDuration="4.667995559s" podCreationTimestamp="2025-09-12 22:52:55 +0000 UTC" firstStartedPulling="2025-09-12 22:52:56.027685923 +0000 UTC m=+5.597436070" lastFinishedPulling="2025-09-12 22:52:58.278745286 +0000 UTC m=+7.848495432" observedRunningTime="2025-09-12 22:52:58.682483648 +0000 UTC m=+8.252233805" watchObservedRunningTime="2025-09-12 22:52:59.667995559 +0000 UTC m=+9.237745726" Sep 12 22:53:02.024012 systemd[1]: cri-containerd-f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f.scope: Deactivated successfully. Sep 12 22:53:02.160427 containerd[2017]: time="2025-09-12T22:53:02.160355090Z" level=info msg="received exit event container_id:\"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\" id:\"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\" pid:3864 exit_status:1 exited_at:{seconds:1757717582 nanos:30739115}" Sep 12 22:53:02.196757 containerd[2017]: time="2025-09-12T22:53:02.196685930Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\" id:\"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\" pid:3864 exit_status:1 exited_at:{seconds:1757717582 nanos:30739115}" Sep 12 22:53:02.244818 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f-rootfs.mount: Deactivated successfully. Sep 12 22:53:02.698430 kubelet[3352]: I0912 22:53:02.698323 3352 scope.go:117] "RemoveContainer" containerID="f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f" Sep 12 22:53:02.706381 containerd[2017]: time="2025-09-12T22:53:02.706313632Z" level=info msg="CreateContainer within sandbox \"aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Sep 12 22:53:02.727933 containerd[2017]: time="2025-09-12T22:53:02.727876300Z" level=info msg="Container efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:02.740589 containerd[2017]: time="2025-09-12T22:53:02.740522941Z" level=info msg="CreateContainer within sandbox \"aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544\"" Sep 12 22:53:02.741765 containerd[2017]: time="2025-09-12T22:53:02.741728312Z" level=info msg="StartContainer for \"efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544\"" Sep 12 22:53:02.746885 containerd[2017]: time="2025-09-12T22:53:02.746822887Z" level=info msg="connecting to shim efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544" address="unix:///run/containerd/s/6e4ad363abbdf3e512a4e72be044a5595373f9ec95405b2e2b35491f25276544" protocol=ttrpc version=3 Sep 12 22:53:02.794888 systemd[1]: Started cri-containerd-efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544.scope - libcontainer container efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544. Sep 12 22:53:02.873810 containerd[2017]: time="2025-09-12T22:53:02.873763866Z" level=info msg="StartContainer for \"efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544\" returns successfully" Sep 12 22:53:05.716294 sudo[2387]: pam_unix(sudo:session): session closed for user root Sep 12 22:53:05.738674 sshd[2386]: Connection closed by 139.178.89.65 port 39590 Sep 12 22:53:05.740778 sshd-session[2383]: pam_unix(sshd:session): session closed for user core Sep 12 22:53:05.762158 systemd[1]: sshd@8-172.31.29.105:22-139.178.89.65:39590.service: Deactivated successfully. Sep 12 22:53:05.771213 systemd[1]: session-9.scope: Deactivated successfully. Sep 12 22:53:05.771803 systemd[1]: session-9.scope: Consumed 5.200s CPU time, 152.5M memory peak. Sep 12 22:53:05.773994 systemd-logind[1983]: Session 9 logged out. Waiting for processes to exit. Sep 12 22:53:05.775944 systemd-logind[1983]: Removed session 9. Sep 12 22:53:12.380533 systemd[1]: Created slice kubepods-besteffort-pode5e5a21e_37e6_4a38_a1fe_6365d14368bf.slice - libcontainer container kubepods-besteffort-pode5e5a21e_37e6_4a38_a1fe_6365d14368bf.slice. Sep 12 22:53:12.416943 kubelet[3352]: I0912 22:53:12.416908 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tztcj\" (UniqueName: \"kubernetes.io/projected/e5e5a21e-37e6-4a38-a1fe-6365d14368bf-kube-api-access-tztcj\") pod \"calico-typha-c497bf79c-s826b\" (UID: \"e5e5a21e-37e6-4a38-a1fe-6365d14368bf\") " pod="calico-system/calico-typha-c497bf79c-s826b" Sep 12 22:53:12.419477 kubelet[3352]: I0912 22:53:12.418541 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/e5e5a21e-37e6-4a38-a1fe-6365d14368bf-typha-certs\") pod \"calico-typha-c497bf79c-s826b\" (UID: \"e5e5a21e-37e6-4a38-a1fe-6365d14368bf\") " pod="calico-system/calico-typha-c497bf79c-s826b" Sep 12 22:53:12.419477 kubelet[3352]: I0912 22:53:12.418603 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5e5a21e-37e6-4a38-a1fe-6365d14368bf-tigera-ca-bundle\") pod \"calico-typha-c497bf79c-s826b\" (UID: \"e5e5a21e-37e6-4a38-a1fe-6365d14368bf\") " pod="calico-system/calico-typha-c497bf79c-s826b" Sep 12 22:53:12.622501 systemd[1]: Created slice kubepods-besteffort-podd9133c96_59ef_4269_9edf_fa9cd593c1e6.slice - libcontainer container kubepods-besteffort-podd9133c96_59ef_4269_9edf_fa9cd593c1e6.slice. Sep 12 22:53:12.627735 kubelet[3352]: I0912 22:53:12.627681 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-var-lib-calico\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.627898 kubelet[3352]: I0912 22:53:12.627743 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-cni-net-dir\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.627898 kubelet[3352]: I0912 22:53:12.627764 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/d9133c96-59ef-4269-9edf-fa9cd593c1e6-node-certs\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.627898 kubelet[3352]: I0912 22:53:12.627788 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-lib-modules\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.627898 kubelet[3352]: I0912 22:53:12.627809 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-797c9\" (UniqueName: \"kubernetes.io/projected/d9133c96-59ef-4269-9edf-fa9cd593c1e6-kube-api-access-797c9\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.627898 kubelet[3352]: I0912 22:53:12.627833 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9133c96-59ef-4269-9edf-fa9cd593c1e6-tigera-ca-bundle\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.628089 kubelet[3352]: I0912 22:53:12.627853 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-policysync\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.628089 kubelet[3352]: I0912 22:53:12.627874 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-var-run-calico\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.628089 kubelet[3352]: I0912 22:53:12.627898 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-cni-bin-dir\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.628089 kubelet[3352]: I0912 22:53:12.627924 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-flexvol-driver-host\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.628089 kubelet[3352]: I0912 22:53:12.627949 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-cni-log-dir\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.628303 kubelet[3352]: I0912 22:53:12.627973 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d9133c96-59ef-4269-9edf-fa9cd593c1e6-xtables-lock\") pod \"calico-node-87b9s\" (UID: \"d9133c96-59ef-4269-9edf-fa9cd593c1e6\") " pod="calico-system/calico-node-87b9s" Sep 12 22:53:12.700253 containerd[2017]: time="2025-09-12T22:53:12.700114375Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-c497bf79c-s826b,Uid:e5e5a21e-37e6-4a38-a1fe-6365d14368bf,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:12.736588 kubelet[3352]: E0912 22:53:12.736522 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:12.736847 kubelet[3352]: W0912 22:53:12.736559 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:12.736847 kubelet[3352]: E0912 22:53:12.736797 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:12.752515 kubelet[3352]: E0912 22:53:12.749581 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:12.752515 kubelet[3352]: W0912 22:53:12.749611 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:12.752515 kubelet[3352]: E0912 22:53:12.749641 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:12.755413 kubelet[3352]: E0912 22:53:12.752873 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:12.756013 kubelet[3352]: W0912 22:53:12.755620 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:12.756013 kubelet[3352]: E0912 22:53:12.755685 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:12.759755 containerd[2017]: time="2025-09-12T22:53:12.755640790Z" level=info msg="connecting to shim 3d70fdf1728f5df7beacb3e1040212091464888ae95c689eb088b5387a1fc6d5" address="unix:///run/containerd/s/583e4eb43046ffa498788e155f55329606a570289829befbc0e6c476e1ba954a" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:12.789149 kubelet[3352]: E0912 22:53:12.789076 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:12.790373 kubelet[3352]: W0912 22:53:12.789510 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:12.790373 kubelet[3352]: E0912 22:53:12.789708 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:12.825860 systemd[1]: Started cri-containerd-3d70fdf1728f5df7beacb3e1040212091464888ae95c689eb088b5387a1fc6d5.scope - libcontainer container 3d70fdf1728f5df7beacb3e1040212091464888ae95c689eb088b5387a1fc6d5. Sep 12 22:53:12.940789 containerd[2017]: time="2025-09-12T22:53:12.940729353Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-87b9s,Uid:d9133c96-59ef-4269-9edf-fa9cd593c1e6,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:12.944276 containerd[2017]: time="2025-09-12T22:53:12.944224382Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-c497bf79c-s826b,Uid:e5e5a21e-37e6-4a38-a1fe-6365d14368bf,Namespace:calico-system,Attempt:0,} returns sandbox id \"3d70fdf1728f5df7beacb3e1040212091464888ae95c689eb088b5387a1fc6d5\"" Sep 12 22:53:12.951070 containerd[2017]: time="2025-09-12T22:53:12.950298883Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Sep 12 22:53:12.961113 kubelet[3352]: E0912 22:53:12.961036 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5nmmm" podUID="6db03382-69c2-445f-a6ae-af82129366d4" Sep 12 22:53:13.022909 kubelet[3352]: E0912 22:53:13.022837 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.022909 kubelet[3352]: W0912 22:53:13.022892 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.023501 kubelet[3352]: E0912 22:53:13.022923 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.024411 kubelet[3352]: E0912 22:53:13.024341 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.024411 kubelet[3352]: W0912 22:53:13.024363 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.024411 kubelet[3352]: E0912 22:53:13.024386 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.025004 kubelet[3352]: E0912 22:53:13.024836 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.025004 kubelet[3352]: W0912 22:53:13.024849 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.025256 kubelet[3352]: E0912 22:53:13.025095 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.026422 kubelet[3352]: E0912 22:53:13.026391 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.026638 kubelet[3352]: W0912 22:53:13.026428 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.026638 kubelet[3352]: E0912 22:53:13.026476 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.027794 kubelet[3352]: E0912 22:53:13.027703 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.027794 kubelet[3352]: W0912 22:53:13.027749 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.027794 kubelet[3352]: E0912 22:53:13.027767 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.029000 kubelet[3352]: E0912 22:53:13.028973 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.029000 kubelet[3352]: W0912 22:53:13.028995 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.029130 kubelet[3352]: E0912 22:53:13.029012 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.030823 kubelet[3352]: E0912 22:53:13.030789 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.030823 kubelet[3352]: W0912 22:53:13.030814 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.030967 kubelet[3352]: E0912 22:53:13.030834 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.031766 kubelet[3352]: E0912 22:53:13.031694 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.031766 kubelet[3352]: W0912 22:53:13.031762 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.031892 kubelet[3352]: E0912 22:53:13.031780 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.035626 kubelet[3352]: E0912 22:53:13.035573 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.035816 kubelet[3352]: W0912 22:53:13.035731 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.035816 kubelet[3352]: E0912 22:53:13.035760 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.038609 kubelet[3352]: E0912 22:53:13.038580 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.038609 kubelet[3352]: W0912 22:53:13.038608 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.038780 kubelet[3352]: E0912 22:53:13.038631 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.041148 kubelet[3352]: E0912 22:53:13.041102 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.041148 kubelet[3352]: W0912 22:53:13.041132 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.041500 kubelet[3352]: E0912 22:53:13.041155 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.041933 kubelet[3352]: E0912 22:53:13.041896 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.042032 kubelet[3352]: W0912 22:53:13.041951 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.042032 kubelet[3352]: E0912 22:53:13.041970 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.043168 kubelet[3352]: E0912 22:53:13.043136 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.043168 kubelet[3352]: W0912 22:53:13.043166 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.043328 kubelet[3352]: E0912 22:53:13.043184 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.045638 kubelet[3352]: E0912 22:53:13.045587 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.045638 kubelet[3352]: W0912 22:53:13.045637 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.045823 kubelet[3352]: E0912 22:53:13.045655 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.046139 kubelet[3352]: E0912 22:53:13.046085 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.046139 kubelet[3352]: W0912 22:53:13.046108 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.046139 kubelet[3352]: E0912 22:53:13.046133 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.049818 kubelet[3352]: E0912 22:53:13.049781 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.049818 kubelet[3352]: W0912 22:53:13.049818 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.050038 kubelet[3352]: E0912 22:53:13.049840 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.050288 kubelet[3352]: E0912 22:53:13.050232 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.050288 kubelet[3352]: W0912 22:53:13.050275 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.050405 kubelet[3352]: E0912 22:53:13.050295 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.050981 kubelet[3352]: E0912 22:53:13.050952 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.050981 kubelet[3352]: W0912 22:53:13.050970 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.051123 kubelet[3352]: E0912 22:53:13.050996 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.051422 kubelet[3352]: E0912 22:53:13.051392 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.051422 kubelet[3352]: W0912 22:53:13.051420 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.051623 kubelet[3352]: E0912 22:53:13.051546 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.052694 kubelet[3352]: E0912 22:53:13.052669 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.052694 kubelet[3352]: W0912 22:53:13.052690 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.052839 kubelet[3352]: E0912 22:53:13.052705 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.053574 kubelet[3352]: E0912 22:53:13.053551 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.053574 kubelet[3352]: W0912 22:53:13.053571 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.053707 kubelet[3352]: E0912 22:53:13.053586 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.053707 kubelet[3352]: I0912 22:53:13.053626 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wdhn\" (UniqueName: \"kubernetes.io/projected/6db03382-69c2-445f-a6ae-af82129366d4-kube-api-access-4wdhn\") pod \"csi-node-driver-5nmmm\" (UID: \"6db03382-69c2-445f-a6ae-af82129366d4\") " pod="calico-system/csi-node-driver-5nmmm" Sep 12 22:53:13.056002 kubelet[3352]: E0912 22:53:13.055968 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.056002 kubelet[3352]: W0912 22:53:13.055996 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.056163 kubelet[3352]: E0912 22:53:13.056014 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.056163 kubelet[3352]: I0912 22:53:13.056059 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6db03382-69c2-445f-a6ae-af82129366d4-kubelet-dir\") pod \"csi-node-driver-5nmmm\" (UID: \"6db03382-69c2-445f-a6ae-af82129366d4\") " pod="calico-system/csi-node-driver-5nmmm" Sep 12 22:53:13.056524 containerd[2017]: time="2025-09-12T22:53:13.056430679Z" level=info msg="connecting to shim df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124" address="unix:///run/containerd/s/5a0c2758cbdbb6b2fd4752bbbf45e9c519c1ef7f84702b1111f31349a715c408" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:13.056825 kubelet[3352]: E0912 22:53:13.056639 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.056921 kubelet[3352]: W0912 22:53:13.056902 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.059286 kubelet[3352]: E0912 22:53:13.059090 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.059286 kubelet[3352]: I0912 22:53:13.059143 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6db03382-69c2-445f-a6ae-af82129366d4-socket-dir\") pod \"csi-node-driver-5nmmm\" (UID: \"6db03382-69c2-445f-a6ae-af82129366d4\") " pod="calico-system/csi-node-driver-5nmmm" Sep 12 22:53:13.064582 kubelet[3352]: E0912 22:53:13.063126 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.064582 kubelet[3352]: W0912 22:53:13.063150 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.064582 kubelet[3352]: E0912 22:53:13.063169 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.064582 kubelet[3352]: E0912 22:53:13.063434 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.064582 kubelet[3352]: W0912 22:53:13.063445 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.064582 kubelet[3352]: E0912 22:53:13.063483 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.064582 kubelet[3352]: E0912 22:53:13.063682 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.064582 kubelet[3352]: W0912 22:53:13.063702 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.064582 kubelet[3352]: E0912 22:53:13.063713 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.064582 kubelet[3352]: E0912 22:53:13.063890 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.065293 kubelet[3352]: W0912 22:53:13.063899 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.065293 kubelet[3352]: E0912 22:53:13.063909 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.065293 kubelet[3352]: E0912 22:53:13.064066 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.065293 kubelet[3352]: W0912 22:53:13.064075 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.065293 kubelet[3352]: E0912 22:53:13.064086 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.065293 kubelet[3352]: I0912 22:53:13.064126 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/6db03382-69c2-445f-a6ae-af82129366d4-varrun\") pod \"csi-node-driver-5nmmm\" (UID: \"6db03382-69c2-445f-a6ae-af82129366d4\") " pod="calico-system/csi-node-driver-5nmmm" Sep 12 22:53:13.065293 kubelet[3352]: E0912 22:53:13.064373 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.065293 kubelet[3352]: W0912 22:53:13.064387 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.065293 kubelet[3352]: E0912 22:53:13.064401 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.065675 kubelet[3352]: I0912 22:53:13.064427 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6db03382-69c2-445f-a6ae-af82129366d4-registration-dir\") pod \"csi-node-driver-5nmmm\" (UID: \"6db03382-69c2-445f-a6ae-af82129366d4\") " pod="calico-system/csi-node-driver-5nmmm" Sep 12 22:53:13.066683 kubelet[3352]: E0912 22:53:13.065867 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.066683 kubelet[3352]: W0912 22:53:13.065885 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.066683 kubelet[3352]: E0912 22:53:13.065901 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.066683 kubelet[3352]: E0912 22:53:13.066084 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.066683 kubelet[3352]: W0912 22:53:13.066093 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.066683 kubelet[3352]: E0912 22:53:13.066105 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.066683 kubelet[3352]: E0912 22:53:13.066283 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.066683 kubelet[3352]: W0912 22:53:13.066292 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.066683 kubelet[3352]: E0912 22:53:13.066304 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.066683 kubelet[3352]: E0912 22:53:13.066527 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.068065 kubelet[3352]: W0912 22:53:13.066536 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.068065 kubelet[3352]: E0912 22:53:13.066551 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.068914 kubelet[3352]: E0912 22:53:13.068686 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.068914 kubelet[3352]: W0912 22:53:13.068711 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.068914 kubelet[3352]: E0912 22:53:13.068746 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.069184 kubelet[3352]: E0912 22:53:13.069128 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.069184 kubelet[3352]: W0912 22:53:13.069144 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.069184 kubelet[3352]: E0912 22:53:13.069157 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.122018 systemd[1]: Started cri-containerd-df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124.scope - libcontainer container df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124. Sep 12 22:53:13.166029 kubelet[3352]: E0912 22:53:13.165896 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.166488 kubelet[3352]: W0912 22:53:13.166425 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.166763 kubelet[3352]: E0912 22:53:13.166732 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.167537 kubelet[3352]: E0912 22:53:13.167499 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.167911 kubelet[3352]: W0912 22:53:13.167884 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.168108 kubelet[3352]: E0912 22:53:13.168055 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.169066 kubelet[3352]: E0912 22:53:13.169049 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.169497 kubelet[3352]: W0912 22:53:13.169175 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.169627 kubelet[3352]: E0912 22:53:13.169608 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.170151 kubelet[3352]: E0912 22:53:13.170122 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.170362 kubelet[3352]: W0912 22:53:13.170345 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.170688 kubelet[3352]: E0912 22:53:13.170576 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.171877 kubelet[3352]: E0912 22:53:13.171861 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.172173 kubelet[3352]: W0912 22:53:13.171960 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.172173 kubelet[3352]: E0912 22:53:13.171982 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.173074 kubelet[3352]: E0912 22:53:13.172937 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.173074 kubelet[3352]: W0912 22:53:13.172960 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.173417 kubelet[3352]: E0912 22:53:13.173305 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.174971 kubelet[3352]: E0912 22:53:13.174941 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.175528 kubelet[3352]: W0912 22:53:13.175509 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.175843 kubelet[3352]: E0912 22:53:13.175717 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.176641 kubelet[3352]: E0912 22:53:13.176540 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.177012 kubelet[3352]: W0912 22:53:13.176780 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.177012 kubelet[3352]: E0912 22:53:13.176989 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.177699 kubelet[3352]: E0912 22:53:13.177685 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.178087 kubelet[3352]: W0912 22:53:13.177862 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.178087 kubelet[3352]: E0912 22:53:13.177884 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.179147 kubelet[3352]: E0912 22:53:13.178530 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.179366 kubelet[3352]: W0912 22:53:13.179341 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.179647 kubelet[3352]: E0912 22:53:13.179486 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.180137 kubelet[3352]: E0912 22:53:13.180120 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.180507 kubelet[3352]: W0912 22:53:13.180488 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.181633 kubelet[3352]: E0912 22:53:13.181614 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.182778 kubelet[3352]: E0912 22:53:13.182618 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.182778 kubelet[3352]: W0912 22:53:13.182637 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.182778 kubelet[3352]: E0912 22:53:13.182656 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.183331 kubelet[3352]: E0912 22:53:13.183149 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.183331 kubelet[3352]: W0912 22:53:13.183164 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.183331 kubelet[3352]: E0912 22:53:13.183177 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.184881 kubelet[3352]: E0912 22:53:13.184789 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.184881 kubelet[3352]: W0912 22:53:13.184804 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.184881 kubelet[3352]: E0912 22:53:13.184819 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.185796 kubelet[3352]: E0912 22:53:13.185658 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.185796 kubelet[3352]: W0912 22:53:13.185678 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.185796 kubelet[3352]: E0912 22:53:13.185696 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.187203 kubelet[3352]: E0912 22:53:13.187013 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.187203 kubelet[3352]: W0912 22:53:13.187027 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.187203 kubelet[3352]: E0912 22:53:13.187042 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.188769 kubelet[3352]: E0912 22:53:13.188746 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.188922 kubelet[3352]: W0912 22:53:13.188865 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.188922 kubelet[3352]: E0912 22:53:13.188890 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.189313 kubelet[3352]: E0912 22:53:13.189301 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.189382 kubelet[3352]: W0912 22:53:13.189369 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.189594 kubelet[3352]: E0912 22:53:13.189511 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.189928 kubelet[3352]: E0912 22:53:13.189880 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.189928 kubelet[3352]: W0912 22:53:13.189899 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.189928 kubelet[3352]: E0912 22:53:13.189913 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.190436 kubelet[3352]: E0912 22:53:13.190421 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.190436 kubelet[3352]: W0912 22:53:13.190436 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.190978 kubelet[3352]: E0912 22:53:13.190529 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.193046 kubelet[3352]: E0912 22:53:13.193018 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.193046 kubelet[3352]: W0912 22:53:13.193045 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.193216 kubelet[3352]: E0912 22:53:13.193067 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.193437 kubelet[3352]: E0912 22:53:13.193418 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.193528 kubelet[3352]: W0912 22:53:13.193437 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.193528 kubelet[3352]: E0912 22:53:13.193492 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.194685 kubelet[3352]: E0912 22:53:13.194658 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.194815 kubelet[3352]: W0912 22:53:13.194794 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.194879 kubelet[3352]: E0912 22:53:13.194821 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.195829 kubelet[3352]: E0912 22:53:13.195809 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.195829 kubelet[3352]: W0912 22:53:13.195828 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.195958 kubelet[3352]: E0912 22:53:13.195845 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.197463 kubelet[3352]: E0912 22:53:13.197356 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.197463 kubelet[3352]: W0912 22:53:13.197385 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.197463 kubelet[3352]: E0912 22:53:13.197402 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:13.203650 containerd[2017]: time="2025-09-12T22:53:13.203035951Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-87b9s,Uid:d9133c96-59ef-4269-9edf-fa9cd593c1e6,Namespace:calico-system,Attempt:0,} returns sandbox id \"df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124\"" Sep 12 22:53:13.219908 kubelet[3352]: E0912 22:53:13.219860 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:13.219908 kubelet[3352]: W0912 22:53:13.219905 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:13.220070 kubelet[3352]: E0912 22:53:13.219930 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:14.576774 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3458588494.mount: Deactivated successfully. Sep 12 22:53:14.615931 kubelet[3352]: E0912 22:53:14.615882 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5nmmm" podUID="6db03382-69c2-445f-a6ae-af82129366d4" Sep 12 22:53:16.029106 containerd[2017]: time="2025-09-12T22:53:16.028840536Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:16.031810 containerd[2017]: time="2025-09-12T22:53:16.031636407Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=35237389" Sep 12 22:53:16.040103 containerd[2017]: time="2025-09-12T22:53:16.039153387Z" level=info msg="ImageCreate event name:\"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:16.060029 containerd[2017]: time="2025-09-12T22:53:16.059968904Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:16.061578 containerd[2017]: time="2025-09-12T22:53:16.061523644Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"35237243\" in 3.11116246s" Sep 12 22:53:16.061578 containerd[2017]: time="2025-09-12T22:53:16.061577252Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\"" Sep 12 22:53:16.075518 containerd[2017]: time="2025-09-12T22:53:16.074318348Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Sep 12 22:53:16.110199 containerd[2017]: time="2025-09-12T22:53:16.107431574Z" level=info msg="CreateContainer within sandbox \"3d70fdf1728f5df7beacb3e1040212091464888ae95c689eb088b5387a1fc6d5\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 12 22:53:16.129420 containerd[2017]: time="2025-09-12T22:53:16.129370825Z" level=info msg="Container f23c72649e42553e23bb6ebea992813d8af0a57cdb63f926a2302b8096c1368a: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:16.151788 containerd[2017]: time="2025-09-12T22:53:16.151734491Z" level=info msg="CreateContainer within sandbox \"3d70fdf1728f5df7beacb3e1040212091464888ae95c689eb088b5387a1fc6d5\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"f23c72649e42553e23bb6ebea992813d8af0a57cdb63f926a2302b8096c1368a\"" Sep 12 22:53:16.154210 containerd[2017]: time="2025-09-12T22:53:16.154167601Z" level=info msg="StartContainer for \"f23c72649e42553e23bb6ebea992813d8af0a57cdb63f926a2302b8096c1368a\"" Sep 12 22:53:16.156018 containerd[2017]: time="2025-09-12T22:53:16.155975450Z" level=info msg="connecting to shim f23c72649e42553e23bb6ebea992813d8af0a57cdb63f926a2302b8096c1368a" address="unix:///run/containerd/s/583e4eb43046ffa498788e155f55329606a570289829befbc0e6c476e1ba954a" protocol=ttrpc version=3 Sep 12 22:53:16.273740 systemd[1]: Started cri-containerd-f23c72649e42553e23bb6ebea992813d8af0a57cdb63f926a2302b8096c1368a.scope - libcontainer container f23c72649e42553e23bb6ebea992813d8af0a57cdb63f926a2302b8096c1368a. Sep 12 22:53:16.453183 containerd[2017]: time="2025-09-12T22:53:16.451622278Z" level=info msg="StartContainer for \"f23c72649e42553e23bb6ebea992813d8af0a57cdb63f926a2302b8096c1368a\" returns successfully" Sep 12 22:53:16.619712 kubelet[3352]: E0912 22:53:16.619663 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5nmmm" podUID="6db03382-69c2-445f-a6ae-af82129366d4" Sep 12 22:53:16.780160 kubelet[3352]: E0912 22:53:16.780118 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.780160 kubelet[3352]: W0912 22:53:16.780153 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.780411 kubelet[3352]: E0912 22:53:16.780287 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.780946 kubelet[3352]: E0912 22:53:16.780920 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.780946 kubelet[3352]: W0912 22:53:16.780943 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.781504 kubelet[3352]: E0912 22:53:16.780959 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.782170 kubelet[3352]: E0912 22:53:16.782149 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.782170 kubelet[3352]: W0912 22:53:16.782169 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.782738 kubelet[3352]: E0912 22:53:16.782184 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.782738 kubelet[3352]: E0912 22:53:16.782526 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.782738 kubelet[3352]: W0912 22:53:16.782537 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.782738 kubelet[3352]: E0912 22:53:16.782572 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.783639 kubelet[3352]: E0912 22:53:16.783618 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.783639 kubelet[3352]: W0912 22:53:16.783638 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.783776 kubelet[3352]: E0912 22:53:16.783663 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.784597 kubelet[3352]: E0912 22:53:16.784576 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.784597 kubelet[3352]: W0912 22:53:16.784595 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.784769 kubelet[3352]: E0912 22:53:16.784610 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.785668 kubelet[3352]: E0912 22:53:16.785641 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.785748 kubelet[3352]: W0912 22:53:16.785673 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.785748 kubelet[3352]: E0912 22:53:16.785688 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.785955 kubelet[3352]: E0912 22:53:16.785938 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.785955 kubelet[3352]: W0912 22:53:16.785954 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.786057 kubelet[3352]: E0912 22:53:16.785967 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.786210 kubelet[3352]: E0912 22:53:16.786193 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.786210 kubelet[3352]: W0912 22:53:16.786209 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.786308 kubelet[3352]: E0912 22:53:16.786222 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.786584 kubelet[3352]: E0912 22:53:16.786564 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.786584 kubelet[3352]: W0912 22:53:16.786583 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.786700 kubelet[3352]: E0912 22:53:16.786596 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.787410 kubelet[3352]: E0912 22:53:16.787389 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.787410 kubelet[3352]: W0912 22:53:16.787409 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.787553 kubelet[3352]: E0912 22:53:16.787423 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.787764 kubelet[3352]: E0912 22:53:16.787745 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.787764 kubelet[3352]: W0912 22:53:16.787763 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.787868 kubelet[3352]: E0912 22:53:16.787776 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.788011 kubelet[3352]: E0912 22:53:16.787992 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.788011 kubelet[3352]: W0912 22:53:16.788010 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.788116 kubelet[3352]: E0912 22:53:16.788022 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.788246 kubelet[3352]: E0912 22:53:16.788227 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.788246 kubelet[3352]: W0912 22:53:16.788245 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.788353 kubelet[3352]: E0912 22:53:16.788258 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.789476 kubelet[3352]: E0912 22:53:16.788475 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.789476 kubelet[3352]: W0912 22:53:16.788487 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.789476 kubelet[3352]: E0912 22:53:16.788498 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.798946 kubelet[3352]: I0912 22:53:16.798485 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-c497bf79c-s826b" podStartSLOduration=1.673628889 podStartE2EDuration="4.797016083s" podCreationTimestamp="2025-09-12 22:53:12 +0000 UTC" firstStartedPulling="2025-09-12 22:53:12.949488963 +0000 UTC m=+22.519239122" lastFinishedPulling="2025-09-12 22:53:16.072876172 +0000 UTC m=+25.642626316" observedRunningTime="2025-09-12 22:53:16.795771255 +0000 UTC m=+26.365521414" watchObservedRunningTime="2025-09-12 22:53:16.797016083 +0000 UTC m=+26.366766251" Sep 12 22:53:16.816620 kubelet[3352]: E0912 22:53:16.816222 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.816784 kubelet[3352]: W0912 22:53:16.816637 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.816784 kubelet[3352]: E0912 22:53:16.816664 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.817977 kubelet[3352]: E0912 22:53:16.817950 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.817977 kubelet[3352]: W0912 22:53:16.817974 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.818155 kubelet[3352]: E0912 22:53:16.818013 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.818381 kubelet[3352]: E0912 22:53:16.818358 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.818483 kubelet[3352]: W0912 22:53:16.818379 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.818483 kubelet[3352]: E0912 22:53:16.818404 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.818976 kubelet[3352]: E0912 22:53:16.818952 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.818976 kubelet[3352]: W0912 22:53:16.818975 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.819102 kubelet[3352]: E0912 22:53:16.818990 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.820882 kubelet[3352]: E0912 22:53:16.820839 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.820882 kubelet[3352]: W0912 22:53:16.820881 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.821027 kubelet[3352]: E0912 22:53:16.820897 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.821663 kubelet[3352]: E0912 22:53:16.821641 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.821663 kubelet[3352]: W0912 22:53:16.821662 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.821796 kubelet[3352]: E0912 22:53:16.821677 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.822232 kubelet[3352]: E0912 22:53:16.822212 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.822232 kubelet[3352]: W0912 22:53:16.822230 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.822358 kubelet[3352]: E0912 22:53:16.822244 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.822495 kubelet[3352]: E0912 22:53:16.822478 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.822569 kubelet[3352]: W0912 22:53:16.822496 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.822569 kubelet[3352]: E0912 22:53:16.822509 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.822800 kubelet[3352]: E0912 22:53:16.822782 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.822800 kubelet[3352]: W0912 22:53:16.822800 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.822917 kubelet[3352]: E0912 22:53:16.822815 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.823069 kubelet[3352]: E0912 22:53:16.823049 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.823069 kubelet[3352]: W0912 22:53:16.823066 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.823183 kubelet[3352]: E0912 22:53:16.823081 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.823678 kubelet[3352]: E0912 22:53:16.823621 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.823678 kubelet[3352]: W0912 22:53:16.823636 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.823678 kubelet[3352]: E0912 22:53:16.823663 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.825150 kubelet[3352]: E0912 22:53:16.825066 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.825150 kubelet[3352]: W0912 22:53:16.825082 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.825150 kubelet[3352]: E0912 22:53:16.825097 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.825614 kubelet[3352]: E0912 22:53:16.825590 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.825795 kubelet[3352]: W0912 22:53:16.825696 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.825795 kubelet[3352]: E0912 22:53:16.825714 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.826647 kubelet[3352]: E0912 22:53:16.826510 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.826647 kubelet[3352]: W0912 22:53:16.826525 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.826647 kubelet[3352]: E0912 22:53:16.826542 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.827170 kubelet[3352]: E0912 22:53:16.827001 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.827170 kubelet[3352]: W0912 22:53:16.827023 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.827170 kubelet[3352]: E0912 22:53:16.827036 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.827599 kubelet[3352]: E0912 22:53:16.827528 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.827599 kubelet[3352]: W0912 22:53:16.827542 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.827599 kubelet[3352]: E0912 22:53:16.827556 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.829561 kubelet[3352]: E0912 22:53:16.829544 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.829663 kubelet[3352]: W0912 22:53:16.829650 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.829753 kubelet[3352]: E0912 22:53:16.829720 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:16.830191 kubelet[3352]: E0912 22:53:16.830091 3352 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 12 22:53:16.830334 kubelet[3352]: W0912 22:53:16.830286 3352 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 12 22:53:16.830334 kubelet[3352]: E0912 22:53:16.830308 3352 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 12 22:53:17.437865 containerd[2017]: time="2025-09-12T22:53:17.437808830Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:17.439809 containerd[2017]: time="2025-09-12T22:53:17.439765848Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4446660" Sep 12 22:53:17.441902 containerd[2017]: time="2025-09-12T22:53:17.441847565Z" level=info msg="ImageCreate event name:\"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:17.444955 containerd[2017]: time="2025-09-12T22:53:17.444892505Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:17.445542 containerd[2017]: time="2025-09-12T22:53:17.445507697Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5939323\" in 1.371122894s" Sep 12 22:53:17.445593 containerd[2017]: time="2025-09-12T22:53:17.445543664Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\"" Sep 12 22:53:17.452772 containerd[2017]: time="2025-09-12T22:53:17.452731319Z" level=info msg="CreateContainer within sandbox \"df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 12 22:53:17.469116 containerd[2017]: time="2025-09-12T22:53:17.467578559Z" level=info msg="Container 173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:17.486055 containerd[2017]: time="2025-09-12T22:53:17.486012365Z" level=info msg="CreateContainer within sandbox \"df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc\"" Sep 12 22:53:17.486796 containerd[2017]: time="2025-09-12T22:53:17.486712498Z" level=info msg="StartContainer for \"173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc\"" Sep 12 22:53:17.489392 containerd[2017]: time="2025-09-12T22:53:17.489328610Z" level=info msg="connecting to shim 173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc" address="unix:///run/containerd/s/5a0c2758cbdbb6b2fd4752bbbf45e9c519c1ef7f84702b1111f31349a715c408" protocol=ttrpc version=3 Sep 12 22:53:17.510668 systemd[1]: Started cri-containerd-173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc.scope - libcontainer container 173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc. Sep 12 22:53:17.572333 containerd[2017]: time="2025-09-12T22:53:17.572285927Z" level=info msg="StartContainer for \"173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc\" returns successfully" Sep 12 22:53:17.588922 systemd[1]: cri-containerd-173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc.scope: Deactivated successfully. Sep 12 22:53:17.593083 containerd[2017]: time="2025-09-12T22:53:17.593035120Z" level=info msg="received exit event container_id:\"173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc\" id:\"173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc\" pid:4262 exited_at:{seconds:1757717597 nanos:592695223}" Sep 12 22:53:17.593425 containerd[2017]: time="2025-09-12T22:53:17.593372421Z" level=info msg="TaskExit event in podsandbox handler container_id:\"173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc\" id:\"173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc\" pid:4262 exited_at:{seconds:1757717597 nanos:592695223}" Sep 12 22:53:17.635009 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-173eaa766c8d7a5d951ffd4558bd39c9f41a8437d802bca4a5614bc6c61036fc-rootfs.mount: Deactivated successfully. Sep 12 22:53:18.614171 kubelet[3352]: E0912 22:53:18.613801 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5nmmm" podUID="6db03382-69c2-445f-a6ae-af82129366d4" Sep 12 22:53:18.774899 containerd[2017]: time="2025-09-12T22:53:18.774608767Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Sep 12 22:53:20.625518 kubelet[3352]: E0912 22:53:20.624486 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5nmmm" podUID="6db03382-69c2-445f-a6ae-af82129366d4" Sep 12 22:53:22.615111 kubelet[3352]: E0912 22:53:22.615035 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-5nmmm" podUID="6db03382-69c2-445f-a6ae-af82129366d4" Sep 12 22:53:22.878355 containerd[2017]: time="2025-09-12T22:53:22.878207742Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:22.880123 containerd[2017]: time="2025-09-12T22:53:22.880076060Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=70440613" Sep 12 22:53:22.882695 containerd[2017]: time="2025-09-12T22:53:22.882611878Z" level=info msg="ImageCreate event name:\"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:22.888615 containerd[2017]: time="2025-09-12T22:53:22.888543935Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:22.889585 containerd[2017]: time="2025-09-12T22:53:22.889142605Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"71933316\" in 4.114466312s" Sep 12 22:53:22.889585 containerd[2017]: time="2025-09-12T22:53:22.889176056Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\"" Sep 12 22:53:22.895115 containerd[2017]: time="2025-09-12T22:53:22.895055902Z" level=info msg="CreateContainer within sandbox \"df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 12 22:53:22.915480 containerd[2017]: time="2025-09-12T22:53:22.911178379Z" level=info msg="Container dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:22.929880 containerd[2017]: time="2025-09-12T22:53:22.929835699Z" level=info msg="CreateContainer within sandbox \"df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a\"" Sep 12 22:53:22.932505 containerd[2017]: time="2025-09-12T22:53:22.930388050Z" level=info msg="StartContainer for \"dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a\"" Sep 12 22:53:22.933418 containerd[2017]: time="2025-09-12T22:53:22.933369336Z" level=info msg="connecting to shim dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a" address="unix:///run/containerd/s/5a0c2758cbdbb6b2fd4752bbbf45e9c519c1ef7f84702b1111f31349a715c408" protocol=ttrpc version=3 Sep 12 22:53:22.965726 systemd[1]: Started cri-containerd-dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a.scope - libcontainer container dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a. Sep 12 22:53:23.039246 containerd[2017]: time="2025-09-12T22:53:23.039197881Z" level=info msg="StartContainer for \"dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a\" returns successfully" Sep 12 22:53:23.758125 systemd[1]: cri-containerd-dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a.scope: Deactivated successfully. Sep 12 22:53:23.758767 systemd[1]: cri-containerd-dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a.scope: Consumed 607ms CPU time, 161.2M memory peak, 4.6M read from disk, 171.3M written to disk. Sep 12 22:53:23.891791 kubelet[3352]: I0912 22:53:23.891763 3352 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Sep 12 22:53:23.901582 containerd[2017]: time="2025-09-12T22:53:23.901531457Z" level=info msg="received exit event container_id:\"dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a\" id:\"dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a\" pid:4322 exited_at:{seconds:1757717603 nanos:901243333}" Sep 12 22:53:23.902156 containerd[2017]: time="2025-09-12T22:53:23.902119618Z" level=info msg="TaskExit event in podsandbox handler container_id:\"dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a\" id:\"dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a\" pid:4322 exited_at:{seconds:1757717603 nanos:901243333}" Sep 12 22:53:23.959351 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-dc9265d051a4220715e4959c43353ca2a94d1377841caa0848e7452e07fa8e0a-rootfs.mount: Deactivated successfully. Sep 12 22:53:23.980103 systemd[1]: Created slice kubepods-burstable-pod18f3337e_9bb9_4b9a_87ce_a92880024dc0.slice - libcontainer container kubepods-burstable-pod18f3337e_9bb9_4b9a_87ce_a92880024dc0.slice. Sep 12 22:53:24.020583 systemd[1]: Created slice kubepods-burstable-poda783c047_d3a2_4f47_9f90_b7b15f3cce29.slice - libcontainer container kubepods-burstable-poda783c047_d3a2_4f47_9f90_b7b15f3cce29.slice. Sep 12 22:53:24.043799 systemd[1]: Created slice kubepods-besteffort-podc41fa72e_0e56_4db0_8fb9_ce3d488fe85d.slice - libcontainer container kubepods-besteffort-podc41fa72e_0e56_4db0_8fb9_ce3d488fe85d.slice. Sep 12 22:53:24.049531 kubelet[3352]: E0912 22:53:24.049479 3352 reflector.go:200] "Failed to watch" err="failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:ip-172-31-29-105\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"calico-apiserver\": no relationship found between node 'ip-172-31-29-105' and this object" logger="UnhandledError" reflector="object-\"calico-apiserver\"/\"kube-root-ca.crt\"" type="*v1.ConfigMap" Sep 12 22:53:24.066896 systemd[1]: Created slice kubepods-besteffort-podc295bbb2_37a6_4298_a6f6_c7ed24a19fa7.slice - libcontainer container kubepods-besteffort-podc295bbb2_37a6_4298_a6f6_c7ed24a19fa7.slice. Sep 12 22:53:24.077077 kubelet[3352]: I0912 22:53:24.076580 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/b2baf514-f775-48c0-bd00-f9a124668792-calico-apiserver-certs\") pod \"calico-apiserver-5c7666b7bb-sbkxs\" (UID: \"b2baf514-f775-48c0-bd00-f9a124668792\") " pod="calico-apiserver/calico-apiserver-5c7666b7bb-sbkxs" Sep 12 22:53:24.079703 kubelet[3352]: I0912 22:53:24.078109 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qckxw\" (UniqueName: \"kubernetes.io/projected/b2baf514-f775-48c0-bd00-f9a124668792-kube-api-access-qckxw\") pod \"calico-apiserver-5c7666b7bb-sbkxs\" (UID: \"b2baf514-f775-48c0-bd00-f9a124668792\") " pod="calico-apiserver/calico-apiserver-5c7666b7bb-sbkxs" Sep 12 22:53:24.079703 kubelet[3352]: I0912 22:53:24.078210 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f284a142-6481-40f9-8099-8b7bffb670e7-tigera-ca-bundle\") pod \"calico-kube-controllers-74b867d4d7-w6r2c\" (UID: \"f284a142-6481-40f9-8099-8b7bffb670e7\") " pod="calico-system/calico-kube-controllers-74b867d4d7-w6r2c" Sep 12 22:53:24.079703 kubelet[3352]: I0912 22:53:24.078284 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fstm2\" (UniqueName: \"kubernetes.io/projected/f284a142-6481-40f9-8099-8b7bffb670e7-kube-api-access-fstm2\") pod \"calico-kube-controllers-74b867d4d7-w6r2c\" (UID: \"f284a142-6481-40f9-8099-8b7bffb670e7\") " pod="calico-system/calico-kube-controllers-74b867d4d7-w6r2c" Sep 12 22:53:24.079703 kubelet[3352]: I0912 22:53:24.078369 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgbj4\" (UniqueName: \"kubernetes.io/projected/18f3337e-9bb9-4b9a-87ce-a92880024dc0-kube-api-access-rgbj4\") pod \"coredns-674b8bbfcf-z7kf4\" (UID: \"18f3337e-9bb9-4b9a-87ce-a92880024dc0\") " pod="kube-system/coredns-674b8bbfcf-z7kf4" Sep 12 22:53:24.079703 kubelet[3352]: I0912 22:53:24.078394 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/229ca78f-57bf-4bb2-ac2b-38651bdb311d-goldmane-key-pair\") pod \"goldmane-54d579b49d-4265x\" (UID: \"229ca78f-57bf-4bb2-ac2b-38651bdb311d\") " pod="calico-system/goldmane-54d579b49d-4265x" Sep 12 22:53:24.079992 kubelet[3352]: I0912 22:53:24.079063 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgb2z\" (UniqueName: \"kubernetes.io/projected/229ca78f-57bf-4bb2-ac2b-38651bdb311d-kube-api-access-kgb2z\") pod \"goldmane-54d579b49d-4265x\" (UID: \"229ca78f-57bf-4bb2-ac2b-38651bdb311d\") " pod="calico-system/goldmane-54d579b49d-4265x" Sep 12 22:53:24.079992 kubelet[3352]: I0912 22:53:24.079131 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a783c047-d3a2-4f47-9f90-b7b15f3cce29-config-volume\") pod \"coredns-674b8bbfcf-rr8gx\" (UID: \"a783c047-d3a2-4f47-9f90-b7b15f3cce29\") " pod="kube-system/coredns-674b8bbfcf-rr8gx" Sep 12 22:53:24.079992 kubelet[3352]: I0912 22:53:24.079159 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/229ca78f-57bf-4bb2-ac2b-38651bdb311d-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-4265x\" (UID: \"229ca78f-57bf-4bb2-ac2b-38651bdb311d\") " pod="calico-system/goldmane-54d579b49d-4265x" Sep 12 22:53:24.079992 kubelet[3352]: I0912 22:53:24.079217 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-backend-key-pair\") pod \"whisker-5884b8d7f6-n7mvz\" (UID: \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\") " pod="calico-system/whisker-5884b8d7f6-n7mvz" Sep 12 22:53:24.079992 kubelet[3352]: I0912 22:53:24.079241 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmxk4\" (UniqueName: \"kubernetes.io/projected/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-kube-api-access-lmxk4\") pod \"whisker-5884b8d7f6-n7mvz\" (UID: \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\") " pod="calico-system/whisker-5884b8d7f6-n7mvz" Sep 12 22:53:24.080180 kubelet[3352]: I0912 22:53:24.079508 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c41fa72e-0e56-4db0-8fb9-ce3d488fe85d-calico-apiserver-certs\") pod \"calico-apiserver-5c7666b7bb-d8qbk\" (UID: \"c41fa72e-0e56-4db0-8fb9-ce3d488fe85d\") " pod="calico-apiserver/calico-apiserver-5c7666b7bb-d8qbk" Sep 12 22:53:24.080749 kubelet[3352]: I0912 22:53:24.080308 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stl89\" (UniqueName: \"kubernetes.io/projected/c41fa72e-0e56-4db0-8fb9-ce3d488fe85d-kube-api-access-stl89\") pod \"calico-apiserver-5c7666b7bb-d8qbk\" (UID: \"c41fa72e-0e56-4db0-8fb9-ce3d488fe85d\") " pod="calico-apiserver/calico-apiserver-5c7666b7bb-d8qbk" Sep 12 22:53:24.080749 kubelet[3352]: I0912 22:53:24.080389 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tslsc\" (UniqueName: \"kubernetes.io/projected/a783c047-d3a2-4f47-9f90-b7b15f3cce29-kube-api-access-tslsc\") pod \"coredns-674b8bbfcf-rr8gx\" (UID: \"a783c047-d3a2-4f47-9f90-b7b15f3cce29\") " pod="kube-system/coredns-674b8bbfcf-rr8gx" Sep 12 22:53:24.080749 kubelet[3352]: I0912 22:53:24.080492 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-ca-bundle\") pod \"whisker-5884b8d7f6-n7mvz\" (UID: \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\") " pod="calico-system/whisker-5884b8d7f6-n7mvz" Sep 12 22:53:24.080749 kubelet[3352]: I0912 22:53:24.080534 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229ca78f-57bf-4bb2-ac2b-38651bdb311d-config\") pod \"goldmane-54d579b49d-4265x\" (UID: \"229ca78f-57bf-4bb2-ac2b-38651bdb311d\") " pod="calico-system/goldmane-54d579b49d-4265x" Sep 12 22:53:24.080749 kubelet[3352]: I0912 22:53:24.080668 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/18f3337e-9bb9-4b9a-87ce-a92880024dc0-config-volume\") pod \"coredns-674b8bbfcf-z7kf4\" (UID: \"18f3337e-9bb9-4b9a-87ce-a92880024dc0\") " pod="kube-system/coredns-674b8bbfcf-z7kf4" Sep 12 22:53:24.085116 systemd[1]: Created slice kubepods-besteffort-pod229ca78f_57bf_4bb2_ac2b_38651bdb311d.slice - libcontainer container kubepods-besteffort-pod229ca78f_57bf_4bb2_ac2b_38651bdb311d.slice. Sep 12 22:53:24.102303 systemd[1]: Created slice kubepods-besteffort-podf284a142_6481_40f9_8099_8b7bffb670e7.slice - libcontainer container kubepods-besteffort-podf284a142_6481_40f9_8099_8b7bffb670e7.slice. Sep 12 22:53:24.109876 systemd[1]: Created slice kubepods-besteffort-podb2baf514_f775_48c0_bd00_f9a124668792.slice - libcontainer container kubepods-besteffort-podb2baf514_f775_48c0_bd00_f9a124668792.slice. Sep 12 22:53:24.297223 containerd[2017]: time="2025-09-12T22:53:24.296872139Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-z7kf4,Uid:18f3337e-9bb9-4b9a-87ce-a92880024dc0,Namespace:kube-system,Attempt:0,}" Sep 12 22:53:24.336814 containerd[2017]: time="2025-09-12T22:53:24.336776235Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-rr8gx,Uid:a783c047-d3a2-4f47-9f90-b7b15f3cce29,Namespace:kube-system,Attempt:0,}" Sep 12 22:53:24.376545 containerd[2017]: time="2025-09-12T22:53:24.376466344Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5884b8d7f6-n7mvz,Uid:c295bbb2-37a6-4298-a6f6-c7ed24a19fa7,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:24.397735 containerd[2017]: time="2025-09-12T22:53:24.397701416Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-4265x,Uid:229ca78f-57bf-4bb2-ac2b-38651bdb311d,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:24.413917 containerd[2017]: time="2025-09-12T22:53:24.413873858Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74b867d4d7-w6r2c,Uid:f284a142-6481-40f9-8099-8b7bffb670e7,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:24.651186 containerd[2017]: time="2025-09-12T22:53:24.650124666Z" level=error msg="Failed to destroy network for sandbox \"f78dd6d184194b97d3f0396b12bbd9169d4d4e48e41f81a8464bd033fda99cab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.660484 containerd[2017]: time="2025-09-12T22:53:24.652930454Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-z7kf4,Uid:18f3337e-9bb9-4b9a-87ce-a92880024dc0,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f78dd6d184194b97d3f0396b12bbd9169d4d4e48e41f81a8464bd033fda99cab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.660319 systemd[1]: Created slice kubepods-besteffort-pod6db03382_69c2_445f_a6ae_af82129366d4.slice - libcontainer container kubepods-besteffort-pod6db03382_69c2_445f_a6ae_af82129366d4.slice. Sep 12 22:53:24.673325 kubelet[3352]: E0912 22:53:24.673262 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f78dd6d184194b97d3f0396b12bbd9169d4d4e48e41f81a8464bd033fda99cab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.673604 kubelet[3352]: E0912 22:53:24.673352 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f78dd6d184194b97d3f0396b12bbd9169d4d4e48e41f81a8464bd033fda99cab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-z7kf4" Sep 12 22:53:24.673604 kubelet[3352]: E0912 22:53:24.673377 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f78dd6d184194b97d3f0396b12bbd9169d4d4e48e41f81a8464bd033fda99cab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-z7kf4" Sep 12 22:53:24.683484 kubelet[3352]: E0912 22:53:24.683154 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-z7kf4_kube-system(18f3337e-9bb9-4b9a-87ce-a92880024dc0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-z7kf4_kube-system(18f3337e-9bb9-4b9a-87ce-a92880024dc0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f78dd6d184194b97d3f0396b12bbd9169d4d4e48e41f81a8464bd033fda99cab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-z7kf4" podUID="18f3337e-9bb9-4b9a-87ce-a92880024dc0" Sep 12 22:53:24.696925 containerd[2017]: time="2025-09-12T22:53:24.696859388Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5nmmm,Uid:6db03382-69c2-445f-a6ae-af82129366d4,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:24.745475 containerd[2017]: time="2025-09-12T22:53:24.744166921Z" level=error msg="Failed to destroy network for sandbox \"7f1e63af44aa873dcc356f4a9346f024c4e13bc268bbbf14fe969f65390ff2d5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.749083 containerd[2017]: time="2025-09-12T22:53:24.749015873Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-rr8gx,Uid:a783c047-d3a2-4f47-9f90-b7b15f3cce29,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"7f1e63af44aa873dcc356f4a9346f024c4e13bc268bbbf14fe969f65390ff2d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.750194 kubelet[3352]: E0912 22:53:24.749429 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7f1e63af44aa873dcc356f4a9346f024c4e13bc268bbbf14fe969f65390ff2d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.750194 kubelet[3352]: E0912 22:53:24.749969 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7f1e63af44aa873dcc356f4a9346f024c4e13bc268bbbf14fe969f65390ff2d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-rr8gx" Sep 12 22:53:24.750194 kubelet[3352]: E0912 22:53:24.750022 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7f1e63af44aa873dcc356f4a9346f024c4e13bc268bbbf14fe969f65390ff2d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-rr8gx" Sep 12 22:53:24.750417 kubelet[3352]: E0912 22:53:24.750126 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-rr8gx_kube-system(a783c047-d3a2-4f47-9f90-b7b15f3cce29)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-rr8gx_kube-system(a783c047-d3a2-4f47-9f90-b7b15f3cce29)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7f1e63af44aa873dcc356f4a9346f024c4e13bc268bbbf14fe969f65390ff2d5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-rr8gx" podUID="a783c047-d3a2-4f47-9f90-b7b15f3cce29" Sep 12 22:53:24.773747 containerd[2017]: time="2025-09-12T22:53:24.773675506Z" level=error msg="Failed to destroy network for sandbox \"b5ee602bc13364c035bf321de022befc49bdcc86079023929dbca5eade7dcb77\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.777284 containerd[2017]: time="2025-09-12T22:53:24.777223425Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74b867d4d7-w6r2c,Uid:f284a142-6481-40f9-8099-8b7bffb670e7,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b5ee602bc13364c035bf321de022befc49bdcc86079023929dbca5eade7dcb77\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.778130 kubelet[3352]: E0912 22:53:24.778008 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b5ee602bc13364c035bf321de022befc49bdcc86079023929dbca5eade7dcb77\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.778130 kubelet[3352]: E0912 22:53:24.778084 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b5ee602bc13364c035bf321de022befc49bdcc86079023929dbca5eade7dcb77\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-74b867d4d7-w6r2c" Sep 12 22:53:24.778130 kubelet[3352]: E0912 22:53:24.778115 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b5ee602bc13364c035bf321de022befc49bdcc86079023929dbca5eade7dcb77\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-74b867d4d7-w6r2c" Sep 12 22:53:24.778491 kubelet[3352]: E0912 22:53:24.778183 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-74b867d4d7-w6r2c_calico-system(f284a142-6481-40f9-8099-8b7bffb670e7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-74b867d4d7-w6r2c_calico-system(f284a142-6481-40f9-8099-8b7bffb670e7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b5ee602bc13364c035bf321de022befc49bdcc86079023929dbca5eade7dcb77\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-74b867d4d7-w6r2c" podUID="f284a142-6481-40f9-8099-8b7bffb670e7" Sep 12 22:53:24.779767 containerd[2017]: time="2025-09-12T22:53:24.779721338Z" level=error msg="Failed to destroy network for sandbox \"90bf22a7bb9cf22e139fc941a8065cbd297cce6a4c2a24273171a21bb6f33b70\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.783214 containerd[2017]: time="2025-09-12T22:53:24.783159996Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-4265x,Uid:229ca78f-57bf-4bb2-ac2b-38651bdb311d,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"90bf22a7bb9cf22e139fc941a8065cbd297cce6a4c2a24273171a21bb6f33b70\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.783815 kubelet[3352]: E0912 22:53:24.783763 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"90bf22a7bb9cf22e139fc941a8065cbd297cce6a4c2a24273171a21bb6f33b70\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.783910 kubelet[3352]: E0912 22:53:24.783836 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"90bf22a7bb9cf22e139fc941a8065cbd297cce6a4c2a24273171a21bb6f33b70\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-4265x" Sep 12 22:53:24.783910 kubelet[3352]: E0912 22:53:24.783869 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"90bf22a7bb9cf22e139fc941a8065cbd297cce6a4c2a24273171a21bb6f33b70\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-4265x" Sep 12 22:53:24.784604 kubelet[3352]: E0912 22:53:24.784562 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-4265x_calico-system(229ca78f-57bf-4bb2-ac2b-38651bdb311d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-4265x_calico-system(229ca78f-57bf-4bb2-ac2b-38651bdb311d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"90bf22a7bb9cf22e139fc941a8065cbd297cce6a4c2a24273171a21bb6f33b70\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-4265x" podUID="229ca78f-57bf-4bb2-ac2b-38651bdb311d" Sep 12 22:53:24.786328 containerd[2017]: time="2025-09-12T22:53:24.786199324Z" level=error msg="Failed to destroy network for sandbox \"07f1ca6a951857535a35ffa204dfcf2efc7f8e96717be586a1e41a518b8035a0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.790095 containerd[2017]: time="2025-09-12T22:53:24.789877682Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-5884b8d7f6-n7mvz,Uid:c295bbb2-37a6-4298-a6f6-c7ed24a19fa7,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"07f1ca6a951857535a35ffa204dfcf2efc7f8e96717be586a1e41a518b8035a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.791141 kubelet[3352]: E0912 22:53:24.791092 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07f1ca6a951857535a35ffa204dfcf2efc7f8e96717be586a1e41a518b8035a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.791242 kubelet[3352]: E0912 22:53:24.791159 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07f1ca6a951857535a35ffa204dfcf2efc7f8e96717be586a1e41a518b8035a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5884b8d7f6-n7mvz" Sep 12 22:53:24.791242 kubelet[3352]: E0912 22:53:24.791186 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07f1ca6a951857535a35ffa204dfcf2efc7f8e96717be586a1e41a518b8035a0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-5884b8d7f6-n7mvz" Sep 12 22:53:24.791827 kubelet[3352]: E0912 22:53:24.791303 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-5884b8d7f6-n7mvz_calico-system(c295bbb2-37a6-4298-a6f6-c7ed24a19fa7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-5884b8d7f6-n7mvz_calico-system(c295bbb2-37a6-4298-a6f6-c7ed24a19fa7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"07f1ca6a951857535a35ffa204dfcf2efc7f8e96717be586a1e41a518b8035a0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-5884b8d7f6-n7mvz" podUID="c295bbb2-37a6-4298-a6f6-c7ed24a19fa7" Sep 12 22:53:24.826670 containerd[2017]: time="2025-09-12T22:53:24.826525950Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Sep 12 22:53:24.874648 containerd[2017]: time="2025-09-12T22:53:24.874590466Z" level=error msg="Failed to destroy network for sandbox \"45a54ead8690e028d87606d85865cb591ff82777467f85e5455dc2973aaac692\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.877395 containerd[2017]: time="2025-09-12T22:53:24.877331250Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5nmmm,Uid:6db03382-69c2-445f-a6ae-af82129366d4,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"45a54ead8690e028d87606d85865cb591ff82777467f85e5455dc2973aaac692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.877831 kubelet[3352]: E0912 22:53:24.877648 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45a54ead8690e028d87606d85865cb591ff82777467f85e5455dc2973aaac692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:24.877831 kubelet[3352]: E0912 22:53:24.877718 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45a54ead8690e028d87606d85865cb591ff82777467f85e5455dc2973aaac692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5nmmm" Sep 12 22:53:24.877831 kubelet[3352]: E0912 22:53:24.877744 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45a54ead8690e028d87606d85865cb591ff82777467f85e5455dc2973aaac692\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-5nmmm" Sep 12 22:53:24.877972 kubelet[3352]: E0912 22:53:24.877805 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-5nmmm_calico-system(6db03382-69c2-445f-a6ae-af82129366d4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-5nmmm_calico-system(6db03382-69c2-445f-a6ae-af82129366d4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"45a54ead8690e028d87606d85865cb591ff82777467f85e5455dc2973aaac692\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-5nmmm" podUID="6db03382-69c2-445f-a6ae-af82129366d4" Sep 12 22:53:25.262425 containerd[2017]: time="2025-09-12T22:53:25.262289628Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-d8qbk,Uid:c41fa72e-0e56-4db0-8fb9-ce3d488fe85d,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:53:25.315945 containerd[2017]: time="2025-09-12T22:53:25.315899125Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-sbkxs,Uid:b2baf514-f775-48c0-bd00-f9a124668792,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:53:25.325620 containerd[2017]: time="2025-09-12T22:53:25.325560961Z" level=error msg="Failed to destroy network for sandbox \"2234e4d3cf2da4b72bc177081141f1ec2f2d77e1ce28be3748f5a093bd296074\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:25.328493 containerd[2017]: time="2025-09-12T22:53:25.328145563Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-d8qbk,Uid:c41fa72e-0e56-4db0-8fb9-ce3d488fe85d,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"2234e4d3cf2da4b72bc177081141f1ec2f2d77e1ce28be3748f5a093bd296074\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:25.329683 kubelet[3352]: E0912 22:53:25.328750 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2234e4d3cf2da4b72bc177081141f1ec2f2d77e1ce28be3748f5a093bd296074\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:25.329683 kubelet[3352]: E0912 22:53:25.328810 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2234e4d3cf2da4b72bc177081141f1ec2f2d77e1ce28be3748f5a093bd296074\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7666b7bb-d8qbk" Sep 12 22:53:25.329683 kubelet[3352]: E0912 22:53:25.328830 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2234e4d3cf2da4b72bc177081141f1ec2f2d77e1ce28be3748f5a093bd296074\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7666b7bb-d8qbk" Sep 12 22:53:25.330063 kubelet[3352]: E0912 22:53:25.328878 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5c7666b7bb-d8qbk_calico-apiserver(c41fa72e-0e56-4db0-8fb9-ce3d488fe85d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5c7666b7bb-d8qbk_calico-apiserver(c41fa72e-0e56-4db0-8fb9-ce3d488fe85d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2234e4d3cf2da4b72bc177081141f1ec2f2d77e1ce28be3748f5a093bd296074\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5c7666b7bb-d8qbk" podUID="c41fa72e-0e56-4db0-8fb9-ce3d488fe85d" Sep 12 22:53:25.385575 containerd[2017]: time="2025-09-12T22:53:25.385515457Z" level=error msg="Failed to destroy network for sandbox \"81ce710fce5a08ee789120ee9eff6bf0ed74d6512578fdde1839e4948fd973ab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:25.387889 containerd[2017]: time="2025-09-12T22:53:25.387829708Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-sbkxs,Uid:b2baf514-f775-48c0-bd00-f9a124668792,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"81ce710fce5a08ee789120ee9eff6bf0ed74d6512578fdde1839e4948fd973ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:25.388568 kubelet[3352]: E0912 22:53:25.388082 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81ce710fce5a08ee789120ee9eff6bf0ed74d6512578fdde1839e4948fd973ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 12 22:53:25.388568 kubelet[3352]: E0912 22:53:25.388147 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81ce710fce5a08ee789120ee9eff6bf0ed74d6512578fdde1839e4948fd973ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7666b7bb-sbkxs" Sep 12 22:53:25.388568 kubelet[3352]: E0912 22:53:25.388168 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81ce710fce5a08ee789120ee9eff6bf0ed74d6512578fdde1839e4948fd973ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-5c7666b7bb-sbkxs" Sep 12 22:53:25.388686 kubelet[3352]: E0912 22:53:25.388222 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-5c7666b7bb-sbkxs_calico-apiserver(b2baf514-f775-48c0-bd00-f9a124668792)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-5c7666b7bb-sbkxs_calico-apiserver(b2baf514-f775-48c0-bd00-f9a124668792)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"81ce710fce5a08ee789120ee9eff6bf0ed74d6512578fdde1839e4948fd973ab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-5c7666b7bb-sbkxs" podUID="b2baf514-f775-48c0-bd00-f9a124668792" Sep 12 22:53:25.958194 systemd[1]: run-netns-cni\x2d5e92c050\x2d4f64\x2d0afb\x2db688\x2dabb1a9415b4a.mount: Deactivated successfully. Sep 12 22:53:25.958301 systemd[1]: run-netns-cni\x2d539ed844\x2dc73e\x2d5664\x2d137c\x2d1a7327c22532.mount: Deactivated successfully. Sep 12 22:53:32.849138 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1846791822.mount: Deactivated successfully. Sep 12 22:53:32.908083 containerd[2017]: time="2025-09-12T22:53:32.907986756Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:32.914058 containerd[2017]: time="2025-09-12T22:53:32.913997665Z" level=info msg="ImageCreate event name:\"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:32.919857 containerd[2017]: time="2025-09-12T22:53:32.918751819Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=157078339" Sep 12 22:53:32.919857 containerd[2017]: time="2025-09-12T22:53:32.918960369Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:32.922928 containerd[2017]: time="2025-09-12T22:53:32.922853682Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"157078201\" in 8.093107779s" Sep 12 22:53:32.922928 containerd[2017]: time="2025-09-12T22:53:32.922928927Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\"" Sep 12 22:53:32.952584 containerd[2017]: time="2025-09-12T22:53:32.952526289Z" level=info msg="CreateContainer within sandbox \"df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 12 22:53:32.997839 containerd[2017]: time="2025-09-12T22:53:32.997637577Z" level=info msg="Container 88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:33.000900 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2038485665.mount: Deactivated successfully. Sep 12 22:53:33.038573 containerd[2017]: time="2025-09-12T22:53:33.038502992Z" level=info msg="CreateContainer within sandbox \"df9c9d17c11c57cd4bc37e96c760b1ff4dfd914d40c670975c29a03615582124\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\"" Sep 12 22:53:33.039720 containerd[2017]: time="2025-09-12T22:53:33.039671580Z" level=info msg="StartContainer for \"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\"" Sep 12 22:53:33.051818 containerd[2017]: time="2025-09-12T22:53:33.051765854Z" level=info msg="connecting to shim 88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0" address="unix:///run/containerd/s/5a0c2758cbdbb6b2fd4752bbbf45e9c519c1ef7f84702b1111f31349a715c408" protocol=ttrpc version=3 Sep 12 22:53:33.193701 systemd[1]: Started cri-containerd-88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0.scope - libcontainer container 88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0. Sep 12 22:53:33.284599 containerd[2017]: time="2025-09-12T22:53:33.283721899Z" level=info msg="StartContainer for \"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\" returns successfully" Sep 12 22:53:33.418257 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 12 22:53:33.419939 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 12 22:53:33.858942 kubelet[3352]: I0912 22:53:33.858884 3352 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-ca-bundle\") pod \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\" (UID: \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\") " Sep 12 22:53:33.860281 kubelet[3352]: I0912 22:53:33.860088 3352 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-backend-key-pair\") pod \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\" (UID: \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\") " Sep 12 22:53:33.860281 kubelet[3352]: I0912 22:53:33.860143 3352 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmxk4\" (UniqueName: \"kubernetes.io/projected/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-kube-api-access-lmxk4\") pod \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\" (UID: \"c295bbb2-37a6-4298-a6f6-c7ed24a19fa7\") " Sep 12 22:53:33.877044 kubelet[3352]: I0912 22:53:33.876925 3352 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "c295bbb2-37a6-4298-a6f6-c7ed24a19fa7" (UID: "c295bbb2-37a6-4298-a6f6-c7ed24a19fa7"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Sep 12 22:53:33.882898 kubelet[3352]: I0912 22:53:33.882774 3352 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-kube-api-access-lmxk4" (OuterVolumeSpecName: "kube-api-access-lmxk4") pod "c295bbb2-37a6-4298-a6f6-c7ed24a19fa7" (UID: "c295bbb2-37a6-4298-a6f6-c7ed24a19fa7"). InnerVolumeSpecName "kube-api-access-lmxk4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Sep 12 22:53:33.885505 systemd[1]: var-lib-kubelet-pods-c295bbb2\x2d37a6\x2d4298\x2da6f6\x2dc7ed24a19fa7-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dlmxk4.mount: Deactivated successfully. Sep 12 22:53:33.887493 kubelet[3352]: I0912 22:53:33.886734 3352 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "c295bbb2-37a6-4298-a6f6-c7ed24a19fa7" (UID: "c295bbb2-37a6-4298-a6f6-c7ed24a19fa7"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Sep 12 22:53:33.895152 systemd[1]: var-lib-kubelet-pods-c295bbb2\x2d37a6\x2d4298\x2da6f6\x2dc7ed24a19fa7-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Sep 12 22:53:33.960743 kubelet[3352]: I0912 22:53:33.960497 3352 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lmxk4\" (UniqueName: \"kubernetes.io/projected/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-kube-api-access-lmxk4\") on node \"ip-172-31-29-105\" DevicePath \"\"" Sep 12 22:53:33.960743 kubelet[3352]: I0912 22:53:33.960529 3352 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-ca-bundle\") on node \"ip-172-31-29-105\" DevicePath \"\"" Sep 12 22:53:33.960743 kubelet[3352]: I0912 22:53:33.960546 3352 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7-whisker-backend-key-pair\") on node \"ip-172-31-29-105\" DevicePath \"\"" Sep 12 22:53:33.962147 systemd[1]: Removed slice kubepods-besteffort-podc295bbb2_37a6_4298_a6f6_c7ed24a19fa7.slice - libcontainer container kubepods-besteffort-podc295bbb2_37a6_4298_a6f6_c7ed24a19fa7.slice. Sep 12 22:53:33.993573 kubelet[3352]: I0912 22:53:33.988977 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-87b9s" podStartSLOduration=2.279601923 podStartE2EDuration="21.988947605s" podCreationTimestamp="2025-09-12 22:53:12 +0000 UTC" firstStartedPulling="2025-09-12 22:53:13.214875798 +0000 UTC m=+22.784625959" lastFinishedPulling="2025-09-12 22:53:32.924221496 +0000 UTC m=+42.493971641" observedRunningTime="2025-09-12 22:53:33.985712287 +0000 UTC m=+43.555462455" watchObservedRunningTime="2025-09-12 22:53:33.988947605 +0000 UTC m=+43.558697768" Sep 12 22:53:34.169089 systemd[1]: Created slice kubepods-besteffort-pode3491158_089c_4a3b_b6fd_ec1d5c9eac23.slice - libcontainer container kubepods-besteffort-pode3491158_089c_4a3b_b6fd_ec1d5c9eac23.slice. Sep 12 22:53:34.265009 kubelet[3352]: I0912 22:53:34.264910 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3491158-089c-4a3b-b6fd-ec1d5c9eac23-whisker-ca-bundle\") pod \"whisker-55bb7dd67f-h76qr\" (UID: \"e3491158-089c-4a3b-b6fd-ec1d5c9eac23\") " pod="calico-system/whisker-55bb7dd67f-h76qr" Sep 12 22:53:34.265206 kubelet[3352]: I0912 22:53:34.265060 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e3491158-089c-4a3b-b6fd-ec1d5c9eac23-whisker-backend-key-pair\") pod \"whisker-55bb7dd67f-h76qr\" (UID: \"e3491158-089c-4a3b-b6fd-ec1d5c9eac23\") " pod="calico-system/whisker-55bb7dd67f-h76qr" Sep 12 22:53:34.265206 kubelet[3352]: I0912 22:53:34.265086 3352 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l875x\" (UniqueName: \"kubernetes.io/projected/e3491158-089c-4a3b-b6fd-ec1d5c9eac23-kube-api-access-l875x\") pod \"whisker-55bb7dd67f-h76qr\" (UID: \"e3491158-089c-4a3b-b6fd-ec1d5c9eac23\") " pod="calico-system/whisker-55bb7dd67f-h76qr" Sep 12 22:53:34.311880 containerd[2017]: time="2025-09-12T22:53:34.311785422Z" level=info msg="TaskExit event in podsandbox handler container_id:\"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\" id:\"75bc585f3cc3713ff7841c5fc77c96c188597285b179506f7dda53f5e36f3b8b\" pid:4667 exit_status:1 exited_at:{seconds:1757717614 nanos:311296278}" Sep 12 22:53:34.490394 containerd[2017]: time="2025-09-12T22:53:34.490350565Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-55bb7dd67f-h76qr,Uid:e3491158-089c-4a3b-b6fd-ec1d5c9eac23,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:34.617167 kubelet[3352]: I0912 22:53:34.617119 3352 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c295bbb2-37a6-4298-a6f6-c7ed24a19fa7" path="/var/lib/kubelet/pods/c295bbb2-37a6-4298-a6f6-c7ed24a19fa7/volumes" Sep 12 22:53:35.078611 (udev-worker)[4622]: Network interface NamePolicy= disabled on kernel command line. Sep 12 22:53:35.082923 systemd-networkd[1897]: calib3dca84709b: Link UP Sep 12 22:53:35.084058 systemd-networkd[1897]: calib3dca84709b: Gained carrier Sep 12 22:53:35.096164 containerd[2017]: time="2025-09-12T22:53:35.096100258Z" level=info msg="TaskExit event in podsandbox handler container_id:\"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\" id:\"af4d35789dfd014a1c49565343e15eb6082a26ebce91ddbc7ceae95c34a86187\" pid:4712 exit_status:1 exited_at:{seconds:1757717615 nanos:95709551}" Sep 12 22:53:35.112263 containerd[2017]: 2025-09-12 22:53:34.543 [INFO][4685] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 12 22:53:35.112263 containerd[2017]: 2025-09-12 22:53:34.600 [INFO][4685] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0 whisker-55bb7dd67f- calico-system e3491158-089c-4a3b-b6fd-ec1d5c9eac23 939 0 2025-09-12 22:53:34 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:55bb7dd67f projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ip-172-31-29-105 whisker-55bb7dd67f-h76qr eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calib3dca84709b [] [] }} ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-" Sep 12 22:53:35.112263 containerd[2017]: 2025-09-12 22:53:34.600 [INFO][4685] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" Sep 12 22:53:35.112263 containerd[2017]: 2025-09-12 22:53:34.977 [INFO][4694] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" HandleID="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Workload="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:34.982 [INFO][4694] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" HandleID="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Workload="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003b78e0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-29-105", "pod":"whisker-55bb7dd67f-h76qr", "timestamp":"2025-09-12 22:53:34.977715515 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:34.982 [INFO][4694] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:34.983 [INFO][4694] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:34.984 [INFO][4694] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:35.002 [INFO][4694] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" host="ip-172-31-29-105" Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:35.018 [INFO][4694] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:35.027 [INFO][4694] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:35.030 [INFO][4694] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:35.113576 containerd[2017]: 2025-09-12 22:53:35.033 [INFO][4694] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:35.113992 containerd[2017]: 2025-09-12 22:53:35.033 [INFO][4694] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" host="ip-172-31-29-105" Sep 12 22:53:35.113992 containerd[2017]: 2025-09-12 22:53:35.035 [INFO][4694] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752 Sep 12 22:53:35.113992 containerd[2017]: 2025-09-12 22:53:35.046 [INFO][4694] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" host="ip-172-31-29-105" Sep 12 22:53:35.113992 containerd[2017]: 2025-09-12 22:53:35.055 [INFO][4694] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.129/26] block=192.168.125.128/26 handle="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" host="ip-172-31-29-105" Sep 12 22:53:35.113992 containerd[2017]: 2025-09-12 22:53:35.056 [INFO][4694] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.129/26] handle="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" host="ip-172-31-29-105" Sep 12 22:53:35.113992 containerd[2017]: 2025-09-12 22:53:35.056 [INFO][4694] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:35.113992 containerd[2017]: 2025-09-12 22:53:35.056 [INFO][4694] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.129/26] IPv6=[] ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" HandleID="k8s-pod-network.42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Workload="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" Sep 12 22:53:35.117219 containerd[2017]: 2025-09-12 22:53:35.060 [INFO][4685] cni-plugin/k8s.go 418: Populated endpoint ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0", GenerateName:"whisker-55bb7dd67f-", Namespace:"calico-system", SelfLink:"", UID:"e3491158-089c-4a3b-b6fd-ec1d5c9eac23", ResourceVersion:"939", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"55bb7dd67f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"whisker-55bb7dd67f-h76qr", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.125.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calib3dca84709b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:35.117219 containerd[2017]: 2025-09-12 22:53:35.060 [INFO][4685] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.129/32] ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" Sep 12 22:53:35.117388 containerd[2017]: 2025-09-12 22:53:35.061 [INFO][4685] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib3dca84709b ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" Sep 12 22:53:35.117388 containerd[2017]: 2025-09-12 22:53:35.081 [INFO][4685] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" Sep 12 22:53:35.118349 containerd[2017]: 2025-09-12 22:53:35.082 [INFO][4685] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0", GenerateName:"whisker-55bb7dd67f-", Namespace:"calico-system", SelfLink:"", UID:"e3491158-089c-4a3b-b6fd-ec1d5c9eac23", ResourceVersion:"939", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"55bb7dd67f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752", Pod:"whisker-55bb7dd67f-h76qr", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.125.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calib3dca84709b", MAC:"ea:7f:aa:1e:20:9f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:35.118542 containerd[2017]: 2025-09-12 22:53:35.104 [INFO][4685] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" Namespace="calico-system" Pod="whisker-55bb7dd67f-h76qr" WorkloadEndpoint="ip--172--31--29--105-k8s-whisker--55bb7dd67f--h76qr-eth0" Sep 12 22:53:35.339845 containerd[2017]: time="2025-09-12T22:53:35.339722393Z" level=info msg="connecting to shim 42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752" address="unix:///run/containerd/s/9ff71b9eebcf79e0d2f744403665595ba286313db55d33808b7259670a0ab9dc" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:35.416733 systemd[1]: Started cri-containerd-42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752.scope - libcontainer container 42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752. Sep 12 22:53:35.615946 containerd[2017]: time="2025-09-12T22:53:35.615536568Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-z7kf4,Uid:18f3337e-9bb9-4b9a-87ce-a92880024dc0,Namespace:kube-system,Attempt:0,}" Sep 12 22:53:35.618421 containerd[2017]: time="2025-09-12T22:53:35.618144541Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-rr8gx,Uid:a783c047-d3a2-4f47-9f90-b7b15f3cce29,Namespace:kube-system,Attempt:0,}" Sep 12 22:53:35.621231 containerd[2017]: time="2025-09-12T22:53:35.621071502Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-55bb7dd67f-h76qr,Uid:e3491158-089c-4a3b-b6fd-ec1d5c9eac23,Namespace:calico-system,Attempt:0,} returns sandbox id \"42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752\"" Sep 12 22:53:35.657496 containerd[2017]: time="2025-09-12T22:53:35.654860361Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Sep 12 22:53:36.016149 systemd-networkd[1897]: cali61c56c1997a: Link UP Sep 12 22:53:36.016728 systemd-networkd[1897]: cali61c56c1997a: Gained carrier Sep 12 22:53:36.065994 containerd[2017]: 2025-09-12 22:53:35.745 [INFO][4868] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Sep 12 22:53:36.065994 containerd[2017]: 2025-09-12 22:53:35.792 [INFO][4868] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0 coredns-674b8bbfcf- kube-system 18f3337e-9bb9-4b9a-87ce-a92880024dc0 860 0 2025-09-12 22:52:55 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-29-105 coredns-674b8bbfcf-z7kf4 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali61c56c1997a [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-" Sep 12 22:53:36.065994 containerd[2017]: 2025-09-12 22:53:35.792 [INFO][4868] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" Sep 12 22:53:36.065994 containerd[2017]: 2025-09-12 22:53:35.844 [INFO][4897] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" HandleID="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.845 [INFO][4897] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" HandleID="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5600), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-29-105", "pod":"coredns-674b8bbfcf-z7kf4", "timestamp":"2025-09-12 22:53:35.844850226 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.845 [INFO][4897] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.845 [INFO][4897] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.845 [INFO][4897] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.858 [INFO][4897] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" host="ip-172-31-29-105" Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.874 [INFO][4897] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.891 [INFO][4897] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.896 [INFO][4897] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:36.067709 containerd[2017]: 2025-09-12 22:53:35.903 [INFO][4897] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:36.069333 containerd[2017]: 2025-09-12 22:53:35.907 [INFO][4897] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" host="ip-172-31-29-105" Sep 12 22:53:36.069333 containerd[2017]: 2025-09-12 22:53:35.916 [INFO][4897] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc Sep 12 22:53:36.069333 containerd[2017]: 2025-09-12 22:53:35.927 [INFO][4897] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" host="ip-172-31-29-105" Sep 12 22:53:36.069333 containerd[2017]: 2025-09-12 22:53:35.966 [INFO][4897] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.130/26] block=192.168.125.128/26 handle="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" host="ip-172-31-29-105" Sep 12 22:53:36.069333 containerd[2017]: 2025-09-12 22:53:35.969 [INFO][4897] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.130/26] handle="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" host="ip-172-31-29-105" Sep 12 22:53:36.069333 containerd[2017]: 2025-09-12 22:53:35.976 [INFO][4897] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:36.069333 containerd[2017]: 2025-09-12 22:53:35.976 [INFO][4897] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.130/26] IPv6=[] ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" HandleID="k8s-pod-network.30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" Sep 12 22:53:36.070028 containerd[2017]: 2025-09-12 22:53:35.996 [INFO][4868] cni-plugin/k8s.go 418: Populated endpoint ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"18f3337e-9bb9-4b9a-87ce-a92880024dc0", ResourceVersion:"860", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 52, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"coredns-674b8bbfcf-z7kf4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.125.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali61c56c1997a", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:36.070028 containerd[2017]: 2025-09-12 22:53:35.997 [INFO][4868] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.130/32] ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" Sep 12 22:53:36.070028 containerd[2017]: 2025-09-12 22:53:35.997 [INFO][4868] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali61c56c1997a ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" Sep 12 22:53:36.070028 containerd[2017]: 2025-09-12 22:53:36.020 [INFO][4868] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" Sep 12 22:53:36.070028 containerd[2017]: 2025-09-12 22:53:36.020 [INFO][4868] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"18f3337e-9bb9-4b9a-87ce-a92880024dc0", ResourceVersion:"860", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 52, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc", Pod:"coredns-674b8bbfcf-z7kf4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.125.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali61c56c1997a", MAC:"be:8d:31:40:ba:75", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:36.070028 containerd[2017]: 2025-09-12 22:53:36.056 [INFO][4868] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" Namespace="kube-system" Pod="coredns-674b8bbfcf-z7kf4" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--z7kf4-eth0" Sep 12 22:53:36.186989 containerd[2017]: time="2025-09-12T22:53:36.186892819Z" level=info msg="connecting to shim 30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc" address="unix:///run/containerd/s/5502f1411b5eabe18e9c155d32e8fc6f0275c47401e4b6338f145721cc0c9a75" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:36.268725 systemd[1]: Started cri-containerd-30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc.scope - libcontainer container 30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc. Sep 12 22:53:36.295714 systemd-networkd[1897]: calib3dca84709b: Gained IPv6LL Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.326 [INFO][4927] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.328 [INFO][4927] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" iface="eth0" netns="/var/run/netns/cni-f72d6466-5d63-21e5-c931-67617c753e8b" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.329 [INFO][4927] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" iface="eth0" netns="/var/run/netns/cni-f72d6466-5d63-21e5-c931-67617c753e8b" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.330 [INFO][4927] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" iface="eth0" netns="/var/run/netns/cni-f72d6466-5d63-21e5-c931-67617c753e8b" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.330 [INFO][4927] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.330 [INFO][4927] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.380 [INFO][4997] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" HandleID="k8s-pod-network.eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.381 [INFO][4997] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.381 [INFO][4997] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.409 [WARNING][4997] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" HandleID="k8s-pod-network.eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.409 [INFO][4997] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" HandleID="k8s-pod-network.eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.415 [INFO][4997] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:36.433863 containerd[2017]: 2025-09-12 22:53:36.425 [INFO][4927] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311" Sep 12 22:53:36.442051 systemd[1]: run-netns-cni\x2df72d6466\x2d5d63\x2d21e5\x2dc931\x2d67617c753e8b.mount: Deactivated successfully. Sep 12 22:53:36.444247 containerd[2017]: time="2025-09-12T22:53:36.441442845Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-z7kf4,Uid:18f3337e-9bb9-4b9a-87ce-a92880024dc0,Namespace:kube-system,Attempt:0,} returns sandbox id \"30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc\"" Sep 12 22:53:36.450904 containerd[2017]: time="2025-09-12T22:53:36.448652810Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-rr8gx,Uid:a783c047-d3a2-4f47-9f90-b7b15f3cce29,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311\": plugin type=\"calico\" failed (add): Get \"https://10.96.0.1:443/api/v1/namespaces/kube-system\": tls: failed to verify certificate: x509: certificate signed by unknown authority" Sep 12 22:53:36.451227 kubelet[3352]: E0912 22:53:36.451046 3352 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311\": plugin type=\"calico\" failed (add): Get \"https://10.96.0.1:443/api/v1/namespaces/kube-system\": tls: failed to verify certificate: x509: certificate signed by unknown authority" Sep 12 22:53:36.451227 kubelet[3352]: E0912 22:53:36.451121 3352 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311\": plugin type=\"calico\" failed (add): Get \"https://10.96.0.1:443/api/v1/namespaces/kube-system\": tls: failed to verify certificate: x509: certificate signed by unknown authority" pod="kube-system/coredns-674b8bbfcf-rr8gx" Sep 12 22:53:36.451227 kubelet[3352]: E0912 22:53:36.451151 3352 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311\": plugin type=\"calico\" failed (add): Get \"https://10.96.0.1:443/api/v1/namespaces/kube-system\": tls: failed to verify certificate: x509: certificate signed by unknown authority" pod="kube-system/coredns-674b8bbfcf-rr8gx" Sep 12 22:53:36.452914 kubelet[3352]: E0912 22:53:36.451225 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-rr8gx_kube-system(a783c047-d3a2-4f47-9f90-b7b15f3cce29)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-rr8gx_kube-system(a783c047-d3a2-4f47-9f90-b7b15f3cce29)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"eac95eaf6aac369e666b69e282a67917e7b959baa0dc3e1342d72647d1a7c311\\\": plugin type=\\\"calico\\\" failed (add): Get \\\"https://10.96.0.1:443/api/v1/namespaces/kube-system\\\": tls: failed to verify certificate: x509: certificate signed by unknown authority\"" pod="kube-system/coredns-674b8bbfcf-rr8gx" podUID="a783c047-d3a2-4f47-9f90-b7b15f3cce29" Sep 12 22:53:36.459837 containerd[2017]: time="2025-09-12T22:53:36.459604368Z" level=info msg="CreateContainer within sandbox \"30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 22:53:36.502462 containerd[2017]: time="2025-09-12T22:53:36.499653284Z" level=info msg="Container 4fb1c847244905041c626dd8bf7f5a1af204b0175666771db9a36365a2ef7d4b: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:36.507177 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4057612412.mount: Deactivated successfully. Sep 12 22:53:36.522888 containerd[2017]: time="2025-09-12T22:53:36.519760691Z" level=info msg="CreateContainer within sandbox \"30773b5581b616c728c4f712aae8e133a765642f91e47937622dece1466bcfbc\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"4fb1c847244905041c626dd8bf7f5a1af204b0175666771db9a36365a2ef7d4b\"" Sep 12 22:53:36.522888 containerd[2017]: time="2025-09-12T22:53:36.522602368Z" level=info msg="StartContainer for \"4fb1c847244905041c626dd8bf7f5a1af204b0175666771db9a36365a2ef7d4b\"" Sep 12 22:53:36.524520 containerd[2017]: time="2025-09-12T22:53:36.524438249Z" level=info msg="connecting to shim 4fb1c847244905041c626dd8bf7f5a1af204b0175666771db9a36365a2ef7d4b" address="unix:///run/containerd/s/5502f1411b5eabe18e9c155d32e8fc6f0275c47401e4b6338f145721cc0c9a75" protocol=ttrpc version=3 Sep 12 22:53:36.549990 systemd[1]: Started cri-containerd-4fb1c847244905041c626dd8bf7f5a1af204b0175666771db9a36365a2ef7d4b.scope - libcontainer container 4fb1c847244905041c626dd8bf7f5a1af204b0175666771db9a36365a2ef7d4b. Sep 12 22:53:36.610400 containerd[2017]: time="2025-09-12T22:53:36.610356101Z" level=info msg="StartContainer for \"4fb1c847244905041c626dd8bf7f5a1af204b0175666771db9a36365a2ef7d4b\" returns successfully" Sep 12 22:53:36.617945 containerd[2017]: time="2025-09-12T22:53:36.617897135Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5nmmm,Uid:6db03382-69c2-445f-a6ae-af82129366d4,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:36.837834 (udev-worker)[4631]: Network interface NamePolicy= disabled on kernel command line. Sep 12 22:53:36.848339 systemd-networkd[1897]: cali48bb707781e: Link UP Sep 12 22:53:36.852210 systemd-networkd[1897]: cali48bb707781e: Gained carrier Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.700 [INFO][5046] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0 csi-node-driver- calico-system 6db03382-69c2-445f-a6ae-af82129366d4 751 0 2025-09-12 22:53:12 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ip-172-31-29-105 csi-node-driver-5nmmm eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali48bb707781e [] [] }} ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.700 [INFO][5046] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.762 [INFO][5061] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" HandleID="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Workload="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.762 [INFO][5061] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" HandleID="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Workload="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5950), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-29-105", "pod":"csi-node-driver-5nmmm", "timestamp":"2025-09-12 22:53:36.762021074 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.762 [INFO][5061] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.762 [INFO][5061] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.762 [INFO][5061] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.772 [INFO][5061] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.782 [INFO][5061] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.789 [INFO][5061] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.793 [INFO][5061] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.797 [INFO][5061] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.797 [INFO][5061] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.800 [INFO][5061] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27 Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.808 [INFO][5061] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.819 [INFO][5061] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.131/26] block=192.168.125.128/26 handle="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.819 [INFO][5061] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.131/26] handle="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" host="ip-172-31-29-105" Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.819 [INFO][5061] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:36.881417 containerd[2017]: 2025-09-12 22:53:36.819 [INFO][5061] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.131/26] IPv6=[] ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" HandleID="k8s-pod-network.bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Workload="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" Sep 12 22:53:36.882569 containerd[2017]: 2025-09-12 22:53:36.826 [INFO][5046] cni-plugin/k8s.go 418: Populated endpoint ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"6db03382-69c2-445f-a6ae-af82129366d4", ResourceVersion:"751", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"csi-node-driver-5nmmm", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.125.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali48bb707781e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:36.882569 containerd[2017]: 2025-09-12 22:53:36.827 [INFO][5046] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.131/32] ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" Sep 12 22:53:36.882569 containerd[2017]: 2025-09-12 22:53:36.828 [INFO][5046] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali48bb707781e ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" Sep 12 22:53:36.882569 containerd[2017]: 2025-09-12 22:53:36.851 [INFO][5046] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" Sep 12 22:53:36.882569 containerd[2017]: 2025-09-12 22:53:36.853 [INFO][5046] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"6db03382-69c2-445f-a6ae-af82129366d4", ResourceVersion:"751", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27", Pod:"csi-node-driver-5nmmm", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.125.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali48bb707781e", MAC:"12:6c:56:06:61:81", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:36.882569 containerd[2017]: 2025-09-12 22:53:36.874 [INFO][5046] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" Namespace="calico-system" Pod="csi-node-driver-5nmmm" WorkloadEndpoint="ip--172--31--29--105-k8s-csi--node--driver--5nmmm-eth0" Sep 12 22:53:36.933781 containerd[2017]: time="2025-09-12T22:53:36.933710893Z" level=info msg="connecting to shim bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27" address="unix:///run/containerd/s/75f368a6ae892d71aae4fa58d4d8fe31ea2551143be0d7b587664a183ce637cd" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:36.999371 containerd[2017]: time="2025-09-12T22:53:36.998369574Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-rr8gx,Uid:a783c047-d3a2-4f47-9f90-b7b15f3cce29,Namespace:kube-system,Attempt:0,}" Sep 12 22:53:36.999868 systemd[1]: Started cri-containerd-bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27.scope - libcontainer container bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27. Sep 12 22:53:37.046244 kubelet[3352]: I0912 22:53:37.045817 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-z7kf4" podStartSLOduration=42.04575928 podStartE2EDuration="42.04575928s" podCreationTimestamp="2025-09-12 22:52:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:53:37.043409346 +0000 UTC m=+46.613159511" watchObservedRunningTime="2025-09-12 22:53:37.04575928 +0000 UTC m=+46.615509447" Sep 12 22:53:37.089889 systemd-networkd[1897]: vxlan.calico: Link UP Sep 12 22:53:37.090495 systemd-networkd[1897]: vxlan.calico: Gained carrier Sep 12 22:53:37.177684 systemd-networkd[1897]: cali61c56c1997a: Gained IPv6LL Sep 12 22:53:37.409250 containerd[2017]: time="2025-09-12T22:53:37.409115536Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-5nmmm,Uid:6db03382-69c2-445f-a6ae-af82129366d4,Namespace:calico-system,Attempt:0,} returns sandbox id \"bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27\"" Sep 12 22:53:37.511768 systemd-networkd[1897]: cali6a8822319a6: Link UP Sep 12 22:53:37.514045 systemd-networkd[1897]: cali6a8822319a6: Gained carrier Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.184 [INFO][5123] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0 coredns-674b8bbfcf- kube-system a783c047-d3a2-4f47-9f90-b7b15f3cce29 952 0 2025-09-12 22:52:55 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-29-105 coredns-674b8bbfcf-rr8gx eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali6a8822319a6 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.185 [INFO][5123] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.344 [INFO][5155] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" HandleID="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.345 [INFO][5155] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" HandleID="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e2c0), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-29-105", "pod":"coredns-674b8bbfcf-rr8gx", "timestamp":"2025-09-12 22:53:37.344436511 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.345 [INFO][5155] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.345 [INFO][5155] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.345 [INFO][5155] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.382 [INFO][5155] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.411 [INFO][5155] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.435 [INFO][5155] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.440 [INFO][5155] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.445 [INFO][5155] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.447 [INFO][5155] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.453 [INFO][5155] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.463 [INFO][5155] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.478 [INFO][5155] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.132/26] block=192.168.125.128/26 handle="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.481 [INFO][5155] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.132/26] handle="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" host="ip-172-31-29-105" Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.481 [INFO][5155] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:37.559185 containerd[2017]: 2025-09-12 22:53:37.481 [INFO][5155] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.132/26] IPv6=[] ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" HandleID="k8s-pod-network.f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Workload="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:37.560532 containerd[2017]: 2025-09-12 22:53:37.495 [INFO][5123] cni-plugin/k8s.go 418: Populated endpoint ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a783c047-d3a2-4f47-9f90-b7b15f3cce29", ResourceVersion:"952", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 52, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"coredns-674b8bbfcf-rr8gx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.125.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6a8822319a6", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:37.560532 containerd[2017]: 2025-09-12 22:53:37.495 [INFO][5123] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.132/32] ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:37.560532 containerd[2017]: 2025-09-12 22:53:37.495 [INFO][5123] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6a8822319a6 ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:37.560532 containerd[2017]: 2025-09-12 22:53:37.516 [INFO][5123] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:37.560532 containerd[2017]: 2025-09-12 22:53:37.517 [INFO][5123] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"a783c047-d3a2-4f47-9f90-b7b15f3cce29", ResourceVersion:"952", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 52, 55, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b", Pod:"coredns-674b8bbfcf-rr8gx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.125.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali6a8822319a6", MAC:"3e:59:7d:b7:ac:5a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:37.560532 containerd[2017]: 2025-09-12 22:53:37.543 [INFO][5123] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" Namespace="kube-system" Pod="coredns-674b8bbfcf-rr8gx" WorkloadEndpoint="ip--172--31--29--105-k8s-coredns--674b8bbfcf--rr8gx-eth0" Sep 12 22:53:37.624790 containerd[2017]: time="2025-09-12T22:53:37.623709183Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74b867d4d7-w6r2c,Uid:f284a142-6481-40f9-8099-8b7bffb670e7,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:37.639660 containerd[2017]: time="2025-09-12T22:53:37.637648498Z" level=info msg="connecting to shim f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b" address="unix:///run/containerd/s/5f01af13ea82e5e4c82f1a2ab4e5ffac9e3b606d95226c20514f31037ffb9a20" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:37.711919 containerd[2017]: time="2025-09-12T22:53:37.711793486Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:37.715169 containerd[2017]: time="2025-09-12T22:53:37.715123649Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4661291" Sep 12 22:53:37.721876 containerd[2017]: time="2025-09-12T22:53:37.721687401Z" level=info msg="ImageCreate event name:\"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:37.778197 containerd[2017]: time="2025-09-12T22:53:37.778146739Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:37.781323 containerd[2017]: time="2025-09-12T22:53:37.780511647Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"6153986\" in 2.124863854s" Sep 12 22:53:37.781323 containerd[2017]: time="2025-09-12T22:53:37.780562882Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\"" Sep 12 22:53:37.787542 systemd[1]: Started cri-containerd-f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b.scope - libcontainer container f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b. Sep 12 22:53:37.812313 containerd[2017]: time="2025-09-12T22:53:37.811277007Z" level=info msg="CreateContainer within sandbox \"42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Sep 12 22:53:37.816029 containerd[2017]: time="2025-09-12T22:53:37.815980009Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Sep 12 22:53:37.848836 containerd[2017]: time="2025-09-12T22:53:37.848021291Z" level=info msg="Container 8768867f29153dff62834b95e269ffa20998c76d602e965d08ad7a03f54d3f7b: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:37.892266 containerd[2017]: time="2025-09-12T22:53:37.892054128Z" level=info msg="CreateContainer within sandbox \"42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"8768867f29153dff62834b95e269ffa20998c76d602e965d08ad7a03f54d3f7b\"" Sep 12 22:53:37.893704 containerd[2017]: time="2025-09-12T22:53:37.893650409Z" level=info msg="StartContainer for \"8768867f29153dff62834b95e269ffa20998c76d602e965d08ad7a03f54d3f7b\"" Sep 12 22:53:37.902611 containerd[2017]: time="2025-09-12T22:53:37.899355057Z" level=info msg="connecting to shim 8768867f29153dff62834b95e269ffa20998c76d602e965d08ad7a03f54d3f7b" address="unix:///run/containerd/s/9ff71b9eebcf79e0d2f744403665595ba286313db55d33808b7259670a0ab9dc" protocol=ttrpc version=3 Sep 12 22:53:37.960887 systemd[1]: Started cri-containerd-8768867f29153dff62834b95e269ffa20998c76d602e965d08ad7a03f54d3f7b.scope - libcontainer container 8768867f29153dff62834b95e269ffa20998c76d602e965d08ad7a03f54d3f7b. Sep 12 22:53:38.034710 containerd[2017]: time="2025-09-12T22:53:38.034527157Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-rr8gx,Uid:a783c047-d3a2-4f47-9f90-b7b15f3cce29,Namespace:kube-system,Attempt:0,} returns sandbox id \"f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b\"" Sep 12 22:53:38.045303 containerd[2017]: time="2025-09-12T22:53:38.045243272Z" level=info msg="CreateContainer within sandbox \"f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 12 22:53:38.064747 containerd[2017]: time="2025-09-12T22:53:38.063934560Z" level=info msg="Container cc0703112c79d060a077de4225a0449f147d5c4d1cb3d26be511cf0cec1e5e23: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:38.078811 containerd[2017]: time="2025-09-12T22:53:38.078770333Z" level=info msg="CreateContainer within sandbox \"f35b882a153a9a1c3153f7c089151484da7cff12b9e9de58646f6636ee030b6b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"cc0703112c79d060a077de4225a0449f147d5c4d1cb3d26be511cf0cec1e5e23\"" Sep 12 22:53:38.080353 containerd[2017]: time="2025-09-12T22:53:38.080319078Z" level=info msg="StartContainer for \"cc0703112c79d060a077de4225a0449f147d5c4d1cb3d26be511cf0cec1e5e23\"" Sep 12 22:53:38.082881 containerd[2017]: time="2025-09-12T22:53:38.082841402Z" level=info msg="connecting to shim cc0703112c79d060a077de4225a0449f147d5c4d1cb3d26be511cf0cec1e5e23" address="unix:///run/containerd/s/5f01af13ea82e5e4c82f1a2ab4e5ffac9e3b606d95226c20514f31037ffb9a20" protocol=ttrpc version=3 Sep 12 22:53:38.103619 systemd-networkd[1897]: cali21225bcdcde: Link UP Sep 12 22:53:38.109047 systemd-networkd[1897]: cali21225bcdcde: Gained carrier Sep 12 22:53:38.125122 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3552804170.mount: Deactivated successfully. Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.834 [INFO][5202] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0 calico-kube-controllers-74b867d4d7- calico-system f284a142-6481-40f9-8099-8b7bffb670e7 866 0 2025-09-12 22:53:13 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:74b867d4d7 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ip-172-31-29-105 calico-kube-controllers-74b867d4d7-w6r2c eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali21225bcdcde [] [] }} ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.834 [INFO][5202] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.960 [INFO][5246] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" HandleID="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Workload="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.960 [INFO][5246] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" HandleID="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Workload="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fab0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-29-105", "pod":"calico-kube-controllers-74b867d4d7-w6r2c", "timestamp":"2025-09-12 22:53:37.960253407 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.961 [INFO][5246] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.961 [INFO][5246] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.961 [INFO][5246] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.974 [INFO][5246] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:37.995 [INFO][5246] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.009 [INFO][5246] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.015 [INFO][5246] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.028 [INFO][5246] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.028 [INFO][5246] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.038 [INFO][5246] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4 Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.051 [INFO][5246] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.072 [INFO][5246] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.133/26] block=192.168.125.128/26 handle="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.073 [INFO][5246] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.133/26] handle="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" host="ip-172-31-29-105" Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.074 [INFO][5246] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:38.170137 containerd[2017]: 2025-09-12 22:53:38.074 [INFO][5246] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.133/26] IPv6=[] ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" HandleID="k8s-pod-network.ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Workload="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" Sep 12 22:53:38.173149 containerd[2017]: 2025-09-12 22:53:38.083 [INFO][5202] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0", GenerateName:"calico-kube-controllers-74b867d4d7-", Namespace:"calico-system", SelfLink:"", UID:"f284a142-6481-40f9-8099-8b7bffb670e7", ResourceVersion:"866", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"74b867d4d7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"calico-kube-controllers-74b867d4d7-w6r2c", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.125.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali21225bcdcde", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:38.173149 containerd[2017]: 2025-09-12 22:53:38.084 [INFO][5202] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.133/32] ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" Sep 12 22:53:38.173149 containerd[2017]: 2025-09-12 22:53:38.084 [INFO][5202] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali21225bcdcde ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" Sep 12 22:53:38.173149 containerd[2017]: 2025-09-12 22:53:38.115 [INFO][5202] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" Sep 12 22:53:38.173149 containerd[2017]: 2025-09-12 22:53:38.120 [INFO][5202] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0", GenerateName:"calico-kube-controllers-74b867d4d7-", Namespace:"calico-system", SelfLink:"", UID:"f284a142-6481-40f9-8099-8b7bffb670e7", ResourceVersion:"866", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"74b867d4d7", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4", Pod:"calico-kube-controllers-74b867d4d7-w6r2c", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.125.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali21225bcdcde", MAC:"42:4d:bf:f7:c8:b1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:38.173149 containerd[2017]: 2025-09-12 22:53:38.165 [INFO][5202] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" Namespace="calico-system" Pod="calico-kube-controllers-74b867d4d7-w6r2c" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--kube--controllers--74b867d4d7--w6r2c-eth0" Sep 12 22:53:38.177710 systemd[1]: Started cri-containerd-cc0703112c79d060a077de4225a0449f147d5c4d1cb3d26be511cf0cec1e5e23.scope - libcontainer container cc0703112c79d060a077de4225a0449f147d5c4d1cb3d26be511cf0cec1e5e23. Sep 12 22:53:38.277760 containerd[2017]: time="2025-09-12T22:53:38.277710566Z" level=info msg="connecting to shim ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4" address="unix:///run/containerd/s/87c2bad3a04f3e75c82a447de9d5419cc40b51a29c5650fbe243fcd8aa8cc847" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:38.337737 systemd[1]: Started cri-containerd-ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4.scope - libcontainer container ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4. Sep 12 22:53:38.380329 containerd[2017]: time="2025-09-12T22:53:38.380274617Z" level=info msg="StartContainer for \"cc0703112c79d060a077de4225a0449f147d5c4d1cb3d26be511cf0cec1e5e23\" returns successfully" Sep 12 22:53:38.522916 containerd[2017]: time="2025-09-12T22:53:38.522861548Z" level=info msg="StartContainer for \"8768867f29153dff62834b95e269ffa20998c76d602e965d08ad7a03f54d3f7b\" returns successfully" Sep 12 22:53:38.618341 containerd[2017]: time="2025-09-12T22:53:38.618291683Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-4265x,Uid:229ca78f-57bf-4bb2-ac2b-38651bdb311d,Namespace:calico-system,Attempt:0,}" Sep 12 22:53:38.620403 containerd[2017]: time="2025-09-12T22:53:38.619297501Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-sbkxs,Uid:b2baf514-f775-48c0-bd00-f9a124668792,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:53:38.649644 systemd-networkd[1897]: vxlan.calico: Gained IPv6LL Sep 12 22:53:38.714279 systemd-networkd[1897]: cali48bb707781e: Gained IPv6LL Sep 12 22:53:38.941636 containerd[2017]: time="2025-09-12T22:53:38.940813049Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-74b867d4d7-w6r2c,Uid:f284a142-6481-40f9-8099-8b7bffb670e7,Namespace:calico-system,Attempt:0,} returns sandbox id \"ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4\"" Sep 12 22:53:38.997974 systemd-networkd[1897]: cali07be9a1513e: Link UP Sep 12 22:53:38.998269 systemd-networkd[1897]: cali07be9a1513e: Gained carrier Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.763 [INFO][5388] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0 calico-apiserver-5c7666b7bb- calico-apiserver b2baf514-f775-48c0-bd00-f9a124668792 870 0 2025-09-12 22:53:07 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5c7666b7bb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-29-105 calico-apiserver-5c7666b7bb-sbkxs eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali07be9a1513e [] [] }} ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.764 [INFO][5388] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.868 [INFO][5410] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" HandleID="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Workload="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.868 [INFO][5410] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" HandleID="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Workload="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000381b60), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-29-105", "pod":"calico-apiserver-5c7666b7bb-sbkxs", "timestamp":"2025-09-12 22:53:38.865956746 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.868 [INFO][5410] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.868 [INFO][5410] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.868 [INFO][5410] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.901 [INFO][5410] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.924 [INFO][5410] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.942 [INFO][5410] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.948 [INFO][5410] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.953 [INFO][5410] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.953 [INFO][5410] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.957 [INFO][5410] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022 Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.968 [INFO][5410] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.984 [INFO][5410] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.134/26] block=192.168.125.128/26 handle="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.985 [INFO][5410] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.134/26] handle="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" host="ip-172-31-29-105" Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.985 [INFO][5410] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:39.033292 containerd[2017]: 2025-09-12 22:53:38.985 [INFO][5410] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.134/26] IPv6=[] ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" HandleID="k8s-pod-network.8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Workload="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" Sep 12 22:53:39.035965 containerd[2017]: 2025-09-12 22:53:38.992 [INFO][5388] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0", GenerateName:"calico-apiserver-5c7666b7bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"b2baf514-f775-48c0-bd00-f9a124668792", ResourceVersion:"870", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7666b7bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"calico-apiserver-5c7666b7bb-sbkxs", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.125.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali07be9a1513e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:39.035965 containerd[2017]: 2025-09-12 22:53:38.992 [INFO][5388] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.134/32] ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" Sep 12 22:53:39.035965 containerd[2017]: 2025-09-12 22:53:38.992 [INFO][5388] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali07be9a1513e ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" Sep 12 22:53:39.035965 containerd[2017]: 2025-09-12 22:53:38.995 [INFO][5388] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" Sep 12 22:53:39.035965 containerd[2017]: 2025-09-12 22:53:38.995 [INFO][5388] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0", GenerateName:"calico-apiserver-5c7666b7bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"b2baf514-f775-48c0-bd00-f9a124668792", ResourceVersion:"870", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7666b7bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022", Pod:"calico-apiserver-5c7666b7bb-sbkxs", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.125.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali07be9a1513e", MAC:"da:12:54:99:2e:55", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:39.035965 containerd[2017]: 2025-09-12 22:53:39.028 [INFO][5388] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-sbkxs" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--sbkxs-eth0" Sep 12 22:53:39.107248 containerd[2017]: time="2025-09-12T22:53:39.107180283Z" level=info msg="connecting to shim 8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022" address="unix:///run/containerd/s/8a1053c7bb5f99f4164311cad50885aac214da7f33318db865ed07297e69945e" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:39.207138 systemd-networkd[1897]: cali3c68d22c2ae: Link UP Sep 12 22:53:39.215865 systemd-networkd[1897]: cali3c68d22c2ae: Gained carrier Sep 12 22:53:39.226131 systemd-networkd[1897]: cali6a8822319a6: Gained IPv6LL Sep 12 22:53:39.260714 systemd[1]: Started cri-containerd-8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022.scope - libcontainer container 8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022. Sep 12 22:53:39.286240 kubelet[3352]: I0912 22:53:39.286114 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-rr8gx" podStartSLOduration=44.286089881 podStartE2EDuration="44.286089881s" podCreationTimestamp="2025-09-12 22:52:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-12 22:53:39.179692716 +0000 UTC m=+48.749442883" watchObservedRunningTime="2025-09-12 22:53:39.286089881 +0000 UTC m=+48.855840047" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:38.765 [INFO][5386] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0 goldmane-54d579b49d- calico-system 229ca78f-57bf-4bb2-ac2b-38651bdb311d 869 0 2025-09-12 22:53:12 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ip-172-31-29-105 goldmane-54d579b49d-4265x eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali3c68d22c2ae [] [] }} ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:38.769 [INFO][5386] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:38.870 [INFO][5415] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" HandleID="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Workload="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:38.871 [INFO][5415] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" HandleID="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Workload="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002c70a0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-29-105", "pod":"goldmane-54d579b49d-4265x", "timestamp":"2025-09-12 22:53:38.870883223 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:38.871 [INFO][5415] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:38.985 [INFO][5415] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:38.985 [INFO][5415] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.007 [INFO][5415] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.025 [INFO][5415] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.046 [INFO][5415] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.053 [INFO][5415] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.065 [INFO][5415] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.065 [INFO][5415] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.068 [INFO][5415] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554 Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.077 [INFO][5415] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.154 [INFO][5415] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.135/26] block=192.168.125.128/26 handle="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.155 [INFO][5415] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.135/26] handle="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" host="ip-172-31-29-105" Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.155 [INFO][5415] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:39.295310 containerd[2017]: 2025-09-12 22:53:39.155 [INFO][5415] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.135/26] IPv6=[] ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" HandleID="k8s-pod-network.b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Workload="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" Sep 12 22:53:39.298203 containerd[2017]: 2025-09-12 22:53:39.182 [INFO][5386] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"229ca78f-57bf-4bb2-ac2b-38651bdb311d", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"goldmane-54d579b49d-4265x", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.125.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3c68d22c2ae", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:39.298203 containerd[2017]: 2025-09-12 22:53:39.182 [INFO][5386] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.135/32] ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" Sep 12 22:53:39.298203 containerd[2017]: 2025-09-12 22:53:39.182 [INFO][5386] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3c68d22c2ae ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" Sep 12 22:53:39.298203 containerd[2017]: 2025-09-12 22:53:39.220 [INFO][5386] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" Sep 12 22:53:39.298203 containerd[2017]: 2025-09-12 22:53:39.237 [INFO][5386] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"229ca78f-57bf-4bb2-ac2b-38651bdb311d", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 12, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554", Pod:"goldmane-54d579b49d-4265x", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.125.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali3c68d22c2ae", MAC:"da:8e:d1:e6:90:39", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:39.298203 containerd[2017]: 2025-09-12 22:53:39.289 [INFO][5386] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" Namespace="calico-system" Pod="goldmane-54d579b49d-4265x" WorkloadEndpoint="ip--172--31--29--105-k8s-goldmane--54d579b49d--4265x-eth0" Sep 12 22:53:39.393261 containerd[2017]: time="2025-09-12T22:53:39.393197774Z" level=info msg="connecting to shim b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554" address="unix:///run/containerd/s/e9ec3c350099cfcfa4473e9a4156504030d54ecf72057c526396a8c8a88ba657" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:39.418645 systemd-networkd[1897]: cali21225bcdcde: Gained IPv6LL Sep 12 22:53:39.504889 systemd[1]: Started cri-containerd-b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554.scope - libcontainer container b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554. Sep 12 22:53:39.617942 containerd[2017]: time="2025-09-12T22:53:39.617605714Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-d8qbk,Uid:c41fa72e-0e56-4db0-8fb9-ce3d488fe85d,Namespace:calico-apiserver,Attempt:0,}" Sep 12 22:53:39.988980 systemd-networkd[1897]: cali696ebcdf1be: Link UP Sep 12 22:53:39.990836 systemd-networkd[1897]: cali696ebcdf1be: Gained carrier Sep 12 22:53:40.003661 containerd[2017]: time="2025-09-12T22:53:40.003396012Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:40.006056 containerd[2017]: time="2025-09-12T22:53:40.005928880Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8760527" Sep 12 22:53:40.009378 containerd[2017]: time="2025-09-12T22:53:40.009321913Z" level=info msg="ImageCreate event name:\"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:40.019581 containerd[2017]: time="2025-09-12T22:53:40.019501860Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:40.027073 containerd[2017]: time="2025-09-12T22:53:40.027012832Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"10253230\" in 2.210979717s" Sep 12 22:53:40.027073 containerd[2017]: time="2025-09-12T22:53:40.027077252Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\"" Sep 12 22:53:40.039050 containerd[2017]: time="2025-09-12T22:53:40.039003288Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.786 [INFO][5555] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0 calico-apiserver-5c7666b7bb- calico-apiserver c41fa72e-0e56-4db0-8fb9-ce3d488fe85d 864 0 2025-09-12 22:53:07 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:5c7666b7bb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-29-105 calico-apiserver-5c7666b7bb-d8qbk eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali696ebcdf1be [] [] }} ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.786 [INFO][5555] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.875 [INFO][5573] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" HandleID="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Workload="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.877 [INFO][5573] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" HandleID="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Workload="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004e810), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-29-105", "pod":"calico-apiserver-5c7666b7bb-d8qbk", "timestamp":"2025-09-12 22:53:39.875113086 +0000 UTC"}, Hostname:"ip-172-31-29-105", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.877 [INFO][5573] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.877 [INFO][5573] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.877 [INFO][5573] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-29-105' Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.892 [INFO][5573] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.902 [INFO][5573] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.912 [INFO][5573] ipam/ipam.go 511: Trying affinity for 192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.917 [INFO][5573] ipam/ipam.go 158: Attempting to load block cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.924 [INFO][5573] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.125.128/26 host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.924 [INFO][5573] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.125.128/26 handle="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.930 [INFO][5573] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432 Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.944 [INFO][5573] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.125.128/26 handle="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.967 [INFO][5573] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.125.136/26] block=192.168.125.128/26 handle="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.968 [INFO][5573] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.125.136/26] handle="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" host="ip-172-31-29-105" Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.969 [INFO][5573] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Sep 12 22:53:40.042500 containerd[2017]: 2025-09-12 22:53:39.969 [INFO][5573] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.125.136/26] IPv6=[] ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" HandleID="k8s-pod-network.4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Workload="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" Sep 12 22:53:40.045991 containerd[2017]: 2025-09-12 22:53:39.982 [INFO][5555] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0", GenerateName:"calico-apiserver-5c7666b7bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"c41fa72e-0e56-4db0-8fb9-ce3d488fe85d", ResourceVersion:"864", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7666b7bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"", Pod:"calico-apiserver-5c7666b7bb-d8qbk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.125.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali696ebcdf1be", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:40.045991 containerd[2017]: 2025-09-12 22:53:39.982 [INFO][5555] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.125.136/32] ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" Sep 12 22:53:40.045991 containerd[2017]: 2025-09-12 22:53:39.982 [INFO][5555] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali696ebcdf1be ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" Sep 12 22:53:40.045991 containerd[2017]: 2025-09-12 22:53:39.986 [INFO][5555] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" Sep 12 22:53:40.045991 containerd[2017]: 2025-09-12 22:53:39.986 [INFO][5555] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0", GenerateName:"calico-apiserver-5c7666b7bb-", Namespace:"calico-apiserver", SelfLink:"", UID:"c41fa72e-0e56-4db0-8fb9-ce3d488fe85d", ResourceVersion:"864", Generation:0, CreationTimestamp:time.Date(2025, time.September, 12, 22, 53, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"5c7666b7bb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-29-105", ContainerID:"4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432", Pod:"calico-apiserver-5c7666b7bb-d8qbk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.125.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali696ebcdf1be", MAC:"86:6c:7b:b6:f0:b1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Sep 12 22:53:40.045991 containerd[2017]: 2025-09-12 22:53:40.019 [INFO][5555] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" Namespace="calico-apiserver" Pod="calico-apiserver-5c7666b7bb-d8qbk" WorkloadEndpoint="ip--172--31--29--105-k8s-calico--apiserver--5c7666b7bb--d8qbk-eth0" Sep 12 22:53:40.047440 containerd[2017]: time="2025-09-12T22:53:40.046704364Z" level=info msg="CreateContainer within sandbox \"bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 12 22:53:40.137614 containerd[2017]: time="2025-09-12T22:53:40.137558834Z" level=info msg="Container b681beb507dcd1fc64efcaae2433a177f4e27d7f810407af8f61b76d415b3e8c: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:40.148830 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount48343077.mount: Deactivated successfully. Sep 12 22:53:40.193860 containerd[2017]: time="2025-09-12T22:53:40.193536433Z" level=info msg="CreateContainer within sandbox \"bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"b681beb507dcd1fc64efcaae2433a177f4e27d7f810407af8f61b76d415b3e8c\"" Sep 12 22:53:40.195471 containerd[2017]: time="2025-09-12T22:53:40.195090023Z" level=info msg="StartContainer for \"b681beb507dcd1fc64efcaae2433a177f4e27d7f810407af8f61b76d415b3e8c\"" Sep 12 22:53:40.216530 containerd[2017]: time="2025-09-12T22:53:40.215929955Z" level=info msg="connecting to shim b681beb507dcd1fc64efcaae2433a177f4e27d7f810407af8f61b76d415b3e8c" address="unix:///run/containerd/s/75f368a6ae892d71aae4fa58d4d8fe31ea2551143be0d7b587664a183ce637cd" protocol=ttrpc version=3 Sep 12 22:53:40.220781 containerd[2017]: time="2025-09-12T22:53:40.220720102Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-sbkxs,Uid:b2baf514-f775-48c0-bd00-f9a124668792,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022\"" Sep 12 22:53:40.277813 systemd[1]: Started sshd@9-172.31.29.105:22-139.178.89.65:33354.service - OpenSSH per-connection server daemon (139.178.89.65:33354). Sep 12 22:53:40.301053 containerd[2017]: time="2025-09-12T22:53:40.296793004Z" level=info msg="connecting to shim 4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432" address="unix:///run/containerd/s/9896e2c02032a0ff36f9eb1b99594a43a6b8e290bfb6917933b25ef9f1599257" namespace=k8s.io protocol=ttrpc version=3 Sep 12 22:53:40.357724 systemd[1]: Started cri-containerd-b681beb507dcd1fc64efcaae2433a177f4e27d7f810407af8f61b76d415b3e8c.scope - libcontainer container b681beb507dcd1fc64efcaae2433a177f4e27d7f810407af8f61b76d415b3e8c. Sep 12 22:53:40.490960 containerd[2017]: time="2025-09-12T22:53:40.489780605Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-4265x,Uid:229ca78f-57bf-4bb2-ac2b-38651bdb311d,Namespace:calico-system,Attempt:0,} returns sandbox id \"b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554\"" Sep 12 22:53:40.496804 systemd[1]: Started cri-containerd-4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432.scope - libcontainer container 4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432. Sep 12 22:53:40.505977 systemd-networkd[1897]: cali07be9a1513e: Gained IPv6LL Sep 12 22:53:40.622545 sshd[5609]: Accepted publickey for core from 139.178.89.65 port 33354 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:53:40.627610 sshd-session[5609]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:53:40.642704 systemd-logind[1983]: New session 10 of user core. Sep 12 22:53:40.645715 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 12 22:53:40.788666 containerd[2017]: time="2025-09-12T22:53:40.788619252Z" level=info msg="StartContainer for \"b681beb507dcd1fc64efcaae2433a177f4e27d7f810407af8f61b76d415b3e8c\" returns successfully" Sep 12 22:53:40.870895 containerd[2017]: time="2025-09-12T22:53:40.870701290Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-5c7666b7bb-d8qbk,Uid:c41fa72e-0e56-4db0-8fb9-ce3d488fe85d,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432\"" Sep 12 22:53:40.889725 systemd-networkd[1897]: cali3c68d22c2ae: Gained IPv6LL Sep 12 22:53:41.210077 systemd-networkd[1897]: cali696ebcdf1be: Gained IPv6LL Sep 12 22:53:41.796668 sshd[5671]: Connection closed by 139.178.89.65 port 33354 Sep 12 22:53:41.797099 sshd-session[5609]: pam_unix(sshd:session): session closed for user core Sep 12 22:53:41.804319 systemd-logind[1983]: Session 10 logged out. Waiting for processes to exit. Sep 12 22:53:41.804905 systemd[1]: sshd@9-172.31.29.105:22-139.178.89.65:33354.service: Deactivated successfully. Sep 12 22:53:41.809276 systemd[1]: session-10.scope: Deactivated successfully. Sep 12 22:53:41.812509 systemd-logind[1983]: Removed session 10. Sep 12 22:53:42.951398 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2439828990.mount: Deactivated successfully. Sep 12 22:53:42.991146 containerd[2017]: time="2025-09-12T22:53:42.991080084Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:42.993244 containerd[2017]: time="2025-09-12T22:53:42.993093278Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=33085545" Sep 12 22:53:42.995436 containerd[2017]: time="2025-09-12T22:53:42.995334853Z" level=info msg="ImageCreate event name:\"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:42.999749 containerd[2017]: time="2025-09-12T22:53:42.998881375Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:42.999749 containerd[2017]: time="2025-09-12T22:53:42.999621385Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"33085375\" in 2.960570034s" Sep 12 22:53:42.999749 containerd[2017]: time="2025-09-12T22:53:42.999649961Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\"" Sep 12 22:53:43.001314 containerd[2017]: time="2025-09-12T22:53:43.001292404Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Sep 12 22:53:43.007302 containerd[2017]: time="2025-09-12T22:53:43.007260361Z" level=info msg="CreateContainer within sandbox \"42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Sep 12 22:53:43.026471 containerd[2017]: time="2025-09-12T22:53:43.024712146Z" level=info msg="Container 92b8c30c57f4fb871f7c7b6e32e0cf57f6428089364faeb3550a7adabb385668: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:43.048812 containerd[2017]: time="2025-09-12T22:53:43.048722382Z" level=info msg="CreateContainer within sandbox \"42097c8419674bace891fa5a7e975069c8cd7abb3757289068731f5d10312752\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"92b8c30c57f4fb871f7c7b6e32e0cf57f6428089364faeb3550a7adabb385668\"" Sep 12 22:53:43.050550 containerd[2017]: time="2025-09-12T22:53:43.050498407Z" level=info msg="StartContainer for \"92b8c30c57f4fb871f7c7b6e32e0cf57f6428089364faeb3550a7adabb385668\"" Sep 12 22:53:43.052085 containerd[2017]: time="2025-09-12T22:53:43.051981219Z" level=info msg="connecting to shim 92b8c30c57f4fb871f7c7b6e32e0cf57f6428089364faeb3550a7adabb385668" address="unix:///run/containerd/s/9ff71b9eebcf79e0d2f744403665595ba286313db55d33808b7259670a0ab9dc" protocol=ttrpc version=3 Sep 12 22:53:43.082021 systemd[1]: Started cri-containerd-92b8c30c57f4fb871f7c7b6e32e0cf57f6428089364faeb3550a7adabb385668.scope - libcontainer container 92b8c30c57f4fb871f7c7b6e32e0cf57f6428089364faeb3550a7adabb385668. Sep 12 22:53:43.171415 containerd[2017]: time="2025-09-12T22:53:43.171374111Z" level=info msg="StartContainer for \"92b8c30c57f4fb871f7c7b6e32e0cf57f6428089364faeb3550a7adabb385668\" returns successfully" Sep 12 22:53:43.310117 ntpd[2166]: Listen normally on 6 vxlan.calico 192.168.125.128:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 6 vxlan.calico 192.168.125.128:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 7 calib3dca84709b [fe80::ecee:eeff:feee:eeee%4]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 8 cali61c56c1997a [fe80::ecee:eeff:feee:eeee%5]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 9 cali48bb707781e [fe80::ecee:eeff:feee:eeee%6]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 10 vxlan.calico [fe80::6429:b0ff:fe75:d977%7]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 11 cali6a8822319a6 [fe80::ecee:eeff:feee:eeee%10]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 12 cali21225bcdcde [fe80::ecee:eeff:feee:eeee%11]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 13 cali07be9a1513e [fe80::ecee:eeff:feee:eeee%12]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 14 cali3c68d22c2ae [fe80::ecee:eeff:feee:eeee%13]:123 Sep 12 22:53:43.310910 ntpd[2166]: 12 Sep 22:53:43 ntpd[2166]: Listen normally on 15 cali696ebcdf1be [fe80::ecee:eeff:feee:eeee%14]:123 Sep 12 22:53:43.310175 ntpd[2166]: Listen normally on 7 calib3dca84709b [fe80::ecee:eeff:feee:eeee%4]:123 Sep 12 22:53:43.310198 ntpd[2166]: Listen normally on 8 cali61c56c1997a [fe80::ecee:eeff:feee:eeee%5]:123 Sep 12 22:53:43.310218 ntpd[2166]: Listen normally on 9 cali48bb707781e [fe80::ecee:eeff:feee:eeee%6]:123 Sep 12 22:53:43.310237 ntpd[2166]: Listen normally on 10 vxlan.calico [fe80::6429:b0ff:fe75:d977%7]:123 Sep 12 22:53:43.310256 ntpd[2166]: Listen normally on 11 cali6a8822319a6 [fe80::ecee:eeff:feee:eeee%10]:123 Sep 12 22:53:43.310274 ntpd[2166]: Listen normally on 12 cali21225bcdcde [fe80::ecee:eeff:feee:eeee%11]:123 Sep 12 22:53:43.310293 ntpd[2166]: Listen normally on 13 cali07be9a1513e [fe80::ecee:eeff:feee:eeee%12]:123 Sep 12 22:53:43.310317 ntpd[2166]: Listen normally on 14 cali3c68d22c2ae [fe80::ecee:eeff:feee:eeee%13]:123 Sep 12 22:53:43.310336 ntpd[2166]: Listen normally on 15 cali696ebcdf1be [fe80::ecee:eeff:feee:eeee%14]:123 Sep 12 22:53:44.134260 kubelet[3352]: I0912 22:53:44.134196 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-55bb7dd67f-h76qr" podStartSLOduration=2.762814607 podStartE2EDuration="10.134177648s" podCreationTimestamp="2025-09-12 22:53:34 +0000 UTC" firstStartedPulling="2025-09-12 22:53:35.629347282 +0000 UTC m=+45.199097425" lastFinishedPulling="2025-09-12 22:53:43.000710308 +0000 UTC m=+52.570460466" observedRunningTime="2025-09-12 22:53:44.133709056 +0000 UTC m=+53.703459223" watchObservedRunningTime="2025-09-12 22:53:44.134177648 +0000 UTC m=+53.703927811" Sep 12 22:53:46.838382 systemd[1]: Started sshd@10-172.31.29.105:22-139.178.89.65:33360.service - OpenSSH per-connection server daemon (139.178.89.65:33360). Sep 12 22:53:47.057775 containerd[2017]: time="2025-09-12T22:53:47.057712156Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:47.059904 containerd[2017]: time="2025-09-12T22:53:47.059871044Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=51277746" Sep 12 22:53:47.063179 containerd[2017]: time="2025-09-12T22:53:47.063119356Z" level=info msg="ImageCreate event name:\"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:47.066989 containerd[2017]: time="2025-09-12T22:53:47.066931053Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:47.067878 containerd[2017]: time="2025-09-12T22:53:47.067526658Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"52770417\" in 4.066017797s" Sep 12 22:53:47.067878 containerd[2017]: time="2025-09-12T22:53:47.067563267Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\"" Sep 12 22:53:47.069425 containerd[2017]: time="2025-09-12T22:53:47.069370816Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 12 22:53:47.104626 sshd[5760]: Accepted publickey for core from 139.178.89.65 port 33360 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:53:47.108802 sshd-session[5760]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:53:47.123407 systemd-logind[1983]: New session 11 of user core. Sep 12 22:53:47.136705 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 12 22:53:47.149335 containerd[2017]: time="2025-09-12T22:53:47.149295504Z" level=info msg="CreateContainer within sandbox \"ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 12 22:53:47.268534 containerd[2017]: time="2025-09-12T22:53:47.268201720Z" level=info msg="Container bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:47.309346 containerd[2017]: time="2025-09-12T22:53:47.309298683Z" level=info msg="CreateContainer within sandbox \"ee26e6f0ea959bea3f209bb1ef682895528be92b350d390a7f924fc17082bfe4\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\"" Sep 12 22:53:47.310936 containerd[2017]: time="2025-09-12T22:53:47.310638703Z" level=info msg="StartContainer for \"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\"" Sep 12 22:53:47.312587 containerd[2017]: time="2025-09-12T22:53:47.312556971Z" level=info msg="connecting to shim bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3" address="unix:///run/containerd/s/87c2bad3a04f3e75c82a447de9d5419cc40b51a29c5650fbe243fcd8aa8cc847" protocol=ttrpc version=3 Sep 12 22:53:47.423787 systemd[1]: Started cri-containerd-bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3.scope - libcontainer container bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3. Sep 12 22:53:47.651077 containerd[2017]: time="2025-09-12T22:53:47.651018607Z" level=info msg="StartContainer for \"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\" returns successfully" Sep 12 22:53:47.980595 sshd[5765]: Connection closed by 139.178.89.65 port 33360 Sep 12 22:53:47.982696 sshd-session[5760]: pam_unix(sshd:session): session closed for user core Sep 12 22:53:47.989153 systemd[1]: sshd@10-172.31.29.105:22-139.178.89.65:33360.service: Deactivated successfully. Sep 12 22:53:47.993440 systemd[1]: session-11.scope: Deactivated successfully. Sep 12 22:53:48.007781 systemd-logind[1983]: Session 11 logged out. Waiting for processes to exit. Sep 12 22:53:48.011731 systemd-logind[1983]: Removed session 11. Sep 12 22:53:48.361349 kubelet[3352]: I0912 22:53:48.359828 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-74b867d4d7-w6r2c" podStartSLOduration=27.2338053 podStartE2EDuration="35.359801413s" podCreationTimestamp="2025-09-12 22:53:13 +0000 UTC" firstStartedPulling="2025-09-12 22:53:38.943051671 +0000 UTC m=+48.512801829" lastFinishedPulling="2025-09-12 22:53:47.069047799 +0000 UTC m=+56.638797942" observedRunningTime="2025-09-12 22:53:48.209210385 +0000 UTC m=+57.778960550" watchObservedRunningTime="2025-09-12 22:53:48.359801413 +0000 UTC m=+57.929551579" Sep 12 22:53:48.374621 containerd[2017]: time="2025-09-12T22:53:48.374567826Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\" id:\"59014afdd3bf7d130be295e554779bc7540d5817d0c10f334a44db35791b375e\" pid:5831 exited_at:{seconds:1757717628 nanos:319230840}" Sep 12 22:53:50.286135 containerd[2017]: time="2025-09-12T22:53:50.286060952Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:50.287967 containerd[2017]: time="2025-09-12T22:53:50.287749627Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=47333864" Sep 12 22:53:50.290050 containerd[2017]: time="2025-09-12T22:53:50.290007435Z" level=info msg="ImageCreate event name:\"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:50.293597 containerd[2017]: time="2025-09-12T22:53:50.293483019Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:50.294047 containerd[2017]: time="2025-09-12T22:53:50.293983401Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 3.22458333s" Sep 12 22:53:50.294047 containerd[2017]: time="2025-09-12T22:53:50.294018203Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 12 22:53:50.296504 containerd[2017]: time="2025-09-12T22:53:50.296410125Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Sep 12 22:53:50.305316 containerd[2017]: time="2025-09-12T22:53:50.305272784Z" level=info msg="CreateContainer within sandbox \"8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 12 22:53:50.445614 containerd[2017]: time="2025-09-12T22:53:50.443379315Z" level=info msg="Container 07c0e8fb4861c4b993050fa1fddcdd8e151e017f045762ac436e98684fd8547f: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:50.500360 containerd[2017]: time="2025-09-12T22:53:50.499494262Z" level=info msg="CreateContainer within sandbox \"8e6042946da054b403a873660560db82cc1f4b30fb85abb2c532c802ac969022\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"07c0e8fb4861c4b993050fa1fddcdd8e151e017f045762ac436e98684fd8547f\"" Sep 12 22:53:50.502562 containerd[2017]: time="2025-09-12T22:53:50.502519933Z" level=info msg="StartContainer for \"07c0e8fb4861c4b993050fa1fddcdd8e151e017f045762ac436e98684fd8547f\"" Sep 12 22:53:50.571139 containerd[2017]: time="2025-09-12T22:53:50.570317120Z" level=info msg="connecting to shim 07c0e8fb4861c4b993050fa1fddcdd8e151e017f045762ac436e98684fd8547f" address="unix:///run/containerd/s/8a1053c7bb5f99f4164311cad50885aac214da7f33318db865ed07297e69945e" protocol=ttrpc version=3 Sep 12 22:53:50.622831 systemd[1]: Started cri-containerd-07c0e8fb4861c4b993050fa1fddcdd8e151e017f045762ac436e98684fd8547f.scope - libcontainer container 07c0e8fb4861c4b993050fa1fddcdd8e151e017f045762ac436e98684fd8547f. Sep 12 22:53:50.755665 containerd[2017]: time="2025-09-12T22:53:50.755618555Z" level=info msg="StartContainer for \"07c0e8fb4861c4b993050fa1fddcdd8e151e017f045762ac436e98684fd8547f\" returns successfully" Sep 12 22:53:53.028636 systemd[1]: Started sshd@11-172.31.29.105:22-139.178.89.65:45350.service - OpenSSH per-connection server daemon (139.178.89.65:45350). Sep 12 22:53:53.377038 sshd[5902]: Accepted publickey for core from 139.178.89.65 port 45350 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:53:53.383073 sshd-session[5902]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:53:53.396867 systemd-logind[1983]: New session 12 of user core. Sep 12 22:53:53.402918 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 12 22:53:53.866037 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount613832135.mount: Deactivated successfully. Sep 12 22:53:54.660076 kubelet[3352]: I0912 22:53:54.659819 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-5c7666b7bb-sbkxs" podStartSLOduration=37.590220178 podStartE2EDuration="47.659693477s" podCreationTimestamp="2025-09-12 22:53:07 +0000 UTC" firstStartedPulling="2025-09-12 22:53:40.225784216 +0000 UTC m=+49.795534362" lastFinishedPulling="2025-09-12 22:53:50.295257517 +0000 UTC m=+59.865007661" observedRunningTime="2025-09-12 22:53:51.224685675 +0000 UTC m=+60.794435842" watchObservedRunningTime="2025-09-12 22:53:54.659693477 +0000 UTC m=+64.229443663" Sep 12 22:53:54.924887 sshd[5905]: Connection closed by 139.178.89.65 port 45350 Sep 12 22:53:54.925954 sshd-session[5902]: pam_unix(sshd:session): session closed for user core Sep 12 22:53:54.939370 systemd[1]: sshd@11-172.31.29.105:22-139.178.89.65:45350.service: Deactivated successfully. Sep 12 22:53:54.948081 systemd[1]: session-12.scope: Deactivated successfully. Sep 12 22:53:54.959125 systemd-logind[1983]: Session 12 logged out. Waiting for processes to exit. Sep 12 22:53:54.984946 systemd[1]: Started sshd@12-172.31.29.105:22-139.178.89.65:45356.service - OpenSSH per-connection server daemon (139.178.89.65:45356). Sep 12 22:53:54.991506 systemd-logind[1983]: Removed session 12. Sep 12 22:53:55.264210 sshd[5927]: Accepted publickey for core from 139.178.89.65 port 45356 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:53:55.269175 sshd-session[5927]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:53:55.289588 systemd-logind[1983]: New session 13 of user core. Sep 12 22:53:55.292700 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 12 22:53:56.199945 containerd[2017]: time="2025-09-12T22:53:56.199651258Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:56.205597 containerd[2017]: time="2025-09-12T22:53:56.205177321Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=66357526" Sep 12 22:53:56.227249 containerd[2017]: time="2025-09-12T22:53:56.226437900Z" level=info msg="ImageCreate event name:\"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:56.232925 containerd[2017]: time="2025-09-12T22:53:56.232864185Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:56.235141 containerd[2017]: time="2025-09-12T22:53:56.234366540Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"66357372\" in 5.937880487s" Sep 12 22:53:56.235141 containerd[2017]: time="2025-09-12T22:53:56.234411221Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\"" Sep 12 22:53:56.358885 sshd[5930]: Connection closed by 139.178.89.65 port 45356 Sep 12 22:53:56.359332 containerd[2017]: time="2025-09-12T22:53:56.358875243Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Sep 12 22:53:56.362396 containerd[2017]: time="2025-09-12T22:53:56.362325091Z" level=info msg="CreateContainer within sandbox \"b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Sep 12 22:53:56.366787 sshd-session[5927]: pam_unix(sshd:session): session closed for user core Sep 12 22:53:56.400411 systemd[1]: sshd@12-172.31.29.105:22-139.178.89.65:45356.service: Deactivated successfully. Sep 12 22:53:56.407105 systemd[1]: session-13.scope: Deactivated successfully. Sep 12 22:53:56.413936 systemd-logind[1983]: Session 13 logged out. Waiting for processes to exit. Sep 12 22:53:56.420773 systemd-logind[1983]: Removed session 13. Sep 12 22:53:56.425612 systemd[1]: Started sshd@13-172.31.29.105:22-139.178.89.65:45360.service - OpenSSH per-connection server daemon (139.178.89.65:45360). Sep 12 22:53:56.512799 containerd[2017]: time="2025-09-12T22:53:56.511755350Z" level=info msg="Container e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:56.520156 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4004896411.mount: Deactivated successfully. Sep 12 22:53:56.616490 containerd[2017]: time="2025-09-12T22:53:56.616410025Z" level=info msg="CreateContainer within sandbox \"b7187d6895860f5c93944bf0fb7bff2d5a140c96cec9f98bb69b4ebfe7559554\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\"" Sep 12 22:53:56.633145 containerd[2017]: time="2025-09-12T22:53:56.633096232Z" level=info msg="StartContainer for \"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\"" Sep 12 22:53:56.638029 containerd[2017]: time="2025-09-12T22:53:56.637973136Z" level=info msg="connecting to shim e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add" address="unix:///run/containerd/s/e9ec3c350099cfcfa4473e9a4156504030d54ecf72057c526396a8c8a88ba657" protocol=ttrpc version=3 Sep 12 22:53:56.769649 sshd[5949]: Accepted publickey for core from 139.178.89.65 port 45360 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:53:56.776225 sshd-session[5949]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:53:56.791813 systemd-logind[1983]: New session 14 of user core. Sep 12 22:53:56.797748 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 12 22:53:56.842610 systemd[1]: Started cri-containerd-e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add.scope - libcontainer container e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add. Sep 12 22:53:57.063032 containerd[2017]: time="2025-09-12T22:53:57.062262700Z" level=info msg="StartContainer for \"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\" returns successfully" Sep 12 22:53:57.228860 sshd[5961]: Connection closed by 139.178.89.65 port 45360 Sep 12 22:53:57.229986 sshd-session[5949]: pam_unix(sshd:session): session closed for user core Sep 12 22:53:57.234920 systemd-logind[1983]: Session 14 logged out. Waiting for processes to exit. Sep 12 22:53:57.235684 systemd[1]: sshd@13-172.31.29.105:22-139.178.89.65:45360.service: Deactivated successfully. Sep 12 22:53:57.239435 systemd[1]: session-14.scope: Deactivated successfully. Sep 12 22:53:57.243548 systemd-logind[1983]: Removed session 14. Sep 12 22:53:58.185024 containerd[2017]: time="2025-09-12T22:53:58.183930752Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:58.186280 containerd[2017]: time="2025-09-12T22:53:58.186126436Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=14698542" Sep 12 22:53:58.188656 containerd[2017]: time="2025-09-12T22:53:58.188522182Z" level=info msg="ImageCreate event name:\"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:58.192634 containerd[2017]: time="2025-09-12T22:53:58.192595035Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:58.193433 containerd[2017]: time="2025-09-12T22:53:58.193197752Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"16191197\" in 1.834281956s" Sep 12 22:53:58.193433 containerd[2017]: time="2025-09-12T22:53:58.193389359Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\"" Sep 12 22:53:58.196305 containerd[2017]: time="2025-09-12T22:53:58.196255802Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Sep 12 22:53:58.205572 containerd[2017]: time="2025-09-12T22:53:58.205510927Z" level=info msg="CreateContainer within sandbox \"bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 12 22:53:58.240731 containerd[2017]: time="2025-09-12T22:53:58.240651911Z" level=info msg="Container 29f9f374a3a6b17c2d6066d179e219323bad65650781d427736e2ae453a66af2: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:58.264023 containerd[2017]: time="2025-09-12T22:53:58.263966507Z" level=info msg="CreateContainer within sandbox \"bce72deeeb413b851c30f60f7284dfb6fd0312b4046223b75bebc42b61d52e27\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"29f9f374a3a6b17c2d6066d179e219323bad65650781d427736e2ae453a66af2\"" Sep 12 22:53:58.265471 containerd[2017]: time="2025-09-12T22:53:58.264824687Z" level=info msg="StartContainer for \"29f9f374a3a6b17c2d6066d179e219323bad65650781d427736e2ae453a66af2\"" Sep 12 22:53:58.268290 containerd[2017]: time="2025-09-12T22:53:58.268214417Z" level=info msg="connecting to shim 29f9f374a3a6b17c2d6066d179e219323bad65650781d427736e2ae453a66af2" address="unix:///run/containerd/s/75f368a6ae892d71aae4fa58d4d8fe31ea2551143be0d7b587664a183ce637cd" protocol=ttrpc version=3 Sep 12 22:53:58.303797 systemd[1]: Started cri-containerd-29f9f374a3a6b17c2d6066d179e219323bad65650781d427736e2ae453a66af2.scope - libcontainer container 29f9f374a3a6b17c2d6066d179e219323bad65650781d427736e2ae453a66af2. Sep 12 22:53:58.377111 containerd[2017]: time="2025-09-12T22:53:58.377055460Z" level=info msg="StartContainer for \"29f9f374a3a6b17c2d6066d179e219323bad65650781d427736e2ae453a66af2\" returns successfully" Sep 12 22:53:58.580520 kubelet[3352]: I0912 22:53:58.580427 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-5nmmm" podStartSLOduration=25.790756328 podStartE2EDuration="46.566460465s" podCreationTimestamp="2025-09-12 22:53:12 +0000 UTC" firstStartedPulling="2025-09-12 22:53:37.420079192 +0000 UTC m=+46.989829338" lastFinishedPulling="2025-09-12 22:53:58.195783318 +0000 UTC m=+67.765533475" observedRunningTime="2025-09-12 22:53:58.552621841 +0000 UTC m=+68.122372007" watchObservedRunningTime="2025-09-12 22:53:58.566460465 +0000 UTC m=+68.136210624" Sep 12 22:53:58.581034 kubelet[3352]: I0912 22:53:58.580935 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-4265x" podStartSLOduration=30.804251988 podStartE2EDuration="46.580917793s" podCreationTimestamp="2025-09-12 22:53:12 +0000 UTC" firstStartedPulling="2025-09-12 22:53:40.496392064 +0000 UTC m=+50.066142219" lastFinishedPulling="2025-09-12 22:53:56.273057858 +0000 UTC m=+65.842808024" observedRunningTime="2025-09-12 22:53:57.680574683 +0000 UTC m=+67.250324849" watchObservedRunningTime="2025-09-12 22:53:58.580917793 +0000 UTC m=+68.150667959" Sep 12 22:53:58.651847 containerd[2017]: time="2025-09-12T22:53:58.651787993Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 12 22:53:58.653728 containerd[2017]: time="2025-09-12T22:53:58.653682294Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=77" Sep 12 22:53:58.656809 containerd[2017]: time="2025-09-12T22:53:58.656764971Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 460.376914ms" Sep 12 22:53:58.656809 containerd[2017]: time="2025-09-12T22:53:58.656816308Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Sep 12 22:53:58.673646 containerd[2017]: time="2025-09-12T22:53:58.673613194Z" level=info msg="CreateContainer within sandbox \"4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 12 22:53:58.703483 containerd[2017]: time="2025-09-12T22:53:58.701680476Z" level=info msg="Container 3ff69efd90640b1d9d1bcf37ddff78be8c78737b4b63af6dd8eccb368ac1096b: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:53:58.706387 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3595394447.mount: Deactivated successfully. Sep 12 22:53:58.726489 containerd[2017]: time="2025-09-12T22:53:58.726415502Z" level=info msg="CreateContainer within sandbox \"4e63017ebcc07f9ce1f4df2d2e292d8af4beba10d9dc3e2a49d70894d1696432\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"3ff69efd90640b1d9d1bcf37ddff78be8c78737b4b63af6dd8eccb368ac1096b\"" Sep 12 22:53:58.727719 containerd[2017]: time="2025-09-12T22:53:58.727582537Z" level=info msg="StartContainer for \"3ff69efd90640b1d9d1bcf37ddff78be8c78737b4b63af6dd8eccb368ac1096b\"" Sep 12 22:53:58.730564 containerd[2017]: time="2025-09-12T22:53:58.730518943Z" level=info msg="connecting to shim 3ff69efd90640b1d9d1bcf37ddff78be8c78737b4b63af6dd8eccb368ac1096b" address="unix:///run/containerd/s/9896e2c02032a0ff36f9eb1b99594a43a6b8e290bfb6917933b25ef9f1599257" protocol=ttrpc version=3 Sep 12 22:53:58.767711 systemd[1]: Started cri-containerd-3ff69efd90640b1d9d1bcf37ddff78be8c78737b4b63af6dd8eccb368ac1096b.scope - libcontainer container 3ff69efd90640b1d9d1bcf37ddff78be8c78737b4b63af6dd8eccb368ac1096b. Sep 12 22:53:58.905515 containerd[2017]: time="2025-09-12T22:53:58.905092220Z" level=info msg="StartContainer for \"3ff69efd90640b1d9d1bcf37ddff78be8c78737b4b63af6dd8eccb368ac1096b\" returns successfully" Sep 12 22:53:58.961918 containerd[2017]: time="2025-09-12T22:53:58.961883425Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\" id:\"af400d50d383c4231b6d74706f03f9b865b8ac89fc10369185ef093866b9a559\" pid:6048 exit_status:1 exited_at:{seconds:1757717638 nanos:961220418}" Sep 12 22:53:59.096476 kubelet[3352]: I0912 22:53:59.087577 3352 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 12 22:53:59.098072 kubelet[3352]: I0912 22:53:59.098030 3352 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 12 22:53:59.545692 kubelet[3352]: I0912 22:53:59.540918 3352 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-5c7666b7bb-d8qbk" podStartSLOduration=34.742612466 podStartE2EDuration="52.532084767s" podCreationTimestamp="2025-09-12 22:53:07 +0000 UTC" firstStartedPulling="2025-09-12 22:53:40.874579176 +0000 UTC m=+50.444329322" lastFinishedPulling="2025-09-12 22:53:58.664051476 +0000 UTC m=+68.233801623" observedRunningTime="2025-09-12 22:53:59.531992267 +0000 UTC m=+69.101742434" watchObservedRunningTime="2025-09-12 22:53:59.532084767 +0000 UTC m=+69.101834933" Sep 12 22:53:59.831161 containerd[2017]: time="2025-09-12T22:53:59.831024775Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\" id:\"1d9facc3dcf51fe9bf297bf045ef74423344fcb584b24c13cc2f55965f285146\" pid:6110 exit_status:1 exited_at:{seconds:1757717639 nanos:830697841}" Sep 12 22:54:00.704059 kubelet[3352]: I0912 22:54:00.701591 3352 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:54:02.274677 systemd[1]: Started sshd@14-172.31.29.105:22-139.178.89.65:34600.service - OpenSSH per-connection server daemon (139.178.89.65:34600). Sep 12 22:54:02.595774 sshd[6126]: Accepted publickey for core from 139.178.89.65 port 34600 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:02.604761 sshd-session[6126]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:02.615367 systemd-logind[1983]: New session 15 of user core. Sep 12 22:54:02.617704 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 12 22:54:04.025481 sshd[6130]: Connection closed by 139.178.89.65 port 34600 Sep 12 22:54:04.031708 sshd-session[6126]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:04.045313 systemd[1]: sshd@14-172.31.29.105:22-139.178.89.65:34600.service: Deactivated successfully. Sep 12 22:54:04.049048 systemd[1]: session-15.scope: Deactivated successfully. Sep 12 22:54:04.051537 systemd-logind[1983]: Session 15 logged out. Waiting for processes to exit. Sep 12 22:54:04.053749 systemd-logind[1983]: Removed session 15. Sep 12 22:54:05.550761 containerd[2017]: time="2025-09-12T22:54:05.550674463Z" level=info msg="TaskExit event in podsandbox handler container_id:\"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\" id:\"0fa3f65f6371565b4267496af845523b0bece1de296880c49c8db99465e45a11\" pid:6153 exited_at:{seconds:1757717645 nanos:548544974}" Sep 12 22:54:09.064535 systemd[1]: Started sshd@15-172.31.29.105:22-139.178.89.65:34614.service - OpenSSH per-connection server daemon (139.178.89.65:34614). Sep 12 22:54:09.371427 sshd[6167]: Accepted publickey for core from 139.178.89.65 port 34614 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:09.377933 sshd-session[6167]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:09.394510 systemd-logind[1983]: New session 16 of user core. Sep 12 22:54:09.397691 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 12 22:54:10.329597 sshd[6170]: Connection closed by 139.178.89.65 port 34614 Sep 12 22:54:10.330938 sshd-session[6167]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:10.339065 systemd[1]: sshd@15-172.31.29.105:22-139.178.89.65:34614.service: Deactivated successfully. Sep 12 22:54:10.343699 systemd[1]: session-16.scope: Deactivated successfully. Sep 12 22:54:10.347902 systemd-logind[1983]: Session 16 logged out. Waiting for processes to exit. Sep 12 22:54:10.351218 systemd-logind[1983]: Removed session 16. Sep 12 22:54:10.370500 systemd[1]: Started sshd@16-172.31.29.105:22-139.178.89.65:56754.service - OpenSSH per-connection server daemon (139.178.89.65:56754). Sep 12 22:54:10.574163 sshd[6182]: Accepted publickey for core from 139.178.89.65 port 56754 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:10.576227 sshd-session[6182]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:10.587076 systemd-logind[1983]: New session 17 of user core. Sep 12 22:54:10.593834 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 12 22:54:11.505528 sshd[6185]: Connection closed by 139.178.89.65 port 56754 Sep 12 22:54:11.508288 sshd-session[6182]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:11.514785 systemd[1]: sshd@16-172.31.29.105:22-139.178.89.65:56754.service: Deactivated successfully. Sep 12 22:54:11.519150 systemd[1]: session-17.scope: Deactivated successfully. Sep 12 22:54:11.526271 systemd-logind[1983]: Session 17 logged out. Waiting for processes to exit. Sep 12 22:54:11.545713 systemd[1]: Started sshd@17-172.31.29.105:22-139.178.89.65:56760.service - OpenSSH per-connection server daemon (139.178.89.65:56760). Sep 12 22:54:11.557739 systemd-logind[1983]: Removed session 17. Sep 12 22:54:11.832000 sshd[6196]: Accepted publickey for core from 139.178.89.65 port 56760 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:11.834184 sshd-session[6196]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:11.848735 systemd-logind[1983]: New session 18 of user core. Sep 12 22:54:11.855715 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 12 22:54:13.436484 sshd[6199]: Connection closed by 139.178.89.65 port 56760 Sep 12 22:54:13.436983 sshd-session[6196]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:13.449001 systemd[1]: sshd@17-172.31.29.105:22-139.178.89.65:56760.service: Deactivated successfully. Sep 12 22:54:13.452219 systemd[1]: session-18.scope: Deactivated successfully. Sep 12 22:54:13.457070 systemd-logind[1983]: Session 18 logged out. Waiting for processes to exit. Sep 12 22:54:13.474052 systemd[1]: Started sshd@18-172.31.29.105:22-139.178.89.65:56774.service - OpenSSH per-connection server daemon (139.178.89.65:56774). Sep 12 22:54:13.478368 systemd-logind[1983]: Removed session 18. Sep 12 22:54:13.711227 sshd[6216]: Accepted publickey for core from 139.178.89.65 port 56774 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:13.716029 sshd-session[6216]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:13.729430 systemd-logind[1983]: New session 19 of user core. Sep 12 22:54:13.735695 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 12 22:54:15.155310 sshd[6222]: Connection closed by 139.178.89.65 port 56774 Sep 12 22:54:15.156222 sshd-session[6216]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:15.162222 systemd[1]: sshd@18-172.31.29.105:22-139.178.89.65:56774.service: Deactivated successfully. Sep 12 22:54:15.169193 systemd[1]: session-19.scope: Deactivated successfully. Sep 12 22:54:15.174258 systemd-logind[1983]: Session 19 logged out. Waiting for processes to exit. Sep 12 22:54:15.195593 systemd[1]: Started sshd@19-172.31.29.105:22-139.178.89.65:56778.service - OpenSSH per-connection server daemon (139.178.89.65:56778). Sep 12 22:54:15.198707 systemd-logind[1983]: Removed session 19. Sep 12 22:54:15.408564 sshd[6232]: Accepted publickey for core from 139.178.89.65 port 56778 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:15.410758 sshd-session[6232]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:15.417025 systemd-logind[1983]: New session 20 of user core. Sep 12 22:54:15.427724 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 12 22:54:15.699557 sshd[6235]: Connection closed by 139.178.89.65 port 56778 Sep 12 22:54:15.700116 sshd-session[6232]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:15.703967 systemd-logind[1983]: Session 20 logged out. Waiting for processes to exit. Sep 12 22:54:15.706831 systemd[1]: sshd@19-172.31.29.105:22-139.178.89.65:56778.service: Deactivated successfully. Sep 12 22:54:15.709893 systemd[1]: session-20.scope: Deactivated successfully. Sep 12 22:54:15.713201 systemd-logind[1983]: Removed session 20. Sep 12 22:54:18.986745 containerd[2017]: time="2025-09-12T22:54:18.986693471Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\" id:\"f7af91a520b23938defe553a6f8f46c75a8a67253c0c2a152aaf47c7cf92c638\" pid:6259 exited_at:{seconds:1757717658 nanos:844299143}" Sep 12 22:54:20.737430 systemd[1]: Started sshd@20-172.31.29.105:22-139.178.89.65:33304.service - OpenSSH per-connection server daemon (139.178.89.65:33304). Sep 12 22:54:21.061414 sshd[6279]: Accepted publickey for core from 139.178.89.65 port 33304 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:21.071008 sshd-session[6279]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:21.083959 systemd-logind[1983]: New session 21 of user core. Sep 12 22:54:21.091697 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 12 22:54:21.899758 sshd[6282]: Connection closed by 139.178.89.65 port 33304 Sep 12 22:54:21.901857 sshd-session[6279]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:21.908886 systemd[1]: sshd@20-172.31.29.105:22-139.178.89.65:33304.service: Deactivated successfully. Sep 12 22:54:21.912886 systemd[1]: session-21.scope: Deactivated successfully. Sep 12 22:54:21.915259 systemd-logind[1983]: Session 21 logged out. Waiting for processes to exit. Sep 12 22:54:21.918060 systemd-logind[1983]: Removed session 21. Sep 12 22:54:24.462634 kubelet[3352]: I0912 22:54:24.462029 3352 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 12 22:54:26.935401 systemd[1]: Started sshd@21-172.31.29.105:22-139.178.89.65:33312.service - OpenSSH per-connection server daemon (139.178.89.65:33312). Sep 12 22:54:27.216497 sshd[6299]: Accepted publickey for core from 139.178.89.65 port 33312 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:27.218614 sshd-session[6299]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:27.225272 systemd-logind[1983]: New session 22 of user core. Sep 12 22:54:27.231366 systemd[1]: Started session-22.scope - Session 22 of User core. Sep 12 22:54:28.497313 sshd[6302]: Connection closed by 139.178.89.65 port 33312 Sep 12 22:54:28.505140 sshd-session[6299]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:28.520131 systemd[1]: sshd@21-172.31.29.105:22-139.178.89.65:33312.service: Deactivated successfully. Sep 12 22:54:28.525745 systemd[1]: session-22.scope: Deactivated successfully. Sep 12 22:54:28.532786 systemd-logind[1983]: Session 22 logged out. Waiting for processes to exit. Sep 12 22:54:28.535537 systemd-logind[1983]: Removed session 22. Sep 12 22:54:29.417328 containerd[2017]: time="2025-09-12T22:54:29.417283887Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\" id:\"95e1bcb6e9a013e4dff83d9aaaba080e2608f6d788893f726e992d470d850f1e\" pid:6331 exited_at:{seconds:1757717669 nanos:416537115}" Sep 12 22:54:29.709532 containerd[2017]: time="2025-09-12T22:54:29.709354794Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\" id:\"196af63a0d5bb74e0c2c5aeef87784c6ce9e9909138c37e267d2b94d16b4b154\" pid:6355 exited_at:{seconds:1757717669 nanos:708768444}" Sep 12 22:54:33.545016 systemd[1]: Started sshd@22-172.31.29.105:22-139.178.89.65:38610.service - OpenSSH per-connection server daemon (139.178.89.65:38610). Sep 12 22:54:33.819631 sshd[6368]: Accepted publickey for core from 139.178.89.65 port 38610 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:33.821604 sshd-session[6368]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:33.829942 systemd-logind[1983]: New session 23 of user core. Sep 12 22:54:33.838697 systemd[1]: Started session-23.scope - Session 23 of User core. Sep 12 22:54:34.576012 sshd[6371]: Connection closed by 139.178.89.65 port 38610 Sep 12 22:54:34.579054 sshd-session[6368]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:34.589930 systemd[1]: sshd@22-172.31.29.105:22-139.178.89.65:38610.service: Deactivated successfully. Sep 12 22:54:34.593918 systemd[1]: session-23.scope: Deactivated successfully. Sep 12 22:54:34.598865 systemd-logind[1983]: Session 23 logged out. Waiting for processes to exit. Sep 12 22:54:34.600906 systemd-logind[1983]: Removed session 23. Sep 12 22:54:36.041373 containerd[2017]: time="2025-09-12T22:54:36.041323436Z" level=info msg="TaskExit event in podsandbox handler container_id:\"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\" id:\"3cec141376089064caf0dc61dd24b9c35fa8b29b3d1e4b5ae6a14edd0bfceca7\" pid:6394 exited_at:{seconds:1757717676 nanos:40830985}" Sep 12 22:54:39.622671 systemd[1]: Started sshd@23-172.31.29.105:22-139.178.89.65:38622.service - OpenSSH per-connection server daemon (139.178.89.65:38622). Sep 12 22:54:39.951681 sshd[6409]: Accepted publickey for core from 139.178.89.65 port 38622 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:39.957918 sshd-session[6409]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:39.966826 systemd-logind[1983]: New session 24 of user core. Sep 12 22:54:39.974957 systemd[1]: Started session-24.scope - Session 24 of User core. Sep 12 22:54:41.208075 sshd[6412]: Connection closed by 139.178.89.65 port 38622 Sep 12 22:54:41.210055 sshd-session[6409]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:41.216539 systemd-logind[1983]: Session 24 logged out. Waiting for processes to exit. Sep 12 22:54:41.218187 systemd[1]: sshd@23-172.31.29.105:22-139.178.89.65:38622.service: Deactivated successfully. Sep 12 22:54:41.222918 systemd[1]: session-24.scope: Deactivated successfully. Sep 12 22:54:41.228866 systemd-logind[1983]: Removed session 24. Sep 12 22:54:46.245761 systemd[1]: Started sshd@24-172.31.29.105:22-139.178.89.65:45854.service - OpenSSH per-connection server daemon (139.178.89.65:45854). Sep 12 22:54:46.452795 sshd[6425]: Accepted publickey for core from 139.178.89.65 port 45854 ssh2: RSA SHA256:6Cuckp9cFHLH3NTfBl1U/KSLCHTjBmHBde3uKlxnZHc Sep 12 22:54:46.456078 sshd-session[6425]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Sep 12 22:54:46.463551 systemd-logind[1983]: New session 25 of user core. Sep 12 22:54:46.470904 systemd[1]: Started session-25.scope - Session 25 of User core. Sep 12 22:54:47.038849 sshd[6428]: Connection closed by 139.178.89.65 port 45854 Sep 12 22:54:47.038745 sshd-session[6425]: pam_unix(sshd:session): session closed for user core Sep 12 22:54:47.049036 systemd[1]: sshd@24-172.31.29.105:22-139.178.89.65:45854.service: Deactivated successfully. Sep 12 22:54:47.050556 systemd-logind[1983]: Session 25 logged out. Waiting for processes to exit. Sep 12 22:54:47.056509 systemd[1]: session-25.scope: Deactivated successfully. Sep 12 22:54:47.063093 systemd-logind[1983]: Removed session 25. Sep 12 22:54:48.293918 containerd[2017]: time="2025-09-12T22:54:48.293863763Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\" id:\"45e5658a6b51602f17cf34b57a9dac8758d66096d1ffcfaef893fbf1f06edf7d\" pid:6452 exited_at:{seconds:1757717688 nanos:267144336}" Sep 12 22:54:49.630560 containerd[2017]: time="2025-09-12T22:54:49.630363807Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\" id:\"0f9325eb9ad7519a53d95c094ab11070c678ad4a605dba776418af9a9c42ff8f\" pid:6475 exited_at:{seconds:1757717689 nanos:629022164}" Sep 12 22:54:59.686508 containerd[2017]: time="2025-09-12T22:54:59.675577659Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e60b006901b0818c46889c7fdfc642599d994b9c8d16b765c58090eef83e5add\" id:\"1dd9fde2630bc73206ad476a4178f113df31594c4c7c9013a1c99b3ad78edff1\" pid:6507 exited_at:{seconds:1757717699 nanos:675128854}" Sep 12 22:55:00.937315 systemd[1]: cri-containerd-efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544.scope: Deactivated successfully. Sep 12 22:55:00.937753 systemd[1]: cri-containerd-efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544.scope: Consumed 15.319s CPU time, 102.7M memory peak, 80.5M read from disk. Sep 12 22:55:01.035766 containerd[2017]: time="2025-09-12T22:55:01.035712157Z" level=info msg="TaskExit event in podsandbox handler container_id:\"efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544\" id:\"efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544\" pid:3930 exit_status:1 exited_at:{seconds:1757717701 nanos:6232209}" Sep 12 22:55:01.046955 containerd[2017]: time="2025-09-12T22:55:01.046490469Z" level=info msg="received exit event container_id:\"efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544\" id:\"efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544\" pid:3930 exit_status:1 exited_at:{seconds:1757717701 nanos:6232209}" Sep 12 22:55:01.200345 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544-rootfs.mount: Deactivated successfully. Sep 12 22:55:02.107135 systemd[1]: cri-containerd-2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a.scope: Deactivated successfully. Sep 12 22:55:02.107560 systemd[1]: cri-containerd-2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a.scope: Consumed 4.331s CPU time, 89.8M memory peak, 113.4M read from disk. Sep 12 22:55:02.221590 containerd[2017]: time="2025-09-12T22:55:02.221370097Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a\" id:\"2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a\" pid:3197 exit_status:1 exited_at:{seconds:1757717702 nanos:185521591}" Sep 12 22:55:02.222773 containerd[2017]: time="2025-09-12T22:55:02.222734597Z" level=info msg="received exit event container_id:\"2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a\" id:\"2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a\" pid:3197 exit_status:1 exited_at:{seconds:1757717702 nanos:185521591}" Sep 12 22:55:02.270987 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a-rootfs.mount: Deactivated successfully. Sep 12 22:55:02.403969 kubelet[3352]: I0912 22:55:02.403818 3352 scope.go:117] "RemoveContainer" containerID="f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f" Sep 12 22:55:02.406989 kubelet[3352]: I0912 22:55:02.404501 3352 scope.go:117] "RemoveContainer" containerID="efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544" Sep 12 22:55:02.435727 kubelet[3352]: E0912 22:55:02.435652 3352 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tigera-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=tigera-operator pod=tigera-operator-755d956888-jj6hp_tigera-operator(8fa588a9-f073-4493-9c89-a0a42ba66c3f)\"" pod="tigera-operator/tigera-operator-755d956888-jj6hp" podUID="8fa588a9-f073-4493-9c89-a0a42ba66c3f" Sep 12 22:55:02.579662 containerd[2017]: time="2025-09-12T22:55:02.579602211Z" level=info msg="RemoveContainer for \"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\"" Sep 12 22:55:02.631933 containerd[2017]: time="2025-09-12T22:55:02.631868308Z" level=info msg="RemoveContainer for \"f3d6ee8bf415e49fe6d7ec99e3e05ac69ab8bc6643ee18bc0752c57b50c0bb8f\" returns successfully" Sep 12 22:55:03.323785 kubelet[3352]: I0912 22:55:03.323697 3352 scope.go:117] "RemoveContainer" containerID="2e3aea56269aed5c51f521c41d18a073d1343cdd77ba5c4bd714b68ee603357a" Sep 12 22:55:03.436997 containerd[2017]: time="2025-09-12T22:55:03.436927318Z" level=info msg="CreateContainer within sandbox \"ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Sep 12 22:55:03.522381 containerd[2017]: time="2025-09-12T22:55:03.522321810Z" level=info msg="Container 23a232736e6025eab64746741f34a7b577238d927bb507c13936f19e2f0591d1: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:55:03.524339 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1479384905.mount: Deactivated successfully. Sep 12 22:55:03.545024 containerd[2017]: time="2025-09-12T22:55:03.544855248Z" level=info msg="CreateContainer within sandbox \"ee82aa961d13dab4a6baf3fce85f3e57953227fefbb3494e34cd01a2f0f49714\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"23a232736e6025eab64746741f34a7b577238d927bb507c13936f19e2f0591d1\"" Sep 12 22:55:03.545542 containerd[2017]: time="2025-09-12T22:55:03.545514700Z" level=info msg="StartContainer for \"23a232736e6025eab64746741f34a7b577238d927bb507c13936f19e2f0591d1\"" Sep 12 22:55:03.546805 containerd[2017]: time="2025-09-12T22:55:03.546763975Z" level=info msg="connecting to shim 23a232736e6025eab64746741f34a7b577238d927bb507c13936f19e2f0591d1" address="unix:///run/containerd/s/077aeb1a6486decd148f8a1d701e244608fdb91e969d1e782287aee03568179f" protocol=ttrpc version=3 Sep 12 22:55:03.643298 systemd[1]: Started cri-containerd-23a232736e6025eab64746741f34a7b577238d927bb507c13936f19e2f0591d1.scope - libcontainer container 23a232736e6025eab64746741f34a7b577238d927bb507c13936f19e2f0591d1. Sep 12 22:55:03.763181 containerd[2017]: time="2025-09-12T22:55:03.763101178Z" level=info msg="StartContainer for \"23a232736e6025eab64746741f34a7b577238d927bb507c13936f19e2f0591d1\" returns successfully" Sep 12 22:55:04.081054 kubelet[3352]: E0912 22:55:04.080978 3352 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.105:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-105?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 12 22:55:05.279720 containerd[2017]: time="2025-09-12T22:55:05.279670166Z" level=info msg="TaskExit event in podsandbox handler container_id:\"88f73c7543ae1ce55bf65c865b1a87ce55bff9d53db7fa65a5d5133af1185ec0\" id:\"abb6505bc55ff5161bcee63520eb1bd0dd52bdbbff37cb2613b3a03f68c12e5f\" pid:6584 exited_at:{seconds:1757717705 nanos:278667639}" Sep 12 22:55:06.658879 systemd[1]: cri-containerd-2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3.scope: Deactivated successfully. Sep 12 22:55:06.659145 systemd[1]: cri-containerd-2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3.scope: Consumed 2.630s CPU time, 39.7M memory peak, 65.6M read from disk. Sep 12 22:55:06.662163 containerd[2017]: time="2025-09-12T22:55:06.662051873Z" level=info msg="received exit event container_id:\"2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3\" id:\"2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3\" pid:3188 exit_status:1 exited_at:{seconds:1757717706 nanos:661355382}" Sep 12 22:55:06.662542 containerd[2017]: time="2025-09-12T22:55:06.662167480Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3\" id:\"2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3\" pid:3188 exit_status:1 exited_at:{seconds:1757717706 nanos:661355382}" Sep 12 22:55:06.696135 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3-rootfs.mount: Deactivated successfully. Sep 12 22:55:07.340886 kubelet[3352]: I0912 22:55:07.340846 3352 scope.go:117] "RemoveContainer" containerID="2f8316da5d87bf38ad242c23b4c385ccfe13209a47c8ff095b3f53b0652096f3" Sep 12 22:55:07.343960 containerd[2017]: time="2025-09-12T22:55:07.343922615Z" level=info msg="CreateContainer within sandbox \"9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Sep 12 22:55:07.368493 containerd[2017]: time="2025-09-12T22:55:07.365539237Z" level=info msg="Container dab3c0531e2d123e342bb05bb4b383e43f0f3a9e5a42ffd2644b47ce1effe21b: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:55:07.384634 containerd[2017]: time="2025-09-12T22:55:07.384575115Z" level=info msg="CreateContainer within sandbox \"9722f9e8bc3f08256f7da7b469179c7cf9300dd47cf6f876368f8cc677ce046f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"dab3c0531e2d123e342bb05bb4b383e43f0f3a9e5a42ffd2644b47ce1effe21b\"" Sep 12 22:55:07.385401 containerd[2017]: time="2025-09-12T22:55:07.385356383Z" level=info msg="StartContainer for \"dab3c0531e2d123e342bb05bb4b383e43f0f3a9e5a42ffd2644b47ce1effe21b\"" Sep 12 22:55:07.386951 containerd[2017]: time="2025-09-12T22:55:07.386904014Z" level=info msg="connecting to shim dab3c0531e2d123e342bb05bb4b383e43f0f3a9e5a42ffd2644b47ce1effe21b" address="unix:///run/containerd/s/c988aa1ce1fcf3f05a530c7d8c38cd9a08fd6181e931f806eda982958cdf9055" protocol=ttrpc version=3 Sep 12 22:55:07.417789 systemd[1]: Started cri-containerd-dab3c0531e2d123e342bb05bb4b383e43f0f3a9e5a42ffd2644b47ce1effe21b.scope - libcontainer container dab3c0531e2d123e342bb05bb4b383e43f0f3a9e5a42ffd2644b47ce1effe21b. Sep 12 22:55:07.489431 containerd[2017]: time="2025-09-12T22:55:07.489319825Z" level=info msg="StartContainer for \"dab3c0531e2d123e342bb05bb4b383e43f0f3a9e5a42ffd2644b47ce1effe21b\" returns successfully" Sep 12 22:55:14.082571 kubelet[3352]: E0912 22:55:14.082029 3352 controller.go:195] "Failed to update lease" err="Put \"https://172.31.29.105:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-29-105?timeout=10s\": context deadline exceeded" Sep 12 22:55:14.615045 kubelet[3352]: I0912 22:55:14.614963 3352 scope.go:117] "RemoveContainer" containerID="efd8d58c9263aed0056faa4658d19d82c8ca05654a680bbe08667e5b57a95544" Sep 12 22:55:14.620429 containerd[2017]: time="2025-09-12T22:55:14.620373801Z" level=info msg="CreateContainer within sandbox \"aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046\" for container &ContainerMetadata{Name:tigera-operator,Attempt:2,}" Sep 12 22:55:14.638320 containerd[2017]: time="2025-09-12T22:55:14.637765391Z" level=info msg="Container aa840f69ae4eca79b087966103fddb716a3e06207122eac083ca94df0b11d267: CDI devices from CRI Config.CDIDevices: []" Sep 12 22:55:14.659157 containerd[2017]: time="2025-09-12T22:55:14.659114644Z" level=info msg="CreateContainer within sandbox \"aafaa77a0ba47117f72541c500e61b58eb7e4f09547bed380c40e306c552f046\" for &ContainerMetadata{Name:tigera-operator,Attempt:2,} returns container id \"aa840f69ae4eca79b087966103fddb716a3e06207122eac083ca94df0b11d267\"" Sep 12 22:55:14.662695 containerd[2017]: time="2025-09-12T22:55:14.662647649Z" level=info msg="StartContainer for \"aa840f69ae4eca79b087966103fddb716a3e06207122eac083ca94df0b11d267\"" Sep 12 22:55:14.664474 containerd[2017]: time="2025-09-12T22:55:14.664190377Z" level=info msg="connecting to shim aa840f69ae4eca79b087966103fddb716a3e06207122eac083ca94df0b11d267" address="unix:///run/containerd/s/6e4ad363abbdf3e512a4e72be044a5595373f9ec95405b2e2b35491f25276544" protocol=ttrpc version=3 Sep 12 22:55:14.704703 systemd[1]: Started cri-containerd-aa840f69ae4eca79b087966103fddb716a3e06207122eac083ca94df0b11d267.scope - libcontainer container aa840f69ae4eca79b087966103fddb716a3e06207122eac083ca94df0b11d267. Sep 12 22:55:14.755737 containerd[2017]: time="2025-09-12T22:55:14.755675906Z" level=info msg="StartContainer for \"aa840f69ae4eca79b087966103fddb716a3e06207122eac083ca94df0b11d267\" returns successfully" Sep 12 22:55:18.237715 containerd[2017]: time="2025-09-12T22:55:18.237659481Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bf61a8a4dea693df1d9b3e3e75125d7af12eb42ab82a06fa5458e0db1d00f0a3\" id:\"e88fe0fe3f1e1ce5d27e3645baf7f2769831ab18040374d059bf9e4979bbfb64\" pid:6705 exit_status:1 exited_at:{seconds:1757717718 nanos:237098186}"