Oct 13 05:51:09.928225 kernel: Linux version 6.12.51-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT_DYNAMIC Sun Oct 12 22:37:12 -00 2025 Oct 13 05:51:09.928264 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=a48d469b0deb49c328e6faf6cf366b11952d47f2d24963c866a0ea8221fb0039 Oct 13 05:51:09.928279 kernel: BIOS-provided physical RAM map: Oct 13 05:51:09.928290 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Oct 13 05:51:09.928300 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable Oct 13 05:51:09.928311 kernel: BIOS-e820: [mem 0x00000000786ce000-0x000000007894dfff] reserved Oct 13 05:51:09.928325 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Oct 13 05:51:09.928336 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Oct 13 05:51:09.928350 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable Oct 13 05:51:09.928362 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Oct 13 05:51:09.928373 kernel: NX (Execute Disable) protection: active Oct 13 05:51:09.928385 kernel: APIC: Static calls initialized Oct 13 05:51:09.928398 kernel: e820: update [mem 0x768c0018-0x768c8e57] usable ==> usable Oct 13 05:51:09.928412 kernel: extended physical RAM map: Oct 13 05:51:09.928430 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Oct 13 05:51:09.928444 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000768c0017] usable Oct 13 05:51:09.928459 kernel: reserve setup_data: [mem 0x00000000768c0018-0x00000000768c8e57] usable Oct 13 05:51:09.928473 kernel: reserve setup_data: [mem 0x00000000768c8e58-0x00000000786cdfff] usable Oct 13 05:51:09.928487 kernel: reserve setup_data: [mem 0x00000000786ce000-0x000000007894dfff] reserved Oct 13 05:51:09.928501 kernel: reserve setup_data: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Oct 13 05:51:09.928515 kernel: reserve setup_data: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Oct 13 05:51:09.928529 kernel: reserve setup_data: [mem 0x00000000789de000-0x000000007c97bfff] usable Oct 13 05:51:09.928543 kernel: reserve setup_data: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Oct 13 05:51:09.928557 kernel: efi: EFI v2.7 by EDK II Oct 13 05:51:09.928575 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77003518 Oct 13 05:51:09.928590 kernel: secureboot: Secure boot disabled Oct 13 05:51:09.928605 kernel: SMBIOS 2.7 present. Oct 13 05:51:09.928620 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Oct 13 05:51:09.928634 kernel: DMI: Memory slots populated: 1/1 Oct 13 05:51:09.928648 kernel: Hypervisor detected: KVM Oct 13 05:51:09.928662 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Oct 13 05:51:09.928676 kernel: kvm-clock: using sched offset of 5696568451 cycles Oct 13 05:51:09.928692 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Oct 13 05:51:09.928706 kernel: tsc: Detected 2499.992 MHz processor Oct 13 05:51:09.928721 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Oct 13 05:51:09.928739 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Oct 13 05:51:09.928753 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 Oct 13 05:51:09.928768 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Oct 13 05:51:09.928802 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Oct 13 05:51:09.928815 kernel: Using GB pages for direct mapping Oct 13 05:51:09.928834 kernel: ACPI: Early table checksum verification disabled Oct 13 05:51:09.928850 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) Oct 13 05:51:09.928864 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) Oct 13 05:51:09.928879 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Oct 13 05:51:09.928892 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Oct 13 05:51:09.928906 kernel: ACPI: FACS 0x00000000789D0000 000040 Oct 13 05:51:09.928920 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Oct 13 05:51:09.928932 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Oct 13 05:51:09.928946 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Oct 13 05:51:09.928964 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Oct 13 05:51:09.928977 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Oct 13 05:51:09.928990 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Oct 13 05:51:09.929005 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Oct 13 05:51:09.929020 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) Oct 13 05:51:09.929035 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] Oct 13 05:51:09.929051 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] Oct 13 05:51:09.929066 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] Oct 13 05:51:09.929085 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] Oct 13 05:51:09.929100 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] Oct 13 05:51:09.929115 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] Oct 13 05:51:09.929130 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] Oct 13 05:51:09.929145 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] Oct 13 05:51:09.929158 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] Oct 13 05:51:09.929170 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] Oct 13 05:51:09.929183 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] Oct 13 05:51:09.929197 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Oct 13 05:51:09.929210 kernel: NUMA: Initialized distance table, cnt=1 Oct 13 05:51:09.929226 kernel: NODE_DATA(0) allocated [mem 0x7a8eddc0-0x7a8f4fff] Oct 13 05:51:09.929239 kernel: Zone ranges: Oct 13 05:51:09.929253 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Oct 13 05:51:09.929267 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] Oct 13 05:51:09.929280 kernel: Normal empty Oct 13 05:51:09.929293 kernel: Device empty Oct 13 05:51:09.929307 kernel: Movable zone start for each node Oct 13 05:51:09.929320 kernel: Early memory node ranges Oct 13 05:51:09.929334 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Oct 13 05:51:09.929350 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] Oct 13 05:51:09.929363 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] Oct 13 05:51:09.929376 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] Oct 13 05:51:09.929390 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Oct 13 05:51:09.929403 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Oct 13 05:51:09.929417 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Oct 13 05:51:09.929431 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges Oct 13 05:51:09.929445 kernel: ACPI: PM-Timer IO Port: 0xb008 Oct 13 05:51:09.929459 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Oct 13 05:51:09.929475 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Oct 13 05:51:09.929489 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Oct 13 05:51:09.929503 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Oct 13 05:51:09.929517 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Oct 13 05:51:09.929531 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Oct 13 05:51:09.929544 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Oct 13 05:51:09.929558 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Oct 13 05:51:09.929572 kernel: TSC deadline timer available Oct 13 05:51:09.929586 kernel: CPU topo: Max. logical packages: 1 Oct 13 05:51:09.929600 kernel: CPU topo: Max. logical dies: 1 Oct 13 05:51:09.929616 kernel: CPU topo: Max. dies per package: 1 Oct 13 05:51:09.929630 kernel: CPU topo: Max. threads per core: 2 Oct 13 05:51:09.929644 kernel: CPU topo: Num. cores per package: 1 Oct 13 05:51:09.929658 kernel: CPU topo: Num. threads per package: 2 Oct 13 05:51:09.929671 kernel: CPU topo: Allowing 2 present CPUs plus 0 hotplug CPUs Oct 13 05:51:09.929686 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Oct 13 05:51:09.929700 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices Oct 13 05:51:09.929713 kernel: Booting paravirtualized kernel on KVM Oct 13 05:51:09.929728 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Oct 13 05:51:09.929744 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Oct 13 05:51:09.929759 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u1048576 Oct 13 05:51:09.929773 kernel: pcpu-alloc: s207832 r8192 d29736 u1048576 alloc=1*2097152 Oct 13 05:51:09.932840 kernel: pcpu-alloc: [0] 0 1 Oct 13 05:51:09.932857 kernel: kvm-guest: PV spinlocks enabled Oct 13 05:51:09.932874 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Oct 13 05:51:09.932892 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=a48d469b0deb49c328e6faf6cf366b11952d47f2d24963c866a0ea8221fb0039 Oct 13 05:51:09.932908 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Oct 13 05:51:09.932928 kernel: random: crng init done Oct 13 05:51:09.932943 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Oct 13 05:51:09.932958 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Oct 13 05:51:09.932974 kernel: Fallback order for Node 0: 0 Oct 13 05:51:09.932989 kernel: Built 1 zonelists, mobility grouping on. Total pages: 509451 Oct 13 05:51:09.933004 kernel: Policy zone: DMA32 Oct 13 05:51:09.933034 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Oct 13 05:51:09.933050 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Oct 13 05:51:09.933066 kernel: Kernel/User page tables isolation: enabled Oct 13 05:51:09.933082 kernel: ftrace: allocating 40139 entries in 157 pages Oct 13 05:51:09.933098 kernel: ftrace: allocated 157 pages with 5 groups Oct 13 05:51:09.933116 kernel: Dynamic Preempt: voluntary Oct 13 05:51:09.933132 kernel: rcu: Preemptible hierarchical RCU implementation. Oct 13 05:51:09.933149 kernel: rcu: RCU event tracing is enabled. Oct 13 05:51:09.933165 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Oct 13 05:51:09.933181 kernel: Trampoline variant of Tasks RCU enabled. Oct 13 05:51:09.933198 kernel: Rude variant of Tasks RCU enabled. Oct 13 05:51:09.933216 kernel: Tracing variant of Tasks RCU enabled. Oct 13 05:51:09.933232 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Oct 13 05:51:09.933248 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Oct 13 05:51:09.933264 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 13 05:51:09.933280 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 13 05:51:09.933296 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Oct 13 05:51:09.933313 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Oct 13 05:51:09.933329 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Oct 13 05:51:09.933347 kernel: Console: colour dummy device 80x25 Oct 13 05:51:09.933363 kernel: printk: legacy console [tty0] enabled Oct 13 05:51:09.933379 kernel: printk: legacy console [ttyS0] enabled Oct 13 05:51:09.933395 kernel: ACPI: Core revision 20240827 Oct 13 05:51:09.933412 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Oct 13 05:51:09.933428 kernel: APIC: Switch to symmetric I/O mode setup Oct 13 05:51:09.933444 kernel: x2apic enabled Oct 13 05:51:09.933460 kernel: APIC: Switched APIC routing to: physical x2apic Oct 13 05:51:09.933476 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093255d7c, max_idle_ns: 440795319144 ns Oct 13 05:51:09.933495 kernel: Calibrating delay loop (skipped) preset value.. 4999.98 BogoMIPS (lpj=2499992) Oct 13 05:51:09.933511 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Oct 13 05:51:09.933527 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Oct 13 05:51:09.933543 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Oct 13 05:51:09.933558 kernel: Spectre V2 : Mitigation: Retpolines Oct 13 05:51:09.933574 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Oct 13 05:51:09.933590 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Oct 13 05:51:09.933606 kernel: RETBleed: Vulnerable Oct 13 05:51:09.933621 kernel: Speculative Store Bypass: Vulnerable Oct 13 05:51:09.933637 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Oct 13 05:51:09.933652 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Oct 13 05:51:09.933671 kernel: GDS: Unknown: Dependent on hypervisor status Oct 13 05:51:09.933686 kernel: active return thunk: its_return_thunk Oct 13 05:51:09.933702 kernel: ITS: Mitigation: Aligned branch/return thunks Oct 13 05:51:09.933717 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Oct 13 05:51:09.933733 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Oct 13 05:51:09.933749 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Oct 13 05:51:09.933764 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Oct 13 05:51:09.933794 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Oct 13 05:51:09.933807 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Oct 13 05:51:09.933820 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Oct 13 05:51:09.934837 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Oct 13 05:51:09.934866 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Oct 13 05:51:09.934881 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Oct 13 05:51:09.934897 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Oct 13 05:51:09.934911 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Oct 13 05:51:09.934926 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Oct 13 05:51:09.934941 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Oct 13 05:51:09.934957 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Oct 13 05:51:09.934973 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Oct 13 05:51:09.934990 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Oct 13 05:51:09.935007 kernel: Freeing SMP alternatives memory: 32K Oct 13 05:51:09.935024 kernel: pid_max: default: 32768 minimum: 301 Oct 13 05:51:09.935044 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Oct 13 05:51:09.935061 kernel: landlock: Up and running. Oct 13 05:51:09.935077 kernel: SELinux: Initializing. Oct 13 05:51:09.935094 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 13 05:51:09.935112 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Oct 13 05:51:09.935128 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) Oct 13 05:51:09.935146 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Oct 13 05:51:09.935162 kernel: signal: max sigframe size: 3632 Oct 13 05:51:09.935178 kernel: rcu: Hierarchical SRCU implementation. Oct 13 05:51:09.935196 kernel: rcu: Max phase no-delay instances is 400. Oct 13 05:51:09.935216 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Oct 13 05:51:09.935233 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Oct 13 05:51:09.935250 kernel: smp: Bringing up secondary CPUs ... Oct 13 05:51:09.935267 kernel: smpboot: x86: Booting SMP configuration: Oct 13 05:51:09.935283 kernel: .... node #0, CPUs: #1 Oct 13 05:51:09.935301 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Oct 13 05:51:09.935320 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Oct 13 05:51:09.935335 kernel: smp: Brought up 1 node, 2 CPUs Oct 13 05:51:09.935349 kernel: smpboot: Total of 2 processors activated (9999.96 BogoMIPS) Oct 13 05:51:09.935365 kernel: Memory: 1908048K/2037804K available (14336K kernel code, 2443K rwdata, 10000K rodata, 54096K init, 2852K bss, 125192K reserved, 0K cma-reserved) Oct 13 05:51:09.935380 kernel: devtmpfs: initialized Oct 13 05:51:09.935397 kernel: x86/mm: Memory block size: 128MB Oct 13 05:51:09.935414 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) Oct 13 05:51:09.935431 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Oct 13 05:51:09.935457 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Oct 13 05:51:09.935474 kernel: pinctrl core: initialized pinctrl subsystem Oct 13 05:51:09.935491 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Oct 13 05:51:09.935508 kernel: audit: initializing netlink subsys (disabled) Oct 13 05:51:09.935528 kernel: audit: type=2000 audit(1760334667.336:1): state=initialized audit_enabled=0 res=1 Oct 13 05:51:09.935545 kernel: thermal_sys: Registered thermal governor 'step_wise' Oct 13 05:51:09.935562 kernel: thermal_sys: Registered thermal governor 'user_space' Oct 13 05:51:09.935579 kernel: cpuidle: using governor menu Oct 13 05:51:09.935595 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Oct 13 05:51:09.935612 kernel: dca service started, version 1.12.1 Oct 13 05:51:09.935628 kernel: PCI: Using configuration type 1 for base access Oct 13 05:51:09.935646 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Oct 13 05:51:09.935662 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Oct 13 05:51:09.935682 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Oct 13 05:51:09.935696 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Oct 13 05:51:09.935711 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Oct 13 05:51:09.935725 kernel: ACPI: Added _OSI(Module Device) Oct 13 05:51:09.939496 kernel: ACPI: Added _OSI(Processor Device) Oct 13 05:51:09.939524 kernel: ACPI: Added _OSI(Processor Aggregator Device) Oct 13 05:51:09.939541 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Oct 13 05:51:09.939557 kernel: ACPI: Interpreter enabled Oct 13 05:51:09.939573 kernel: ACPI: PM: (supports S0 S5) Oct 13 05:51:09.939596 kernel: ACPI: Using IOAPIC for interrupt routing Oct 13 05:51:09.939613 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Oct 13 05:51:09.939629 kernel: PCI: Using E820 reservations for host bridge windows Oct 13 05:51:09.939644 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Oct 13 05:51:09.939661 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Oct 13 05:51:09.939958 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Oct 13 05:51:09.940105 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Oct 13 05:51:09.940248 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Oct 13 05:51:09.940268 kernel: acpiphp: Slot [3] registered Oct 13 05:51:09.940284 kernel: acpiphp: Slot [4] registered Oct 13 05:51:09.940300 kernel: acpiphp: Slot [5] registered Oct 13 05:51:09.940316 kernel: acpiphp: Slot [6] registered Oct 13 05:51:09.940331 kernel: acpiphp: Slot [7] registered Oct 13 05:51:09.940347 kernel: acpiphp: Slot [8] registered Oct 13 05:51:09.940363 kernel: acpiphp: Slot [9] registered Oct 13 05:51:09.940379 kernel: acpiphp: Slot [10] registered Oct 13 05:51:09.940398 kernel: acpiphp: Slot [11] registered Oct 13 05:51:09.940414 kernel: acpiphp: Slot [12] registered Oct 13 05:51:09.940430 kernel: acpiphp: Slot [13] registered Oct 13 05:51:09.940446 kernel: acpiphp: Slot [14] registered Oct 13 05:51:09.940461 kernel: acpiphp: Slot [15] registered Oct 13 05:51:09.940477 kernel: acpiphp: Slot [16] registered Oct 13 05:51:09.940493 kernel: acpiphp: Slot [17] registered Oct 13 05:51:09.940509 kernel: acpiphp: Slot [18] registered Oct 13 05:51:09.940524 kernel: acpiphp: Slot [19] registered Oct 13 05:51:09.940540 kernel: acpiphp: Slot [20] registered Oct 13 05:51:09.940558 kernel: acpiphp: Slot [21] registered Oct 13 05:51:09.940574 kernel: acpiphp: Slot [22] registered Oct 13 05:51:09.940589 kernel: acpiphp: Slot [23] registered Oct 13 05:51:09.940605 kernel: acpiphp: Slot [24] registered Oct 13 05:51:09.940621 kernel: acpiphp: Slot [25] registered Oct 13 05:51:09.940636 kernel: acpiphp: Slot [26] registered Oct 13 05:51:09.940652 kernel: acpiphp: Slot [27] registered Oct 13 05:51:09.940667 kernel: acpiphp: Slot [28] registered Oct 13 05:51:09.940683 kernel: acpiphp: Slot [29] registered Oct 13 05:51:09.940701 kernel: acpiphp: Slot [30] registered Oct 13 05:51:09.940717 kernel: acpiphp: Slot [31] registered Oct 13 05:51:09.940733 kernel: PCI host bridge to bus 0000:00 Oct 13 05:51:09.940901 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Oct 13 05:51:09.941031 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Oct 13 05:51:09.941155 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Oct 13 05:51:09.941273 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Oct 13 05:51:09.941385 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] Oct 13 05:51:09.941501 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Oct 13 05:51:09.941655 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 conventional PCI endpoint Oct 13 05:51:09.943398 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 conventional PCI endpoint Oct 13 05:51:09.943603 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 conventional PCI endpoint Oct 13 05:51:09.943734 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Oct 13 05:51:09.943878 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Oct 13 05:51:09.944009 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Oct 13 05:51:09.944132 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Oct 13 05:51:09.944252 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Oct 13 05:51:09.944374 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Oct 13 05:51:09.944514 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Oct 13 05:51:09.944656 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 conventional PCI endpoint Oct 13 05:51:09.944850 kernel: pci 0000:00:03.0: BAR 0 [mem 0x80000000-0x803fffff pref] Oct 13 05:51:09.945002 kernel: pci 0000:00:03.0: ROM [mem 0xffff0000-0xffffffff pref] Oct 13 05:51:09.945139 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Oct 13 05:51:09.946918 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 PCIe Endpoint Oct 13 05:51:09.947076 kernel: pci 0000:00:04.0: BAR 0 [mem 0x80404000-0x80407fff] Oct 13 05:51:09.947226 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 PCIe Endpoint Oct 13 05:51:09.947362 kernel: pci 0000:00:05.0: BAR 0 [mem 0x80400000-0x80403fff] Oct 13 05:51:09.947387 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Oct 13 05:51:09.947416 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Oct 13 05:51:09.947430 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Oct 13 05:51:09.947543 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Oct 13 05:51:09.947560 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Oct 13 05:51:09.947574 kernel: iommu: Default domain type: Translated Oct 13 05:51:09.947588 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Oct 13 05:51:09.947602 kernel: efivars: Registered efivars operations Oct 13 05:51:09.947617 kernel: PCI: Using ACPI for IRQ routing Oct 13 05:51:09.947635 kernel: PCI: pci_cache_line_size set to 64 bytes Oct 13 05:51:09.947657 kernel: e820: reserve RAM buffer [mem 0x768c0018-0x77ffffff] Oct 13 05:51:09.947670 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] Oct 13 05:51:09.947683 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] Oct 13 05:51:09.947867 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Oct 13 05:51:09.948008 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Oct 13 05:51:09.948146 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Oct 13 05:51:09.948167 kernel: vgaarb: loaded Oct 13 05:51:09.948188 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Oct 13 05:51:09.948205 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Oct 13 05:51:09.948221 kernel: clocksource: Switched to clocksource kvm-clock Oct 13 05:51:09.948238 kernel: VFS: Disk quotas dquot_6.6.0 Oct 13 05:51:09.948254 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Oct 13 05:51:09.948269 kernel: pnp: PnP ACPI init Oct 13 05:51:09.948282 kernel: pnp: PnP ACPI: found 5 devices Oct 13 05:51:09.948297 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Oct 13 05:51:09.948313 kernel: NET: Registered PF_INET protocol family Oct 13 05:51:09.948332 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Oct 13 05:51:09.948349 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Oct 13 05:51:09.948365 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Oct 13 05:51:09.948381 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Oct 13 05:51:09.948397 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Oct 13 05:51:09.948413 kernel: TCP: Hash tables configured (established 16384 bind 16384) Oct 13 05:51:09.948429 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 13 05:51:09.948444 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Oct 13 05:51:09.948460 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Oct 13 05:51:09.948478 kernel: NET: Registered PF_XDP protocol family Oct 13 05:51:09.948602 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Oct 13 05:51:09.948717 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Oct 13 05:51:09.949902 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Oct 13 05:51:09.950036 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Oct 13 05:51:09.950151 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] Oct 13 05:51:09.950286 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Oct 13 05:51:09.950305 kernel: PCI: CLS 0 bytes, default 64 Oct 13 05:51:09.950326 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Oct 13 05:51:09.950342 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093255d7c, max_idle_ns: 440795319144 ns Oct 13 05:51:09.950356 kernel: clocksource: Switched to clocksource tsc Oct 13 05:51:09.950372 kernel: Initialise system trusted keyrings Oct 13 05:51:09.950386 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Oct 13 05:51:09.950402 kernel: Key type asymmetric registered Oct 13 05:51:09.950417 kernel: Asymmetric key parser 'x509' registered Oct 13 05:51:09.950432 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Oct 13 05:51:09.950448 kernel: io scheduler mq-deadline registered Oct 13 05:51:09.950466 kernel: io scheduler kyber registered Oct 13 05:51:09.950480 kernel: io scheduler bfq registered Oct 13 05:51:09.950496 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Oct 13 05:51:09.950511 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Oct 13 05:51:09.950526 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Oct 13 05:51:09.950541 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Oct 13 05:51:09.950556 kernel: i8042: Warning: Keylock active Oct 13 05:51:09.950571 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Oct 13 05:51:09.950587 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Oct 13 05:51:09.950733 kernel: rtc_cmos 00:00: RTC can wake from S4 Oct 13 05:51:09.952922 kernel: rtc_cmos 00:00: registered as rtc0 Oct 13 05:51:09.953065 kernel: rtc_cmos 00:00: setting system clock to 2025-10-13T05:51:09 UTC (1760334669) Oct 13 05:51:09.953192 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Oct 13 05:51:09.953239 kernel: intel_pstate: CPU model not supported Oct 13 05:51:09.953259 kernel: efifb: probing for efifb Oct 13 05:51:09.953276 kernel: efifb: framebuffer at 0x80000000, using 1876k, total 1875k Oct 13 05:51:09.953293 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Oct 13 05:51:09.953313 kernel: efifb: scrolling: redraw Oct 13 05:51:09.953330 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Oct 13 05:51:09.953346 kernel: Console: switching to colour frame buffer device 100x37 Oct 13 05:51:09.953363 kernel: fb0: EFI VGA frame buffer device Oct 13 05:51:09.953381 kernel: pstore: Using crash dump compression: deflate Oct 13 05:51:09.953397 kernel: pstore: Registered efi_pstore as persistent store backend Oct 13 05:51:09.953414 kernel: NET: Registered PF_INET6 protocol family Oct 13 05:51:09.953431 kernel: Segment Routing with IPv6 Oct 13 05:51:09.953448 kernel: In-situ OAM (IOAM) with IPv6 Oct 13 05:51:09.953468 kernel: NET: Registered PF_PACKET protocol family Oct 13 05:51:09.953485 kernel: Key type dns_resolver registered Oct 13 05:51:09.953502 kernel: IPI shorthand broadcast: enabled Oct 13 05:51:09.953518 kernel: sched_clock: Marking stable (2634003119, 150001222)->(2862751093, -78746752) Oct 13 05:51:09.953535 kernel: registered taskstats version 1 Oct 13 05:51:09.953552 kernel: Loading compiled-in X.509 certificates Oct 13 05:51:09.953569 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.51-flatcar: d8dbf4abead15098249886d373d42a3af4f50ccd' Oct 13 05:51:09.953585 kernel: Demotion targets for Node 0: null Oct 13 05:51:09.953602 kernel: Key type .fscrypt registered Oct 13 05:51:09.953621 kernel: Key type fscrypt-provisioning registered Oct 13 05:51:09.953638 kernel: ima: No TPM chip found, activating TPM-bypass! Oct 13 05:51:09.953655 kernel: ima: Allocated hash algorithm: sha1 Oct 13 05:51:09.953672 kernel: ima: No architecture policies found Oct 13 05:51:09.953689 kernel: clk: Disabling unused clocks Oct 13 05:51:09.953705 kernel: Warning: unable to open an initial console. Oct 13 05:51:09.953722 kernel: Freeing unused kernel image (initmem) memory: 54096K Oct 13 05:51:09.953739 kernel: Write protecting the kernel read-only data: 24576k Oct 13 05:51:09.953758 kernel: Freeing unused kernel image (rodata/data gap) memory: 240K Oct 13 05:51:09.955819 kernel: Run /init as init process Oct 13 05:51:09.955845 kernel: with arguments: Oct 13 05:51:09.955863 kernel: /init Oct 13 05:51:09.955880 kernel: with environment: Oct 13 05:51:09.955896 kernel: HOME=/ Oct 13 05:51:09.955918 kernel: TERM=linux Oct 13 05:51:09.955935 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Oct 13 05:51:09.955954 systemd[1]: Successfully made /usr/ read-only. Oct 13 05:51:09.955976 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 13 05:51:09.955995 systemd[1]: Detected virtualization amazon. Oct 13 05:51:09.956013 systemd[1]: Detected architecture x86-64. Oct 13 05:51:09.956033 systemd[1]: Running in initrd. Oct 13 05:51:09.956053 systemd[1]: No hostname configured, using default hostname. Oct 13 05:51:09.956070 systemd[1]: Hostname set to . Oct 13 05:51:09.956088 systemd[1]: Initializing machine ID from VM UUID. Oct 13 05:51:09.956105 systemd[1]: Queued start job for default target initrd.target. Oct 13 05:51:09.956123 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 13 05:51:09.956142 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 13 05:51:09.956161 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Oct 13 05:51:09.956179 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 13 05:51:09.956199 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Oct 13 05:51:09.956219 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Oct 13 05:51:09.956239 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Oct 13 05:51:09.956257 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Oct 13 05:51:09.956275 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 13 05:51:09.956292 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 13 05:51:09.956311 systemd[1]: Reached target paths.target - Path Units. Oct 13 05:51:09.956332 systemd[1]: Reached target slices.target - Slice Units. Oct 13 05:51:09.956350 systemd[1]: Reached target swap.target - Swaps. Oct 13 05:51:09.956367 systemd[1]: Reached target timers.target - Timer Units. Oct 13 05:51:09.956386 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Oct 13 05:51:09.956404 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 13 05:51:09.956422 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Oct 13 05:51:09.956440 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Oct 13 05:51:09.956458 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 13 05:51:09.956476 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 13 05:51:09.956497 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 13 05:51:09.956515 systemd[1]: Reached target sockets.target - Socket Units. Oct 13 05:51:09.956533 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Oct 13 05:51:09.956551 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 13 05:51:09.956569 systemd[1]: Finished network-cleanup.service - Network Cleanup. Oct 13 05:51:09.956588 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Oct 13 05:51:09.956605 systemd[1]: Starting systemd-fsck-usr.service... Oct 13 05:51:09.956623 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 13 05:51:09.956644 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 13 05:51:09.956662 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 05:51:09.956679 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Oct 13 05:51:09.956699 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 13 05:51:09.956758 systemd-journald[207]: Collecting audit messages is disabled. Oct 13 05:51:09.956818 systemd[1]: Finished systemd-fsck-usr.service. Oct 13 05:51:09.956841 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Oct 13 05:51:09.956860 systemd-journald[207]: Journal started Oct 13 05:51:09.956900 systemd-journald[207]: Runtime Journal (/run/log/journal/ec2f7cd44b9a877b0e35039c6ed85210) is 4.8M, max 38.4M, 33.6M free. Oct 13 05:51:09.947688 systemd-modules-load[208]: Inserted module 'overlay' Oct 13 05:51:09.964798 systemd[1]: Started systemd-journald.service - Journal Service. Oct 13 05:51:09.972984 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 13 05:51:09.988815 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 05:51:09.991903 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Oct 13 05:51:09.992009 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Oct 13 05:51:09.995000 kernel: Bridge firewalling registered Oct 13 05:51:09.992767 systemd-modules-load[208]: Inserted module 'br_netfilter' Oct 13 05:51:09.997969 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 13 05:51:10.005800 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Oct 13 05:51:10.006017 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Oct 13 05:51:10.006252 systemd-tmpfiles[222]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Oct 13 05:51:10.013944 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 13 05:51:10.019960 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 13 05:51:10.024374 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 13 05:51:10.036732 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 13 05:51:10.042939 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 13 05:51:10.046162 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 13 05:51:10.048652 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 13 05:51:10.052764 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Oct 13 05:51:10.076286 dracut-cmdline[248]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=a48d469b0deb49c328e6faf6cf366b11952d47f2d24963c866a0ea8221fb0039 Oct 13 05:51:10.104445 systemd-resolved[246]: Positive Trust Anchors: Oct 13 05:51:10.104462 systemd-resolved[246]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 13 05:51:10.104526 systemd-resolved[246]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 13 05:51:10.113183 systemd-resolved[246]: Defaulting to hostname 'linux'. Oct 13 05:51:10.116246 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 13 05:51:10.116952 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 13 05:51:10.176822 kernel: SCSI subsystem initialized Oct 13 05:51:10.187818 kernel: Loading iSCSI transport class v2.0-870. Oct 13 05:51:10.199813 kernel: iscsi: registered transport (tcp) Oct 13 05:51:10.221024 kernel: iscsi: registered transport (qla4xxx) Oct 13 05:51:10.221098 kernel: QLogic iSCSI HBA Driver Oct 13 05:51:10.240763 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 13 05:51:10.264901 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 13 05:51:10.267666 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 13 05:51:10.315384 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Oct 13 05:51:10.317639 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Oct 13 05:51:10.371833 kernel: raid6: avx512x4 gen() 17307 MB/s Oct 13 05:51:10.389811 kernel: raid6: avx512x2 gen() 17318 MB/s Oct 13 05:51:10.407817 kernel: raid6: avx512x1 gen() 15369 MB/s Oct 13 05:51:10.425832 kernel: raid6: avx2x4 gen() 17063 MB/s Oct 13 05:51:10.443815 kernel: raid6: avx2x2 gen() 16911 MB/s Oct 13 05:51:10.462083 kernel: raid6: avx2x1 gen() 13688 MB/s Oct 13 05:51:10.462154 kernel: raid6: using algorithm avx512x2 gen() 17318 MB/s Oct 13 05:51:10.481182 kernel: raid6: .... xor() 24124 MB/s, rmw enabled Oct 13 05:51:10.481264 kernel: raid6: using avx512x2 recovery algorithm Oct 13 05:51:10.501862 kernel: xor: automatically using best checksumming function avx Oct 13 05:51:10.670815 kernel: Btrfs loaded, zoned=no, fsverity=no Oct 13 05:51:10.677570 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Oct 13 05:51:10.679966 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 13 05:51:10.706503 systemd-udevd[457]: Using default interface naming scheme 'v255'. Oct 13 05:51:10.713374 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 13 05:51:10.718567 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Oct 13 05:51:10.746125 dracut-pre-trigger[463]: rd.md=0: removing MD RAID activation Oct 13 05:51:10.773507 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Oct 13 05:51:10.775682 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 13 05:51:10.839932 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 13 05:51:10.844010 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Oct 13 05:51:10.934836 kernel: cryptd: max_cpu_qlen set to 1000 Oct 13 05:51:10.962925 kernel: ena 0000:00:05.0: ENA device version: 0.10 Oct 13 05:51:10.963203 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Oct 13 05:51:10.976813 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Oct 13 05:51:10.979804 kernel: nvme nvme0: pci function 0000:00:04.0 Oct 13 05:51:10.980181 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:d8:10:60:82:cb Oct 13 05:51:10.985075 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Oct 13 05:51:10.985138 kernel: AES CTR mode by8 optimization enabled Oct 13 05:51:10.986744 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 13 05:51:10.990885 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input2 Oct 13 05:51:10.989053 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 05:51:10.990286 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 05:51:10.998181 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 05:51:11.008122 kernel: nvme nvme0: 2/0/0 default/read/poll queues Oct 13 05:51:11.006399 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Oct 13 05:51:11.026510 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Oct 13 05:51:11.026547 kernel: GPT:9289727 != 33554431 Oct 13 05:51:11.026566 kernel: GPT:Alternate GPT header not at the end of the disk. Oct 13 05:51:11.026583 kernel: GPT:9289727 != 33554431 Oct 13 05:51:11.026601 kernel: GPT: Use GNU Parted to correct GPT errors. Oct 13 05:51:11.026618 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Oct 13 05:51:11.035624 (udev-worker)[512]: Network interface NamePolicy= disabled on kernel command line. Oct 13 05:51:11.056190 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 05:51:11.068897 kernel: nvme nvme0: using unchecked data buffer Oct 13 05:51:11.205227 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Oct 13 05:51:11.217610 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Oct 13 05:51:11.227982 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Oct 13 05:51:11.228663 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Oct 13 05:51:11.240528 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Oct 13 05:51:11.251444 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Oct 13 05:51:11.252112 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Oct 13 05:51:11.253316 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 13 05:51:11.254460 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 13 05:51:11.256340 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Oct 13 05:51:11.259084 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Oct 13 05:51:11.274070 disk-uuid[693]: Primary Header is updated. Oct 13 05:51:11.274070 disk-uuid[693]: Secondary Entries is updated. Oct 13 05:51:11.274070 disk-uuid[693]: Secondary Header is updated. Oct 13 05:51:11.280847 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Oct 13 05:51:11.285464 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Oct 13 05:51:11.294171 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Oct 13 05:51:12.298877 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Oct 13 05:51:12.299395 disk-uuid[697]: The operation has completed successfully. Oct 13 05:51:12.436842 systemd[1]: disk-uuid.service: Deactivated successfully. Oct 13 05:51:12.436984 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Oct 13 05:51:12.472855 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Oct 13 05:51:12.491769 sh[959]: Success Oct 13 05:51:12.517956 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Oct 13 05:51:12.518035 kernel: device-mapper: uevent: version 1.0.3 Oct 13 05:51:12.518795 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Oct 13 05:51:12.530806 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" Oct 13 05:51:12.626266 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Oct 13 05:51:12.630071 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Oct 13 05:51:12.645137 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Oct 13 05:51:12.667152 kernel: BTRFS: device fsid c8746500-26f5-4ec1-9da8-aef51ec7db92 devid 1 transid 41 /dev/mapper/usr (254:0) scanned by mount (982) Oct 13 05:51:12.667317 kernel: BTRFS info (device dm-0): first mount of filesystem c8746500-26f5-4ec1-9da8-aef51ec7db92 Oct 13 05:51:12.667812 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Oct 13 05:51:12.786939 kernel: BTRFS info (device dm-0): enabling ssd optimizations Oct 13 05:51:12.787006 kernel: BTRFS info (device dm-0): disabling log replay at mount time Oct 13 05:51:12.789446 kernel: BTRFS info (device dm-0): enabling free space tree Oct 13 05:51:12.801968 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Oct 13 05:51:12.802948 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Oct 13 05:51:12.804199 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Oct 13 05:51:12.805143 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Oct 13 05:51:12.806548 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Oct 13 05:51:12.842810 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1017) Oct 13 05:51:12.849856 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 1cd10441-4b32-40b7-b370-b928e4bc90dd Oct 13 05:51:12.849931 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Oct 13 05:51:12.867267 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Oct 13 05:51:12.867350 kernel: BTRFS info (device nvme0n1p6): enabling free space tree Oct 13 05:51:12.874835 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 1cd10441-4b32-40b7-b370-b928e4bc90dd Oct 13 05:51:12.875943 systemd[1]: Finished ignition-setup.service - Ignition (setup). Oct 13 05:51:12.879935 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Oct 13 05:51:12.922089 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 13 05:51:12.925900 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 13 05:51:12.969993 systemd-networkd[1151]: lo: Link UP Oct 13 05:51:12.970005 systemd-networkd[1151]: lo: Gained carrier Oct 13 05:51:12.971909 systemd-networkd[1151]: Enumeration completed Oct 13 05:51:12.972184 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 13 05:51:12.972341 systemd-networkd[1151]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 05:51:12.972347 systemd-networkd[1151]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 13 05:51:12.973169 systemd[1]: Reached target network.target - Network. Oct 13 05:51:12.980499 systemd-networkd[1151]: eth0: Link UP Oct 13 05:51:12.980510 systemd-networkd[1151]: eth0: Gained carrier Oct 13 05:51:12.980531 systemd-networkd[1151]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 05:51:12.997884 systemd-networkd[1151]: eth0: DHCPv4 address 172.31.26.230/20, gateway 172.31.16.1 acquired from 172.31.16.1 Oct 13 05:51:13.298201 ignition[1109]: Ignition 2.22.0 Oct 13 05:51:13.298217 ignition[1109]: Stage: fetch-offline Oct 13 05:51:13.298388 ignition[1109]: no configs at "/usr/lib/ignition/base.d" Oct 13 05:51:13.298397 ignition[1109]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Oct 13 05:51:13.298678 ignition[1109]: Ignition finished successfully Oct 13 05:51:13.300753 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Oct 13 05:51:13.302406 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Oct 13 05:51:13.336938 ignition[1160]: Ignition 2.22.0 Oct 13 05:51:13.336955 ignition[1160]: Stage: fetch Oct 13 05:51:13.337334 ignition[1160]: no configs at "/usr/lib/ignition/base.d" Oct 13 05:51:13.337347 ignition[1160]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Oct 13 05:51:13.337470 ignition[1160]: PUT http://169.254.169.254/latest/api/token: attempt #1 Oct 13 05:51:13.360613 ignition[1160]: PUT result: OK Oct 13 05:51:13.363167 ignition[1160]: parsed url from cmdline: "" Oct 13 05:51:13.363176 ignition[1160]: no config URL provided Oct 13 05:51:13.363184 ignition[1160]: reading system config file "/usr/lib/ignition/user.ign" Oct 13 05:51:13.363195 ignition[1160]: no config at "/usr/lib/ignition/user.ign" Oct 13 05:51:13.363228 ignition[1160]: PUT http://169.254.169.254/latest/api/token: attempt #1 Oct 13 05:51:13.364480 ignition[1160]: PUT result: OK Oct 13 05:51:13.364528 ignition[1160]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Oct 13 05:51:13.365216 ignition[1160]: GET result: OK Oct 13 05:51:13.365313 ignition[1160]: parsing config with SHA512: a30ddcea11af3d33273a8ebea9776162b8ea2fa1c59988c1b1ad85906d95f664dcc9f077f0341b00a214e72c98afddfc6aafa879771eafdc7fe9ce0b4f726f51 Oct 13 05:51:13.373456 unknown[1160]: fetched base config from "system" Oct 13 05:51:13.373469 unknown[1160]: fetched base config from "system" Oct 13 05:51:13.373812 ignition[1160]: fetch: fetch complete Oct 13 05:51:13.373475 unknown[1160]: fetched user config from "aws" Oct 13 05:51:13.373817 ignition[1160]: fetch: fetch passed Oct 13 05:51:13.376286 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Oct 13 05:51:13.373856 ignition[1160]: Ignition finished successfully Oct 13 05:51:13.378598 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Oct 13 05:51:13.415020 ignition[1166]: Ignition 2.22.0 Oct 13 05:51:13.415035 ignition[1166]: Stage: kargs Oct 13 05:51:13.415588 ignition[1166]: no configs at "/usr/lib/ignition/base.d" Oct 13 05:51:13.415612 ignition[1166]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Oct 13 05:51:13.415730 ignition[1166]: PUT http://169.254.169.254/latest/api/token: attempt #1 Oct 13 05:51:13.416610 ignition[1166]: PUT result: OK Oct 13 05:51:13.419298 ignition[1166]: kargs: kargs passed Oct 13 05:51:13.419374 ignition[1166]: Ignition finished successfully Oct 13 05:51:13.421720 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Oct 13 05:51:13.423343 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Oct 13 05:51:13.457838 ignition[1173]: Ignition 2.22.0 Oct 13 05:51:13.457862 ignition[1173]: Stage: disks Oct 13 05:51:13.458235 ignition[1173]: no configs at "/usr/lib/ignition/base.d" Oct 13 05:51:13.458247 ignition[1173]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Oct 13 05:51:13.458362 ignition[1173]: PUT http://169.254.169.254/latest/api/token: attempt #1 Oct 13 05:51:13.459212 ignition[1173]: PUT result: OK Oct 13 05:51:13.461570 ignition[1173]: disks: disks passed Oct 13 05:51:13.461650 ignition[1173]: Ignition finished successfully Oct 13 05:51:13.463715 systemd[1]: Finished ignition-disks.service - Ignition (disks). Oct 13 05:51:13.464583 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Oct 13 05:51:13.464975 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Oct 13 05:51:13.465493 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 13 05:51:13.466024 systemd[1]: Reached target sysinit.target - System Initialization. Oct 13 05:51:13.466553 systemd[1]: Reached target basic.target - Basic System. Oct 13 05:51:13.468371 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Oct 13 05:51:13.500142 systemd-fsck[1181]: ROOT: clean, 15/553520 files, 52789/553472 blocks Oct 13 05:51:13.503953 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Oct 13 05:51:13.506264 systemd[1]: Mounting sysroot.mount - /sysroot... Oct 13 05:51:13.661837 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 8b520359-9763-45f3-b7f7-db1e9fbc640d r/w with ordered data mode. Quota mode: none. Oct 13 05:51:13.662656 systemd[1]: Mounted sysroot.mount - /sysroot. Oct 13 05:51:13.664022 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Oct 13 05:51:13.666063 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 13 05:51:13.668764 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Oct 13 05:51:13.672379 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Oct 13 05:51:13.672455 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Oct 13 05:51:13.672491 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Oct 13 05:51:13.680865 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Oct 13 05:51:13.683042 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Oct 13 05:51:13.695810 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1200) Oct 13 05:51:13.700882 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 1cd10441-4b32-40b7-b370-b928e4bc90dd Oct 13 05:51:13.700939 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Oct 13 05:51:13.708737 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Oct 13 05:51:13.708821 kernel: BTRFS info (device nvme0n1p6): enabling free space tree Oct 13 05:51:13.710266 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 13 05:51:13.944230 initrd-setup-root[1224]: cut: /sysroot/etc/passwd: No such file or directory Oct 13 05:51:13.958923 initrd-setup-root[1231]: cut: /sysroot/etc/group: No such file or directory Oct 13 05:51:13.964774 initrd-setup-root[1238]: cut: /sysroot/etc/shadow: No such file or directory Oct 13 05:51:13.969433 initrd-setup-root[1245]: cut: /sysroot/etc/gshadow: No such file or directory Oct 13 05:51:14.147629 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Oct 13 05:51:14.149846 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Oct 13 05:51:14.152490 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Oct 13 05:51:14.168936 systemd[1]: sysroot-oem.mount: Deactivated successfully. Oct 13 05:51:14.171598 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 1cd10441-4b32-40b7-b370-b928e4bc90dd Oct 13 05:51:14.198292 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Oct 13 05:51:14.208008 ignition[1313]: INFO : Ignition 2.22.0 Oct 13 05:51:14.208008 ignition[1313]: INFO : Stage: mount Oct 13 05:51:14.209467 ignition[1313]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 13 05:51:14.209467 ignition[1313]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Oct 13 05:51:14.209467 ignition[1313]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Oct 13 05:51:14.209467 ignition[1313]: INFO : PUT result: OK Oct 13 05:51:14.212073 ignition[1313]: INFO : mount: mount passed Oct 13 05:51:14.213519 ignition[1313]: INFO : Ignition finished successfully Oct 13 05:51:14.214437 systemd[1]: Finished ignition-mount.service - Ignition (mount). Oct 13 05:51:14.216143 systemd[1]: Starting ignition-files.service - Ignition (files)... Oct 13 05:51:14.626044 systemd-networkd[1151]: eth0: Gained IPv6LL Oct 13 05:51:14.664421 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Oct 13 05:51:14.703804 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 (259:5) scanned by mount (1324) Oct 13 05:51:14.707878 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 1cd10441-4b32-40b7-b370-b928e4bc90dd Oct 13 05:51:14.707949 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Oct 13 05:51:14.715109 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Oct 13 05:51:14.715199 kernel: BTRFS info (device nvme0n1p6): enabling free space tree Oct 13 05:51:14.717355 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Oct 13 05:51:14.752343 ignition[1340]: INFO : Ignition 2.22.0 Oct 13 05:51:14.752343 ignition[1340]: INFO : Stage: files Oct 13 05:51:14.753690 ignition[1340]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 13 05:51:14.753690 ignition[1340]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Oct 13 05:51:14.753690 ignition[1340]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Oct 13 05:51:14.755233 ignition[1340]: INFO : PUT result: OK Oct 13 05:51:14.757005 ignition[1340]: DEBUG : files: compiled without relabeling support, skipping Oct 13 05:51:14.758362 ignition[1340]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Oct 13 05:51:14.758362 ignition[1340]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Oct 13 05:51:14.764045 ignition[1340]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Oct 13 05:51:14.764789 ignition[1340]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Oct 13 05:51:14.764789 ignition[1340]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Oct 13 05:51:14.764513 unknown[1340]: wrote ssh authorized keys file for user: core Oct 13 05:51:14.767158 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 13 05:51:14.767158 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Oct 13 05:51:14.817542 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Oct 13 05:51:15.005496 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Oct 13 05:51:15.005496 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Oct 13 05:51:15.007319 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Oct 13 05:51:15.007319 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Oct 13 05:51:15.007319 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Oct 13 05:51:15.007319 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 13 05:51:15.007319 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Oct 13 05:51:15.007319 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 13 05:51:15.007319 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Oct 13 05:51:15.012534 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Oct 13 05:51:15.012534 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Oct 13 05:51:15.012534 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 13 05:51:15.015119 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 13 05:51:15.015119 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 13 05:51:15.015119 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.32.4-x86-64.raw: attempt #1 Oct 13 05:51:15.329002 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Oct 13 05:51:15.922228 ignition[1340]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.4-x86-64.raw" Oct 13 05:51:15.922228 ignition[1340]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Oct 13 05:51:15.924516 ignition[1340]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 13 05:51:15.928224 ignition[1340]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Oct 13 05:51:15.928224 ignition[1340]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Oct 13 05:51:15.928224 ignition[1340]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Oct 13 05:51:15.932400 ignition[1340]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Oct 13 05:51:15.932400 ignition[1340]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Oct 13 05:51:15.932400 ignition[1340]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Oct 13 05:51:15.932400 ignition[1340]: INFO : files: files passed Oct 13 05:51:15.932400 ignition[1340]: INFO : Ignition finished successfully Oct 13 05:51:15.930072 systemd[1]: Finished ignition-files.service - Ignition (files). Oct 13 05:51:15.935922 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Oct 13 05:51:15.937462 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Oct 13 05:51:15.946691 systemd[1]: ignition-quench.service: Deactivated successfully. Oct 13 05:51:15.946965 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Oct 13 05:51:15.956146 initrd-setup-root-after-ignition[1371]: grep: Oct 13 05:51:15.956146 initrd-setup-root-after-ignition[1375]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 13 05:51:15.958809 initrd-setup-root-after-ignition[1371]: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Oct 13 05:51:15.958809 initrd-setup-root-after-ignition[1371]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Oct 13 05:51:15.958405 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 13 05:51:15.959557 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Oct 13 05:51:15.962457 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Oct 13 05:51:16.016500 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Oct 13 05:51:16.016640 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Oct 13 05:51:16.017942 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Oct 13 05:51:16.019008 systemd[1]: Reached target initrd.target - Initrd Default Target. Oct 13 05:51:16.020012 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Oct 13 05:51:16.021181 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Oct 13 05:51:16.044693 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 13 05:51:16.047007 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Oct 13 05:51:16.065987 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Oct 13 05:51:16.066759 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 13 05:51:16.068003 systemd[1]: Stopped target timers.target - Timer Units. Oct 13 05:51:16.068803 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Oct 13 05:51:16.068988 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Oct 13 05:51:16.070236 systemd[1]: Stopped target initrd.target - Initrd Default Target. Oct 13 05:51:16.071169 systemd[1]: Stopped target basic.target - Basic System. Oct 13 05:51:16.072122 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Oct 13 05:51:16.072923 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Oct 13 05:51:16.073713 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Oct 13 05:51:16.074504 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Oct 13 05:51:16.075343 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Oct 13 05:51:16.076271 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Oct 13 05:51:16.077098 systemd[1]: Stopped target sysinit.target - System Initialization. Oct 13 05:51:16.078172 systemd[1]: Stopped target local-fs.target - Local File Systems. Oct 13 05:51:16.078981 systemd[1]: Stopped target swap.target - Swaps. Oct 13 05:51:16.079872 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Oct 13 05:51:16.080099 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Oct 13 05:51:16.081136 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Oct 13 05:51:16.081907 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 13 05:51:16.082625 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Oct 13 05:51:16.082758 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 13 05:51:16.083586 systemd[1]: dracut-initqueue.service: Deactivated successfully. Oct 13 05:51:16.083851 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Oct 13 05:51:16.084871 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Oct 13 05:51:16.085113 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Oct 13 05:51:16.085753 systemd[1]: ignition-files.service: Deactivated successfully. Oct 13 05:51:16.085970 systemd[1]: Stopped ignition-files.service - Ignition (files). Oct 13 05:51:16.089024 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Oct 13 05:51:16.089572 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Oct 13 05:51:16.089815 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Oct 13 05:51:16.094589 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Oct 13 05:51:16.095968 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Oct 13 05:51:16.096817 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Oct 13 05:51:16.098216 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Oct 13 05:51:16.098988 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Oct 13 05:51:16.105619 systemd[1]: initrd-cleanup.service: Deactivated successfully. Oct 13 05:51:16.106466 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Oct 13 05:51:16.129554 ignition[1395]: INFO : Ignition 2.22.0 Oct 13 05:51:16.129554 ignition[1395]: INFO : Stage: umount Oct 13 05:51:16.132082 ignition[1395]: INFO : no configs at "/usr/lib/ignition/base.d" Oct 13 05:51:16.132082 ignition[1395]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Oct 13 05:51:16.132082 ignition[1395]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Oct 13 05:51:16.132082 ignition[1395]: INFO : PUT result: OK Oct 13 05:51:16.134287 systemd[1]: sysroot-boot.mount: Deactivated successfully. Oct 13 05:51:16.136435 ignition[1395]: INFO : umount: umount passed Oct 13 05:51:16.136435 ignition[1395]: INFO : Ignition finished successfully Oct 13 05:51:16.140114 systemd[1]: ignition-mount.service: Deactivated successfully. Oct 13 05:51:16.140267 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Oct 13 05:51:16.141323 systemd[1]: sysroot-boot.service: Deactivated successfully. Oct 13 05:51:16.141453 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Oct 13 05:51:16.142942 systemd[1]: ignition-disks.service: Deactivated successfully. Oct 13 05:51:16.143036 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Oct 13 05:51:16.143942 systemd[1]: ignition-kargs.service: Deactivated successfully. Oct 13 05:51:16.144005 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Oct 13 05:51:16.144569 systemd[1]: ignition-fetch.service: Deactivated successfully. Oct 13 05:51:16.144627 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Oct 13 05:51:16.145221 systemd[1]: Stopped target network.target - Network. Oct 13 05:51:16.145884 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Oct 13 05:51:16.145953 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Oct 13 05:51:16.146466 systemd[1]: Stopped target paths.target - Path Units. Oct 13 05:51:16.147042 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Oct 13 05:51:16.150909 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 13 05:51:16.151368 systemd[1]: Stopped target slices.target - Slice Units. Oct 13 05:51:16.152630 systemd[1]: Stopped target sockets.target - Socket Units. Oct 13 05:51:16.153364 systemd[1]: iscsid.socket: Deactivated successfully. Oct 13 05:51:16.153435 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Oct 13 05:51:16.154120 systemd[1]: iscsiuio.socket: Deactivated successfully. Oct 13 05:51:16.154175 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Oct 13 05:51:16.154768 systemd[1]: ignition-setup.service: Deactivated successfully. Oct 13 05:51:16.154872 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Oct 13 05:51:16.155690 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Oct 13 05:51:16.155757 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Oct 13 05:51:16.156310 systemd[1]: initrd-setup-root.service: Deactivated successfully. Oct 13 05:51:16.156379 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Oct 13 05:51:16.157146 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Oct 13 05:51:16.157744 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Oct 13 05:51:16.162280 systemd[1]: systemd-resolved.service: Deactivated successfully. Oct 13 05:51:16.162417 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Oct 13 05:51:16.166226 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Oct 13 05:51:16.166548 systemd[1]: systemd-networkd.service: Deactivated successfully. Oct 13 05:51:16.166690 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Oct 13 05:51:16.169192 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Oct 13 05:51:16.170376 systemd[1]: Stopped target network-pre.target - Preparation for Network. Oct 13 05:51:16.170926 systemd[1]: systemd-networkd.socket: Deactivated successfully. Oct 13 05:51:16.170986 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Oct 13 05:51:16.172890 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Oct 13 05:51:16.173238 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Oct 13 05:51:16.173307 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Oct 13 05:51:16.175927 systemd[1]: systemd-sysctl.service: Deactivated successfully. Oct 13 05:51:16.175993 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Oct 13 05:51:16.176768 systemd[1]: systemd-modules-load.service: Deactivated successfully. Oct 13 05:51:16.176900 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Oct 13 05:51:16.177456 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Oct 13 05:51:16.177519 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 13 05:51:16.179593 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 13 05:51:16.185764 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Oct 13 05:51:16.186722 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Oct 13 05:51:16.194323 systemd[1]: systemd-udevd.service: Deactivated successfully. Oct 13 05:51:16.195399 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 13 05:51:16.197403 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Oct 13 05:51:16.197503 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Oct 13 05:51:16.198393 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Oct 13 05:51:16.198443 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Oct 13 05:51:16.199106 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Oct 13 05:51:16.199172 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Oct 13 05:51:16.200566 systemd[1]: dracut-cmdline.service: Deactivated successfully. Oct 13 05:51:16.200631 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Oct 13 05:51:16.202643 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Oct 13 05:51:16.202722 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Oct 13 05:51:16.205580 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Oct 13 05:51:16.206908 systemd[1]: systemd-network-generator.service: Deactivated successfully. Oct 13 05:51:16.206986 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Oct 13 05:51:16.210291 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Oct 13 05:51:16.210942 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 13 05:51:16.212259 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 13 05:51:16.212311 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 05:51:16.214685 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Oct 13 05:51:16.214792 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Oct 13 05:51:16.214855 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Oct 13 05:51:16.215392 systemd[1]: network-cleanup.service: Deactivated successfully. Oct 13 05:51:16.218915 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Oct 13 05:51:16.225727 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Oct 13 05:51:16.225880 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Oct 13 05:51:16.227512 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Oct 13 05:51:16.229531 systemd[1]: Starting initrd-switch-root.service - Switch Root... Oct 13 05:51:16.248641 systemd[1]: Switching root. Oct 13 05:51:16.300689 systemd-journald[207]: Journal stopped Oct 13 05:51:17.668565 systemd-journald[207]: Received SIGTERM from PID 1 (systemd). Oct 13 05:51:17.668623 kernel: SELinux: policy capability network_peer_controls=1 Oct 13 05:51:17.668638 kernel: SELinux: policy capability open_perms=1 Oct 13 05:51:17.668649 kernel: SELinux: policy capability extended_socket_class=1 Oct 13 05:51:17.668660 kernel: SELinux: policy capability always_check_network=0 Oct 13 05:51:17.668674 kernel: SELinux: policy capability cgroup_seclabel=1 Oct 13 05:51:17.668686 kernel: SELinux: policy capability nnp_nosuid_transition=1 Oct 13 05:51:17.668697 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Oct 13 05:51:17.668708 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Oct 13 05:51:17.668719 kernel: SELinux: policy capability userspace_initial_context=0 Oct 13 05:51:17.668732 kernel: audit: type=1403 audit(1760334676.660:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Oct 13 05:51:17.668752 systemd[1]: Successfully loaded SELinux policy in 75.662ms. Oct 13 05:51:17.668775 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 6.402ms. Oct 13 05:51:17.670715 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Oct 13 05:51:17.670758 systemd[1]: Detected virtualization amazon. Oct 13 05:51:17.670800 systemd[1]: Detected architecture x86-64. Oct 13 05:51:17.670822 systemd[1]: Detected first boot. Oct 13 05:51:17.670847 systemd[1]: Initializing machine ID from VM UUID. Oct 13 05:51:17.670878 zram_generator::config[1439]: No configuration found. Oct 13 05:51:17.670905 kernel: Guest personality initialized and is inactive Oct 13 05:51:17.670928 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Oct 13 05:51:17.670954 kernel: Initialized host personality Oct 13 05:51:17.670975 kernel: NET: Registered PF_VSOCK protocol family Oct 13 05:51:17.671000 systemd[1]: Populated /etc with preset unit settings. Oct 13 05:51:17.671024 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Oct 13 05:51:17.671051 systemd[1]: initrd-switch-root.service: Deactivated successfully. Oct 13 05:51:17.671075 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Oct 13 05:51:17.671104 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Oct 13 05:51:17.671138 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Oct 13 05:51:17.671163 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Oct 13 05:51:17.671187 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Oct 13 05:51:17.671211 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Oct 13 05:51:17.671237 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Oct 13 05:51:17.671263 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Oct 13 05:51:17.671287 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Oct 13 05:51:17.671316 systemd[1]: Created slice user.slice - User and Session Slice. Oct 13 05:51:17.671346 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Oct 13 05:51:17.671373 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Oct 13 05:51:17.671397 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Oct 13 05:51:17.671432 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Oct 13 05:51:17.671459 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Oct 13 05:51:17.671483 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Oct 13 05:51:17.671508 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Oct 13 05:51:17.671542 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Oct 13 05:51:17.671568 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Oct 13 05:51:17.671592 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Oct 13 05:51:17.671616 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Oct 13 05:51:17.671640 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Oct 13 05:51:17.671665 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Oct 13 05:51:17.671690 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Oct 13 05:51:17.671715 systemd[1]: Reached target remote-fs.target - Remote File Systems. Oct 13 05:51:17.671740 systemd[1]: Reached target slices.target - Slice Units. Oct 13 05:51:17.671765 systemd[1]: Reached target swap.target - Swaps. Oct 13 05:51:17.684681 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Oct 13 05:51:17.684727 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Oct 13 05:51:17.684752 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Oct 13 05:51:17.684777 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Oct 13 05:51:17.684863 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Oct 13 05:51:17.684890 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Oct 13 05:51:17.684915 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Oct 13 05:51:17.684941 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Oct 13 05:51:17.684976 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Oct 13 05:51:17.685000 systemd[1]: Mounting media.mount - External Media Directory... Oct 13 05:51:17.685026 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 13 05:51:17.685049 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Oct 13 05:51:17.685075 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Oct 13 05:51:17.685100 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Oct 13 05:51:17.685126 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Oct 13 05:51:17.685152 systemd[1]: Reached target machines.target - Containers. Oct 13 05:51:17.685174 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Oct 13 05:51:17.685204 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 13 05:51:17.685222 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Oct 13 05:51:17.685244 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Oct 13 05:51:17.685272 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 13 05:51:17.685297 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 13 05:51:17.685321 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 13 05:51:17.685345 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Oct 13 05:51:17.685369 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 13 05:51:17.685398 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Oct 13 05:51:17.685423 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Oct 13 05:51:17.685447 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Oct 13 05:51:17.685472 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Oct 13 05:51:17.685496 systemd[1]: Stopped systemd-fsck-usr.service. Oct 13 05:51:17.685523 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 13 05:51:17.685546 systemd[1]: Starting systemd-journald.service - Journal Service... Oct 13 05:51:17.685567 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Oct 13 05:51:17.685588 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Oct 13 05:51:17.685618 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Oct 13 05:51:17.685642 kernel: loop: module loaded Oct 13 05:51:17.685670 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Oct 13 05:51:17.685695 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Oct 13 05:51:17.685722 systemd[1]: verity-setup.service: Deactivated successfully. Oct 13 05:51:17.685748 systemd[1]: Stopped verity-setup.service. Oct 13 05:51:17.685773 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 13 05:51:17.687965 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Oct 13 05:51:17.687995 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Oct 13 05:51:17.688014 systemd[1]: Mounted media.mount - External Media Directory. Oct 13 05:51:17.688039 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Oct 13 05:51:17.688060 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Oct 13 05:51:17.688080 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Oct 13 05:51:17.688102 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Oct 13 05:51:17.688121 systemd[1]: modprobe@configfs.service: Deactivated successfully. Oct 13 05:51:17.688141 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Oct 13 05:51:17.688161 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 13 05:51:17.688181 kernel: ACPI: bus type drm_connector registered Oct 13 05:51:17.688201 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 13 05:51:17.688225 kernel: fuse: init (API version 7.41) Oct 13 05:51:17.688243 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 13 05:51:17.688263 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 13 05:51:17.688284 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 13 05:51:17.688307 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 13 05:51:17.688329 systemd[1]: modprobe@fuse.service: Deactivated successfully. Oct 13 05:51:17.688351 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Oct 13 05:51:17.688372 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 13 05:51:17.688398 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 13 05:51:17.688421 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Oct 13 05:51:17.688442 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Oct 13 05:51:17.688462 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Oct 13 05:51:17.688528 systemd-journald[1525]: Collecting audit messages is disabled. Oct 13 05:51:17.688572 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Oct 13 05:51:17.688594 systemd-journald[1525]: Journal started Oct 13 05:51:17.688634 systemd-journald[1525]: Runtime Journal (/run/log/journal/ec2f7cd44b9a877b0e35039c6ed85210) is 4.8M, max 38.4M, 33.6M free. Oct 13 05:51:17.356615 systemd[1]: Queued start job for default target multi-user.target. Oct 13 05:51:17.691295 systemd[1]: Started systemd-journald.service - Journal Service. Oct 13 05:51:17.365432 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Oct 13 05:51:17.365897 systemd[1]: systemd-journald.service: Deactivated successfully. Oct 13 05:51:17.696050 systemd[1]: Reached target network-pre.target - Preparation for Network. Oct 13 05:51:17.701909 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Oct 13 05:51:17.706903 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Oct 13 05:51:17.708892 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Oct 13 05:51:17.708946 systemd[1]: Reached target local-fs.target - Local File Systems. Oct 13 05:51:17.715092 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Oct 13 05:51:17.727060 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Oct 13 05:51:17.729131 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 13 05:51:17.732041 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Oct 13 05:51:17.736485 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Oct 13 05:51:17.738311 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 13 05:51:17.739807 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Oct 13 05:51:17.740722 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 13 05:51:17.744023 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Oct 13 05:51:17.751029 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Oct 13 05:51:17.754759 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Oct 13 05:51:17.758482 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Oct 13 05:51:17.759373 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Oct 13 05:51:17.799102 systemd[1]: Starting systemd-sysusers.service - Create System Users... Oct 13 05:51:17.802647 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Oct 13 05:51:17.810000 systemd-journald[1525]: Time spent on flushing to /var/log/journal/ec2f7cd44b9a877b0e35039c6ed85210 is 88.781ms for 1019 entries. Oct 13 05:51:17.810000 systemd-journald[1525]: System Journal (/var/log/journal/ec2f7cd44b9a877b0e35039c6ed85210) is 8M, max 195.6M, 187.6M free. Oct 13 05:51:17.915706 systemd-journald[1525]: Received client request to flush runtime journal. Oct 13 05:51:17.915772 kernel: loop0: detected capacity change from 0 to 110984 Oct 13 05:51:17.850185 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Oct 13 05:51:17.854417 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Oct 13 05:51:17.864934 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Oct 13 05:51:17.867270 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Oct 13 05:51:17.923356 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Oct 13 05:51:17.933846 systemd[1]: Finished systemd-sysusers.service - Create System Users. Oct 13 05:51:17.938092 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Oct 13 05:51:17.952976 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Oct 13 05:51:17.963803 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Oct 13 05:51:17.989912 kernel: loop1: detected capacity change from 0 to 128016 Oct 13 05:51:18.021756 systemd-tmpfiles[1591]: ACLs are not supported, ignoring. Oct 13 05:51:18.021804 systemd-tmpfiles[1591]: ACLs are not supported, ignoring. Oct 13 05:51:18.032244 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Oct 13 05:51:18.041701 kernel: loop2: detected capacity change from 0 to 224512 Oct 13 05:51:18.193862 kernel: loop3: detected capacity change from 0 to 72368 Oct 13 05:51:18.325807 kernel: loop4: detected capacity change from 0 to 110984 Oct 13 05:51:18.367824 kernel: loop5: detected capacity change from 0 to 128016 Oct 13 05:51:18.368153 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Oct 13 05:51:18.389840 kernel: loop6: detected capacity change from 0 to 224512 Oct 13 05:51:18.434315 kernel: loop7: detected capacity change from 0 to 72368 Oct 13 05:51:18.450897 (sd-merge)[1598]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Oct 13 05:51:18.454593 (sd-merge)[1598]: Merged extensions into '/usr'. Oct 13 05:51:18.471528 systemd[1]: Reload requested from client PID 1555 ('systemd-sysext') (unit systemd-sysext.service)... Oct 13 05:51:18.472934 systemd[1]: Reloading... Oct 13 05:51:18.618810 zram_generator::config[1621]: No configuration found. Oct 13 05:51:18.665134 ldconfig[1550]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Oct 13 05:51:18.921382 systemd[1]: Reloading finished in 447 ms. Oct 13 05:51:18.945371 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Oct 13 05:51:18.946176 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Oct 13 05:51:18.946898 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Oct 13 05:51:18.958136 systemd[1]: Starting ensure-sysext.service... Oct 13 05:51:18.962964 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Oct 13 05:51:18.970948 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Oct 13 05:51:18.983709 systemd[1]: Reload requested from client PID 1677 ('systemctl') (unit ensure-sysext.service)... Oct 13 05:51:18.983723 systemd[1]: Reloading... Oct 13 05:51:18.986364 systemd-tmpfiles[1678]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Oct 13 05:51:18.986627 systemd-tmpfiles[1678]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Oct 13 05:51:18.986973 systemd-tmpfiles[1678]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Oct 13 05:51:18.987315 systemd-tmpfiles[1678]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Oct 13 05:51:18.988279 systemd-tmpfiles[1678]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Oct 13 05:51:18.988633 systemd-tmpfiles[1678]: ACLs are not supported, ignoring. Oct 13 05:51:18.988807 systemd-tmpfiles[1678]: ACLs are not supported, ignoring. Oct 13 05:51:18.994177 systemd-tmpfiles[1678]: Detected autofs mount point /boot during canonicalization of boot. Oct 13 05:51:18.994190 systemd-tmpfiles[1678]: Skipping /boot Oct 13 05:51:19.005470 systemd-tmpfiles[1678]: Detected autofs mount point /boot during canonicalization of boot. Oct 13 05:51:19.005595 systemd-tmpfiles[1678]: Skipping /boot Oct 13 05:51:19.029550 systemd-udevd[1679]: Using default interface naming scheme 'v255'. Oct 13 05:51:19.071801 zram_generator::config[1708]: No configuration found. Oct 13 05:51:19.259295 (udev-worker)[1733]: Network interface NamePolicy= disabled on kernel command line. Oct 13 05:51:19.366806 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Oct 13 05:51:19.381802 kernel: mousedev: PS/2 mouse device common for all mice Oct 13 05:51:19.389807 kernel: ACPI: button: Power Button [PWRF] Oct 13 05:51:19.389878 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input4 Oct 13 05:51:19.427800 kernel: ACPI: button: Sleep Button [SLPF] Oct 13 05:51:19.440810 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Oct 13 05:51:19.567023 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Oct 13 05:51:19.567960 systemd[1]: Reloading finished in 583 ms. Oct 13 05:51:19.580849 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Oct 13 05:51:19.582637 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Oct 13 05:51:19.688223 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 13 05:51:19.693083 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 13 05:51:19.699119 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Oct 13 05:51:19.701126 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Oct 13 05:51:19.707154 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Oct 13 05:51:19.711951 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Oct 13 05:51:19.716751 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Oct 13 05:51:19.726188 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Oct 13 05:51:19.728177 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Oct 13 05:51:19.728394 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Oct 13 05:51:19.732673 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Oct 13 05:51:19.742163 systemd[1]: Starting systemd-networkd.service - Network Configuration... Oct 13 05:51:19.753913 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Oct 13 05:51:19.754648 systemd[1]: Reached target time-set.target - System Time Set. Oct 13 05:51:19.759302 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Oct 13 05:51:19.759986 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Oct 13 05:51:19.774946 systemd[1]: Finished ensure-sysext.service. Oct 13 05:51:19.820987 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Oct 13 05:51:19.830242 systemd[1]: modprobe@drm.service: Deactivated successfully. Oct 13 05:51:19.832905 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Oct 13 05:51:19.837278 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Oct 13 05:51:19.844112 systemd[1]: Starting systemd-update-done.service - Update is Completed... Oct 13 05:51:19.845495 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Oct 13 05:51:19.847009 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Oct 13 05:51:19.865209 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Oct 13 05:51:19.865487 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Oct 13 05:51:19.867737 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Oct 13 05:51:19.879990 systemd[1]: modprobe@loop.service: Deactivated successfully. Oct 13 05:51:19.880291 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Oct 13 05:51:19.882353 systemd[1]: Finished systemd-update-done.service - Update is Completed. Oct 13 05:51:19.896383 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Oct 13 05:51:19.905598 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Oct 13 05:51:19.929368 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Oct 13 05:51:19.930693 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Oct 13 05:51:19.937867 augenrules[1926]: No rules Oct 13 05:51:19.939900 systemd[1]: audit-rules.service: Deactivated successfully. Oct 13 05:51:19.940742 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 13 05:51:19.959230 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 05:51:19.994071 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Oct 13 05:51:19.996859 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Oct 13 05:51:20.026687 systemd[1]: Started systemd-userdbd.service - User Database Manager. Oct 13 05:51:20.056988 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Oct 13 05:51:20.086025 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Oct 13 05:51:20.086452 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 05:51:20.091769 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Oct 13 05:51:20.096975 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Oct 13 05:51:20.194292 systemd-networkd[1896]: lo: Link UP Oct 13 05:51:20.194305 systemd-networkd[1896]: lo: Gained carrier Oct 13 05:51:20.198286 systemd-networkd[1896]: Enumeration completed Oct 13 05:51:20.198803 systemd-networkd[1896]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 05:51:20.198809 systemd-networkd[1896]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Oct 13 05:51:20.200021 systemd[1]: Started systemd-networkd.service - Network Configuration. Oct 13 05:51:20.203692 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Oct 13 05:51:20.207765 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Oct 13 05:51:20.212131 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Oct 13 05:51:20.215179 systemd-networkd[1896]: eth0: Link UP Oct 13 05:51:20.215491 systemd-networkd[1896]: eth0: Gained carrier Oct 13 05:51:20.215604 systemd-networkd[1896]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Oct 13 05:51:20.230991 systemd-networkd[1896]: eth0: DHCPv4 address 172.31.26.230/20, gateway 172.31.16.1 acquired from 172.31.16.1 Oct 13 05:51:20.236543 systemd-resolved[1897]: Positive Trust Anchors: Oct 13 05:51:20.236564 systemd-resolved[1897]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Oct 13 05:51:20.236617 systemd-resolved[1897]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Oct 13 05:51:20.237151 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Oct 13 05:51:20.242508 systemd-resolved[1897]: Defaulting to hostname 'linux'. Oct 13 05:51:20.244335 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Oct 13 05:51:20.244947 systemd[1]: Reached target network.target - Network. Oct 13 05:51:20.245427 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Oct 13 05:51:20.245884 systemd[1]: Reached target sysinit.target - System Initialization. Oct 13 05:51:20.246432 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Oct 13 05:51:20.246888 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Oct 13 05:51:20.247295 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Oct 13 05:51:20.247882 systemd[1]: Started logrotate.timer - Daily rotation of log files. Oct 13 05:51:20.248395 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Oct 13 05:51:20.248817 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Oct 13 05:51:20.249195 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Oct 13 05:51:20.249238 systemd[1]: Reached target paths.target - Path Units. Oct 13 05:51:20.249620 systemd[1]: Reached target timers.target - Timer Units. Oct 13 05:51:20.251508 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Oct 13 05:51:20.253283 systemd[1]: Starting docker.socket - Docker Socket for the API... Oct 13 05:51:20.255889 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Oct 13 05:51:20.256514 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Oct 13 05:51:20.256969 systemd[1]: Reached target ssh-access.target - SSH Access Available. Oct 13 05:51:20.259680 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Oct 13 05:51:20.260594 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Oct 13 05:51:20.261796 systemd[1]: Listening on docker.socket - Docker Socket for the API. Oct 13 05:51:20.263167 systemd[1]: Reached target sockets.target - Socket Units. Oct 13 05:51:20.263638 systemd[1]: Reached target basic.target - Basic System. Oct 13 05:51:20.264114 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Oct 13 05:51:20.264151 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Oct 13 05:51:20.265237 systemd[1]: Starting containerd.service - containerd container runtime... Oct 13 05:51:20.268938 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Oct 13 05:51:20.272036 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Oct 13 05:51:20.277001 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Oct 13 05:51:20.280956 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Oct 13 05:51:20.284957 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Oct 13 05:51:20.288052 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Oct 13 05:51:20.294568 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Oct 13 05:51:20.323334 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Oct 13 05:51:20.333172 systemd[1]: Started ntpd.service - Network Time Service. Oct 13 05:51:20.336519 jq[1964]: false Oct 13 05:51:20.335317 oslogin_cache_refresh[1966]: Refreshing passwd entry cache Oct 13 05:51:20.337023 google_oslogin_nss_cache[1966]: oslogin_cache_refresh[1966]: Refreshing passwd entry cache Oct 13 05:51:20.341315 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Oct 13 05:51:20.342105 oslogin_cache_refresh[1966]: Failure getting users, quitting Oct 13 05:51:20.347432 google_oslogin_nss_cache[1966]: oslogin_cache_refresh[1966]: Failure getting users, quitting Oct 13 05:51:20.347432 google_oslogin_nss_cache[1966]: oslogin_cache_refresh[1966]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Oct 13 05:51:20.347432 google_oslogin_nss_cache[1966]: oslogin_cache_refresh[1966]: Refreshing group entry cache Oct 13 05:51:20.347432 google_oslogin_nss_cache[1966]: oslogin_cache_refresh[1966]: Failure getting groups, quitting Oct 13 05:51:20.347432 google_oslogin_nss_cache[1966]: oslogin_cache_refresh[1966]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Oct 13 05:51:20.342127 oslogin_cache_refresh[1966]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Oct 13 05:51:20.342181 oslogin_cache_refresh[1966]: Refreshing group entry cache Oct 13 05:51:20.346916 oslogin_cache_refresh[1966]: Failure getting groups, quitting Oct 13 05:51:20.346931 oslogin_cache_refresh[1966]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Oct 13 05:51:20.347972 systemd[1]: Starting setup-oem.service - Setup OEM... Oct 13 05:51:20.352867 extend-filesystems[1965]: Found /dev/nvme0n1p6 Oct 13 05:51:20.361843 extend-filesystems[1965]: Found /dev/nvme0n1p9 Oct 13 05:51:20.357651 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Oct 13 05:51:20.365642 extend-filesystems[1965]: Checking size of /dev/nvme0n1p9 Oct 13 05:51:20.364529 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Oct 13 05:51:20.381076 systemd[1]: Starting systemd-logind.service - User Login Management... Oct 13 05:51:20.384030 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Oct 13 05:51:20.389158 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Oct 13 05:51:20.392053 systemd[1]: Starting update-engine.service - Update Engine... Oct 13 05:51:20.398098 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Oct 13 05:51:20.405014 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Oct 13 05:51:20.406013 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Oct 13 05:51:20.406871 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Oct 13 05:51:20.407291 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Oct 13 05:51:20.408840 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Oct 13 05:51:20.423339 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Oct 13 05:51:20.425503 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Oct 13 05:51:20.442140 jq[1985]: true Oct 13 05:51:20.450930 extend-filesystems[1965]: Resized partition /dev/nvme0n1p9 Oct 13 05:51:20.472236 extend-filesystems[2008]: resize2fs 1.47.3 (8-Jul-2025) Oct 13 05:51:20.477565 systemd[1]: motdgen.service: Deactivated successfully. Oct 13 05:51:20.478964 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Oct 13 05:51:20.496380 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 3587067 blocks Oct 13 05:51:20.503264 (ntainerd)[2012]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Oct 13 05:51:20.526124 update_engine[1984]: I20251013 05:51:20.508567 1984 main.cc:92] Flatcar Update Engine starting Oct 13 05:51:20.541648 tar[1989]: linux-amd64/LICENSE Oct 13 05:51:20.544480 ntpd[1972]: ntpd 4.2.8p18@1.4062-o Sun Oct 12 22:02:52 UTC 2025 (1): Starting Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: ntpd 4.2.8p18@1.4062-o Sun Oct 12 22:02:52 UTC 2025 (1): Starting Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: ---------------------------------------------------- Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: ntp-4 is maintained by Network Time Foundation, Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: corporation. Support and training for ntp-4 are Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: available at https://www.nwtime.org/support Oct 13 05:51:20.545164 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: ---------------------------------------------------- Oct 13 05:51:20.544546 ntpd[1972]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Oct 13 05:51:20.549574 tar[1989]: linux-amd64/helm Oct 13 05:51:20.544556 ntpd[1972]: ---------------------------------------------------- Oct 13 05:51:20.544565 ntpd[1972]: ntp-4 is maintained by Network Time Foundation, Oct 13 05:51:20.544575 ntpd[1972]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Oct 13 05:51:20.544583 ntpd[1972]: corporation. Support and training for ntp-4 are Oct 13 05:51:20.544592 ntpd[1972]: available at https://www.nwtime.org/support Oct 13 05:51:20.544601 ntpd[1972]: ---------------------------------------------------- Oct 13 05:51:20.553699 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: proto: precision = 0.090 usec (-23) Oct 13 05:51:20.553740 jq[2004]: true Oct 13 05:51:20.552858 ntpd[1972]: proto: precision = 0.090 usec (-23) Oct 13 05:51:20.558220 ntpd[1972]: basedate set to 2025-09-30 Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: basedate set to 2025-09-30 Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: gps base set to 2025-10-05 (week 2387) Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: Listen and drop on 0 v6wildcard [::]:123 Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: Listen normally on 2 lo 127.0.0.1:123 Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: Listen normally on 3 eth0 172.31.26.230:123 Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: Listen normally on 4 lo [::1]:123 Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: bind(21) AF_INET6 [fe80::4d8:10ff:fe60:82cb%2]:123 flags 0x811 failed: Cannot assign requested address Oct 13 05:51:20.559172 ntpd[1972]: 13 Oct 05:51:20 ntpd[1972]: unable to create socket on eth0 (5) for [fe80::4d8:10ff:fe60:82cb%2]:123 Oct 13 05:51:20.558321 ntpd[1972]: gps base set to 2025-10-05 (week 2387) Oct 13 05:51:20.558478 ntpd[1972]: Listen and drop on 0 v6wildcard [::]:123 Oct 13 05:51:20.558508 ntpd[1972]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Oct 13 05:51:20.558738 ntpd[1972]: Listen normally on 2 lo 127.0.0.1:123 Oct 13 05:51:20.558767 ntpd[1972]: Listen normally on 3 eth0 172.31.26.230:123 Oct 13 05:51:20.558815 ntpd[1972]: Listen normally on 4 lo [::1]:123 Oct 13 05:51:20.558847 ntpd[1972]: bind(21) AF_INET6 [fe80::4d8:10ff:fe60:82cb%2]:123 flags 0x811 failed: Cannot assign requested address Oct 13 05:51:20.560136 kernel: ntpd[1972]: segfault at 24 ip 00005574ed6cfaeb sp 00007ffdaff974f0 error 4 in ntpd[68aeb,5574ed66d000+80000] likely on CPU 1 (core 0, socket 0) Oct 13 05:51:20.558867 ntpd[1972]: unable to create socket on eth0 (5) for [fe80::4d8:10ff:fe60:82cb%2]:123 Oct 13 05:51:20.561913 kernel: Code: 0f 1e fa 41 56 41 55 41 54 55 53 48 89 fb e8 8c eb f9 ff 44 8b 28 49 89 c4 e8 51 6b ff ff 48 89 c5 48 85 db 0f 84 a5 00 00 00 <0f> b7 0b 66 83 f9 02 0f 84 c0 00 00 00 66 83 f9 0a 74 32 66 85 c9 Oct 13 05:51:20.582225 systemd-coredump[2020]: Process 1972 (ntpd) of user 0 terminated abnormally with signal 11/SEGV, processing... Oct 13 05:51:20.609320 dbus-daemon[1962]: [system] SELinux support is enabled Oct 13 05:51:20.587057 systemd[1]: Created slice system-systemd\x2dcoredump.slice - Slice /system/systemd-coredump. Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.599 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.607 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.608 INFO Fetch successful Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.608 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.611 INFO Fetch successful Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.611 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.614 INFO Fetch successful Oct 13 05:51:20.623841 coreos-metadata[1961]: Oct 13 05:51:20.614 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Oct 13 05:51:20.633908 update_engine[1984]: I20251013 05:51:20.632968 1984 update_check_scheduler.cc:74] Next update check in 7m26s Oct 13 05:51:20.630815 dbus-daemon[1962]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.1' (uid=244 pid=1896 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Oct 13 05:51:20.629593 systemd[1]: Started systemd-coredump@0-2020-0.service - Process Core Dump (PID 2020/UID 0). Oct 13 05:51:20.634469 coreos-metadata[1961]: Oct 13 05:51:20.629 INFO Fetch successful Oct 13 05:51:20.634469 coreos-metadata[1961]: Oct 13 05:51:20.629 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Oct 13 05:51:20.630511 systemd[1]: Started dbus.service - D-Bus System Message Bus. Oct 13 05:51:20.639835 coreos-metadata[1961]: Oct 13 05:51:20.638 INFO Fetch failed with 404: resource not found Oct 13 05:51:20.639835 coreos-metadata[1961]: Oct 13 05:51:20.638 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Oct 13 05:51:20.647194 coreos-metadata[1961]: Oct 13 05:51:20.642 INFO Fetch successful Oct 13 05:51:20.647194 coreos-metadata[1961]: Oct 13 05:51:20.643 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Oct 13 05:51:20.643938 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Oct 13 05:51:20.644003 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Oct 13 05:51:20.645952 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Oct 13 05:51:20.645985 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Oct 13 05:51:20.650831 coreos-metadata[1961]: Oct 13 05:51:20.648 INFO Fetch successful Oct 13 05:51:20.650831 coreos-metadata[1961]: Oct 13 05:51:20.648 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Oct 13 05:51:20.657142 coreos-metadata[1961]: Oct 13 05:51:20.654 INFO Fetch successful Oct 13 05:51:20.657142 coreos-metadata[1961]: Oct 13 05:51:20.654 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Oct 13 05:51:20.657897 coreos-metadata[1961]: Oct 13 05:51:20.657 INFO Fetch successful Oct 13 05:51:20.658197 systemd[1]: Started update-engine.service - Update Engine. Oct 13 05:51:20.659752 coreos-metadata[1961]: Oct 13 05:51:20.659 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Oct 13 05:51:20.661062 dbus-daemon[1962]: [system] Successfully activated service 'org.freedesktop.systemd1' Oct 13 05:51:20.662843 coreos-metadata[1961]: Oct 13 05:51:20.661 INFO Fetch successful Oct 13 05:51:20.706460 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 3587067 Oct 13 05:51:20.712028 systemd[1]: Started locksmithd.service - Cluster reboot manager. Oct 13 05:51:20.714417 systemd[1]: Finished setup-oem.service - Setup OEM. Oct 13 05:51:20.721519 systemd-logind[1980]: Watching system buttons on /dev/input/event2 (Power Button) Oct 13 05:51:20.726053 systemd-logind[1980]: Watching system buttons on /dev/input/event3 (Sleep Button) Oct 13 05:51:20.726087 systemd-logind[1980]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Oct 13 05:51:20.726542 systemd-logind[1980]: New seat seat0. Oct 13 05:51:20.732281 systemd[1]: Started systemd-logind.service - User Login Management. Oct 13 05:51:20.741314 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Oct 13 05:51:20.744584 extend-filesystems[2008]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Oct 13 05:51:20.744584 extend-filesystems[2008]: old_desc_blocks = 1, new_desc_blocks = 2 Oct 13 05:51:20.744584 extend-filesystems[2008]: The filesystem on /dev/nvme0n1p9 is now 3587067 (4k) blocks long. Oct 13 05:51:20.754094 extend-filesystems[1965]: Resized filesystem in /dev/nvme0n1p9 Oct 13 05:51:20.747298 systemd[1]: extend-filesystems.service: Deactivated successfully. Oct 13 05:51:20.748329 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Oct 13 05:51:20.762922 bash[2044]: Updated "/home/core/.ssh/authorized_keys" Oct 13 05:51:20.764911 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Oct 13 05:51:20.770928 systemd[1]: Starting sshkeys.service... Oct 13 05:51:20.883942 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Oct 13 05:51:20.890185 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Oct 13 05:51:20.916756 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Oct 13 05:51:20.918208 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Oct 13 05:51:21.187027 coreos-metadata[2087]: Oct 13 05:51:21.186 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Oct 13 05:51:21.188500 coreos-metadata[2087]: Oct 13 05:51:21.188 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Oct 13 05:51:21.189362 coreos-metadata[2087]: Oct 13 05:51:21.189 INFO Fetch successful Oct 13 05:51:21.189362 coreos-metadata[2087]: Oct 13 05:51:21.189 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Oct 13 05:51:21.190335 coreos-metadata[2087]: Oct 13 05:51:21.189 INFO Fetch successful Oct 13 05:51:21.198738 unknown[2087]: wrote ssh authorized keys file for user: core Oct 13 05:51:21.248024 systemd-coredump[2034]: Process 1972 (ntpd) of user 0 dumped core. Module libnss_usrfiles.so.2 without build-id. Module libgcc_s.so.1 without build-id. Module ld-linux-x86-64.so.2 without build-id. Module libc.so.6 without build-id. Module libcrypto.so.3 without build-id. Module libm.so.6 without build-id. Module libcap.so.2 without build-id. Module ntpd without build-id. Stack trace of thread 1972: #0 0x00005574ed6cfaeb n/a (ntpd + 0x68aeb) #1 0x00005574ed678cdf n/a (ntpd + 0x11cdf) #2 0x00005574ed679575 n/a (ntpd + 0x12575) #3 0x00005574ed674d8a n/a (ntpd + 0xdd8a) #4 0x00005574ed6765d3 n/a (ntpd + 0xf5d3) #5 0x00005574ed67efd1 n/a (ntpd + 0x17fd1) #6 0x00005574ed66fc2d n/a (ntpd + 0x8c2d) #7 0x00007f0d51f2f16c n/a (libc.so.6 + 0x2716c) #8 0x00007f0d51f2f229 __libc_start_main (libc.so.6 + 0x27229) #9 0x00005574ed66fc55 n/a (ntpd + 0x8c55) ELF object binary architecture: AMD x86-64 Oct 13 05:51:21.258371 systemd[1]: ntpd.service: Main process exited, code=dumped, status=11/SEGV Oct 13 05:51:21.258562 systemd[1]: ntpd.service: Failed with result 'core-dump'. Oct 13 05:51:21.268204 systemd[1]: systemd-coredump@0-2020-0.service: Deactivated successfully. Oct 13 05:51:21.310997 update-ssh-keys[2161]: Updated "/home/core/.ssh/authorized_keys" Oct 13 05:51:21.313926 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Oct 13 05:51:21.322117 systemd[1]: Finished sshkeys.service. Oct 13 05:51:21.326375 locksmithd[2040]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Oct 13 05:51:21.375863 systemd[1]: ntpd.service: Scheduled restart job, restart counter is at 1. Oct 13 05:51:21.376994 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Oct 13 05:51:21.381696 systemd[1]: Started ntpd.service - Network Time Service. Oct 13 05:51:21.385425 dbus-daemon[1962]: [system] Successfully activated service 'org.freedesktop.hostname1' Oct 13 05:51:21.392210 dbus-daemon[1962]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.7' (uid=0 pid=2050 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Oct 13 05:51:21.403111 systemd[1]: Starting polkit.service - Authorization Manager... Oct 13 05:51:21.406888 containerd[2012]: time="2025-10-13T05:51:21Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Oct 13 05:51:21.413197 containerd[2012]: time="2025-10-13T05:51:21.413140582Z" level=info msg="starting containerd" revision=fb4c30d4ede3531652d86197bf3fc9515e5276d9 version=v2.0.5 Oct 13 05:51:21.469479 ntpd[2173]: ntpd 4.2.8p18@1.4062-o Sun Oct 12 22:02:52 UTC 2025 (1): Starting Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: ntpd 4.2.8p18@1.4062-o Sun Oct 12 22:02:52 UTC 2025 (1): Starting Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: ---------------------------------------------------- Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: ntp-4 is maintained by Network Time Foundation, Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: corporation. Support and training for ntp-4 are Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: available at https://www.nwtime.org/support Oct 13 05:51:21.469937 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: ---------------------------------------------------- Oct 13 05:51:21.469557 ntpd[2173]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Oct 13 05:51:21.469568 ntpd[2173]: ---------------------------------------------------- Oct 13 05:51:21.469578 ntpd[2173]: ntp-4 is maintained by Network Time Foundation, Oct 13 05:51:21.469586 ntpd[2173]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Oct 13 05:51:21.469594 ntpd[2173]: corporation. Support and training for ntp-4 are Oct 13 05:51:21.469601 ntpd[2173]: available at https://www.nwtime.org/support Oct 13 05:51:21.469610 ntpd[2173]: ---------------------------------------------------- Oct 13 05:51:21.473429 ntpd[2173]: proto: precision = 0.076 usec (-24) Oct 13 05:51:21.474940 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: proto: precision = 0.076 usec (-24) Oct 13 05:51:21.474940 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: basedate set to 2025-09-30 Oct 13 05:51:21.474940 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: gps base set to 2025-10-05 (week 2387) Oct 13 05:51:21.474940 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: Listen and drop on 0 v6wildcard [::]:123 Oct 13 05:51:21.474940 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Oct 13 05:51:21.473697 ntpd[2173]: basedate set to 2025-09-30 Oct 13 05:51:21.475195 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: Listen normally on 2 lo 127.0.0.1:123 Oct 13 05:51:21.475195 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: Listen normally on 3 eth0 172.31.26.230:123 Oct 13 05:51:21.475195 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: Listen normally on 4 lo [::1]:123 Oct 13 05:51:21.475195 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: bind(21) AF_INET6 [fe80::4d8:10ff:fe60:82cb%2]:123 flags 0x811 failed: Cannot assign requested address Oct 13 05:51:21.475195 ntpd[2173]: 13 Oct 05:51:21 ntpd[2173]: unable to create socket on eth0 (5) for [fe80::4d8:10ff:fe60:82cb%2]:123 Oct 13 05:51:21.473709 ntpd[2173]: gps base set to 2025-10-05 (week 2387) Oct 13 05:51:21.474842 ntpd[2173]: Listen and drop on 0 v6wildcard [::]:123 Oct 13 05:51:21.474876 ntpd[2173]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Oct 13 05:51:21.475064 ntpd[2173]: Listen normally on 2 lo 127.0.0.1:123 Oct 13 05:51:21.475091 ntpd[2173]: Listen normally on 3 eth0 172.31.26.230:123 Oct 13 05:51:21.475121 ntpd[2173]: Listen normally on 4 lo [::1]:123 Oct 13 05:51:21.475152 ntpd[2173]: bind(21) AF_INET6 [fe80::4d8:10ff:fe60:82cb%2]:123 flags 0x811 failed: Cannot assign requested address Oct 13 05:51:21.475172 ntpd[2173]: unable to create socket on eth0 (5) for [fe80::4d8:10ff:fe60:82cb%2]:123 Oct 13 05:51:21.477479 kernel: ntpd[2173]: segfault at 24 ip 0000558ba4657aeb sp 00007ffdb8ee0030 error 4 in ntpd[68aeb,558ba45f5000+80000] likely on CPU 1 (core 0, socket 0) Oct 13 05:51:21.480391 kernel: Code: 0f 1e fa 41 56 41 55 41 54 55 53 48 89 fb e8 8c eb f9 ff 44 8b 28 49 89 c4 e8 51 6b ff ff 48 89 c5 48 85 db 0f 84 a5 00 00 00 <0f> b7 0b 66 83 f9 02 0f 84 c0 00 00 00 66 83 f9 0a 74 32 66 85 c9 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486564647Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="12.54µs" Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486606299Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486631299Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486830850Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486849990Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486880251Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486940376Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.486957204Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.487240132Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.487258724Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.487273096Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Oct 13 05:51:21.487697 containerd[2012]: time="2025-10-13T05:51:21.487284561Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Oct 13 05:51:21.488216 containerd[2012]: time="2025-10-13T05:51:21.487362633Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Oct 13 05:51:21.488216 containerd[2012]: time="2025-10-13T05:51:21.487597657Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 13 05:51:21.488216 containerd[2012]: time="2025-10-13T05:51:21.487633522Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Oct 13 05:51:21.488216 containerd[2012]: time="2025-10-13T05:51:21.487656772Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Oct 13 05:51:21.492545 containerd[2012]: time="2025-10-13T05:51:21.491899292Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Oct 13 05:51:21.492545 containerd[2012]: time="2025-10-13T05:51:21.492249333Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Oct 13 05:51:21.492545 containerd[2012]: time="2025-10-13T05:51:21.492347508Z" level=info msg="metadata content store policy set" policy=shared Oct 13 05:51:21.496693 containerd[2012]: time="2025-10-13T05:51:21.496562803Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Oct 13 05:51:21.496693 containerd[2012]: time="2025-10-13T05:51:21.496644756Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.496672351Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.496935196Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.496960992Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.496983739Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497000862Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497018709Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497036698Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497051349Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497064454Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497082179Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497229501Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497255996Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497277970Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Oct 13 05:51:21.497858 containerd[2012]: time="2025-10-13T05:51:21.497306823Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497325467Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497341283Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497359817Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497377380Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497395420Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497412872Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497427595Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497512614Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Oct 13 05:51:21.498414 containerd[2012]: time="2025-10-13T05:51:21.497535289Z" level=info msg="Start snapshots syncer" Oct 13 05:51:21.501617 containerd[2012]: time="2025-10-13T05:51:21.500311141Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Oct 13 05:51:21.501617 containerd[2012]: time="2025-10-13T05:51:21.500768621Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Oct 13 05:51:21.501889 containerd[2012]: time="2025-10-13T05:51:21.500857366Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Oct 13 05:51:21.502101 systemd-coredump[2178]: Process 2173 (ntpd) of user 0 terminated abnormally with signal 11/SEGV, processing... Oct 13 05:51:21.504312 containerd[2012]: time="2025-10-13T05:51:21.504270043Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Oct 13 05:51:21.504507 containerd[2012]: time="2025-10-13T05:51:21.504483251Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Oct 13 05:51:21.504557 containerd[2012]: time="2025-10-13T05:51:21.504526057Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Oct 13 05:51:21.504557 containerd[2012]: time="2025-10-13T05:51:21.504547873Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Oct 13 05:51:21.504640 containerd[2012]: time="2025-10-13T05:51:21.504568836Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Oct 13 05:51:21.504640 containerd[2012]: time="2025-10-13T05:51:21.504587291Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Oct 13 05:51:21.504640 containerd[2012]: time="2025-10-13T05:51:21.504603421Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Oct 13 05:51:21.504640 containerd[2012]: time="2025-10-13T05:51:21.504620063Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Oct 13 05:51:21.504772 containerd[2012]: time="2025-10-13T05:51:21.504658144Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Oct 13 05:51:21.504772 containerd[2012]: time="2025-10-13T05:51:21.504674900Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Oct 13 05:51:21.504772 containerd[2012]: time="2025-10-13T05:51:21.504691330Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Oct 13 05:51:21.509524 containerd[2012]: time="2025-10-13T05:51:21.509424402Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 13 05:51:21.509632 containerd[2012]: time="2025-10-13T05:51:21.509534317Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Oct 13 05:51:21.509632 containerd[2012]: time="2025-10-13T05:51:21.509549677Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 13 05:51:21.509632 containerd[2012]: time="2025-10-13T05:51:21.509565763Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Oct 13 05:51:21.509632 containerd[2012]: time="2025-10-13T05:51:21.509577025Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Oct 13 05:51:21.509632 containerd[2012]: time="2025-10-13T05:51:21.509600054Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Oct 13 05:51:21.509632 containerd[2012]: time="2025-10-13T05:51:21.509616757Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Oct 13 05:51:21.509864 containerd[2012]: time="2025-10-13T05:51:21.509638024Z" level=info msg="runtime interface created" Oct 13 05:51:21.509864 containerd[2012]: time="2025-10-13T05:51:21.509645792Z" level=info msg="created NRI interface" Oct 13 05:51:21.509864 containerd[2012]: time="2025-10-13T05:51:21.509658637Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Oct 13 05:51:21.509864 containerd[2012]: time="2025-10-13T05:51:21.509682297Z" level=info msg="Connect containerd service" Oct 13 05:51:21.509864 containerd[2012]: time="2025-10-13T05:51:21.509732078Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Oct 13 05:51:21.510325 systemd[1]: Started systemd-coredump@1-2178-0.service - Process Core Dump (PID 2178/UID 0). Oct 13 05:51:21.516791 containerd[2012]: time="2025-10-13T05:51:21.516735106Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Oct 13 05:51:21.781932 polkitd[2174]: Started polkitd version 126 Oct 13 05:51:21.792516 polkitd[2174]: Loading rules from directory /etc/polkit-1/rules.d Oct 13 05:51:21.803903 polkitd[2174]: Loading rules from directory /run/polkit-1/rules.d Oct 13 05:51:21.803983 polkitd[2174]: Error opening rules directory: Error opening directory “/run/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Oct 13 05:51:21.804403 polkitd[2174]: Loading rules from directory /usr/local/share/polkit-1/rules.d Oct 13 05:51:21.804444 polkitd[2174]: Error opening rules directory: Error opening directory “/usr/local/share/polkit-1/rules.d”: No such file or directory (g-file-error-quark, 4) Oct 13 05:51:21.804498 polkitd[2174]: Loading rules from directory /usr/share/polkit-1/rules.d Oct 13 05:51:21.809706 polkitd[2174]: Finished loading, compiling and executing 2 rules Oct 13 05:51:21.814594 systemd[1]: Started polkit.service - Authorization Manager. Oct 13 05:51:21.823697 dbus-daemon[1962]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Oct 13 05:51:21.824076 polkitd[2174]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Oct 13 05:51:21.857994 systemd-networkd[1896]: eth0: Gained IPv6LL Oct 13 05:51:21.861802 systemd-hostnamed[2050]: Hostname set to (transient) Oct 13 05:51:21.863834 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Oct 13 05:51:21.865888 systemd-resolved[1897]: System hostname changed to 'ip-172-31-26-230'. Oct 13 05:51:21.867338 systemd[1]: Reached target network-online.target - Network is Online. Oct 13 05:51:21.872719 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Oct 13 05:51:21.875984 containerd[2012]: time="2025-10-13T05:51:21.875930956Z" level=info msg="Start subscribing containerd event" Oct 13 05:51:21.876089 containerd[2012]: time="2025-10-13T05:51:21.876009041Z" level=info msg="Start recovering state" Oct 13 05:51:21.876156 containerd[2012]: time="2025-10-13T05:51:21.876138849Z" level=info msg="Start event monitor" Oct 13 05:51:21.876194 containerd[2012]: time="2025-10-13T05:51:21.876163068Z" level=info msg="Start cni network conf syncer for default" Oct 13 05:51:21.876194 containerd[2012]: time="2025-10-13T05:51:21.876176060Z" level=info msg="Start streaming server" Oct 13 05:51:21.876264 containerd[2012]: time="2025-10-13T05:51:21.876200888Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Oct 13 05:51:21.876264 containerd[2012]: time="2025-10-13T05:51:21.876212279Z" level=info msg="runtime interface starting up..." Oct 13 05:51:21.876264 containerd[2012]: time="2025-10-13T05:51:21.876220972Z" level=info msg="starting plugins..." Oct 13 05:51:21.876264 containerd[2012]: time="2025-10-13T05:51:21.876237400Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Oct 13 05:51:21.878142 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 05:51:21.885245 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Oct 13 05:51:21.891016 containerd[2012]: time="2025-10-13T05:51:21.885658540Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Oct 13 05:51:21.891016 containerd[2012]: time="2025-10-13T05:51:21.885851198Z" level=info msg=serving... address=/run/containerd/containerd.sock Oct 13 05:51:21.891016 containerd[2012]: time="2025-10-13T05:51:21.886598583Z" level=info msg="containerd successfully booted in 0.480582s" Oct 13 05:51:21.887554 systemd[1]: Started containerd.service - containerd container runtime. Oct 13 05:51:21.922999 systemd-coredump[2179]: Process 2173 (ntpd) of user 0 dumped core. Module libnss_usrfiles.so.2 without build-id. Module libgcc_s.so.1 without build-id. Module ld-linux-x86-64.so.2 without build-id. Module libc.so.6 without build-id. Module libcrypto.so.3 without build-id. Module libm.so.6 without build-id. Module libcap.so.2 without build-id. Module ntpd without build-id. Stack trace of thread 2173: #0 0x0000558ba4657aeb n/a (ntpd + 0x68aeb) #1 0x0000558ba4600cdf n/a (ntpd + 0x11cdf) #2 0x0000558ba4601575 n/a (ntpd + 0x12575) #3 0x0000558ba45fcd8a n/a (ntpd + 0xdd8a) #4 0x0000558ba45fe5d3 n/a (ntpd + 0xf5d3) #5 0x0000558ba4606fd1 n/a (ntpd + 0x17fd1) #6 0x0000558ba45f7c2d n/a (ntpd + 0x8c2d) #7 0x00007f127da1816c n/a (libc.so.6 + 0x2716c) #8 0x00007f127da18229 __libc_start_main (libc.so.6 + 0x27229) #9 0x0000558ba45f7c55 n/a (ntpd + 0x8c55) ELF object binary architecture: AMD x86-64 Oct 13 05:51:21.928510 systemd[1]: ntpd.service: Main process exited, code=dumped, status=11/SEGV Oct 13 05:51:21.928711 systemd[1]: ntpd.service: Failed with result 'core-dump'. Oct 13 05:51:21.935722 systemd[1]: systemd-coredump@1-2178-0.service: Deactivated successfully. Oct 13 05:51:21.977141 sshd_keygen[2016]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Oct 13 05:51:21.997923 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Oct 13 05:51:22.041946 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Oct 13 05:51:22.045381 systemd[1]: ntpd.service: Scheduled restart job, restart counter is at 2. Oct 13 05:51:22.050627 systemd[1]: Starting issuegen.service - Generate /run/issue... Oct 13 05:51:22.054150 systemd[1]: Started ntpd.service - Network Time Service. Oct 13 05:51:22.089425 systemd[1]: issuegen.service: Deactivated successfully. Oct 13 05:51:22.089700 systemd[1]: Finished issuegen.service - Generate /run/issue. Oct 13 05:51:22.099272 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Oct 13 05:51:22.111169 ntpd[2236]: ntpd 4.2.8p18@1.4062-o Sun Oct 12 22:02:52 UTC 2025 (1): Starting Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: ntpd 4.2.8p18@1.4062-o Sun Oct 12 22:02:52 UTC 2025 (1): Starting Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: ---------------------------------------------------- Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: ntp-4 is maintained by Network Time Foundation, Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: corporation. Support and training for ntp-4 are Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: available at https://www.nwtime.org/support Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: ---------------------------------------------------- Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: proto: precision = 0.095 usec (-23) Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: basedate set to 2025-09-30 Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: gps base set to 2025-10-05 (week 2387) Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Listen and drop on 0 v6wildcard [::]:123 Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Oct 13 05:51:22.113156 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Listen normally on 2 lo 127.0.0.1:123 Oct 13 05:51:22.111258 ntpd[2236]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Oct 13 05:51:22.111270 ntpd[2236]: ---------------------------------------------------- Oct 13 05:51:22.111281 ntpd[2236]: ntp-4 is maintained by Network Time Foundation, Oct 13 05:51:22.111292 ntpd[2236]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Oct 13 05:51:22.111302 ntpd[2236]: corporation. Support and training for ntp-4 are Oct 13 05:51:22.111312 ntpd[2236]: available at https://www.nwtime.org/support Oct 13 05:51:22.111325 ntpd[2236]: ---------------------------------------------------- Oct 13 05:51:22.112188 ntpd[2236]: proto: precision = 0.095 usec (-23) Oct 13 05:51:22.112446 ntpd[2236]: basedate set to 2025-09-30 Oct 13 05:51:22.112459 ntpd[2236]: gps base set to 2025-10-05 (week 2387) Oct 13 05:51:22.112547 ntpd[2236]: Listen and drop on 0 v6wildcard [::]:123 Oct 13 05:51:22.112575 ntpd[2236]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Oct 13 05:51:22.112764 ntpd[2236]: Listen normally on 2 lo 127.0.0.1:123 Oct 13 05:51:22.117924 ntpd[2236]: Listen normally on 3 eth0 172.31.26.230:123 Oct 13 05:51:22.119430 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Listen normally on 3 eth0 172.31.26.230:123 Oct 13 05:51:22.119430 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Listen normally on 4 lo [::1]:123 Oct 13 05:51:22.119430 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Listen normally on 5 eth0 [fe80::4d8:10ff:fe60:82cb%2]:123 Oct 13 05:51:22.119430 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: Listening on routing socket on fd #22 for interface updates Oct 13 05:51:22.117985 ntpd[2236]: Listen normally on 4 lo [::1]:123 Oct 13 05:51:22.118015 ntpd[2236]: Listen normally on 5 eth0 [fe80::4d8:10ff:fe60:82cb%2]:123 Oct 13 05:51:22.118045 ntpd[2236]: Listening on routing socket on fd #22 for interface updates Oct 13 05:51:22.124733 ntpd[2236]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Oct 13 05:51:22.128958 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Oct 13 05:51:22.128958 ntpd[2236]: 13 Oct 05:51:22 ntpd[2236]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Oct 13 05:51:22.127757 ntpd[2236]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Oct 13 05:51:22.131346 amazon-ssm-agent[2202]: Initializing new seelog logger Oct 13 05:51:22.133579 amazon-ssm-agent[2202]: New Seelog Logger Creation Complete Oct 13 05:51:22.133579 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.133579 amazon-ssm-agent[2202]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.133579 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 processing appconfig overrides Oct 13 05:51:22.138807 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.138807 amazon-ssm-agent[2202]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.138807 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 processing appconfig overrides Oct 13 05:51:22.138807 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.138807 amazon-ssm-agent[2202]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.138807 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 processing appconfig overrides Oct 13 05:51:22.141225 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.1362 INFO Proxy environment variables: Oct 13 05:51:22.142935 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.142935 amazon-ssm-agent[2202]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.143053 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 processing appconfig overrides Oct 13 05:51:22.147364 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Oct 13 05:51:22.152279 systemd[1]: Started getty@tty1.service - Getty on tty1. Oct 13 05:51:22.156384 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Oct 13 05:51:22.157960 systemd[1]: Reached target getty.target - Login Prompts. Oct 13 05:51:22.240999 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.1363 INFO http_proxy: Oct 13 05:51:22.243469 tar[1989]: linux-amd64/README.md Oct 13 05:51:22.272218 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Oct 13 05:51:22.339481 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.1363 INFO no_proxy: Oct 13 05:51:22.438500 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.1367 INFO https_proxy: Oct 13 05:51:22.536437 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.1381 INFO Checking if agent identity type OnPrem can be assumed Oct 13 05:51:22.561800 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.561800 amazon-ssm-agent[2202]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Oct 13 05:51:22.563354 amazon-ssm-agent[2202]: 2025/10/13 05:51:22 processing appconfig overrides Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.1383 INFO Checking if agent identity type EC2 can be assumed Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2177 INFO Agent will take identity from EC2 Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2210 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.3.0.0 Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2210 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2210 INFO [amazon-ssm-agent] Starting Core Agent Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2210 INFO [amazon-ssm-agent] Registrar detected. Attempting registration Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2210 INFO [Registrar] Starting registrar module Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2238 INFO [EC2Identity] Checking disk for registration info Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2238 INFO [EC2Identity] No registration info found for ec2 instance, attempting registration Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.2238 INFO [EC2Identity] Generating registration keypair Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5201 INFO [EC2Identity] Checking write access before registering Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5205 INFO [EC2Identity] Registering EC2 instance with Systems Manager Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5613 INFO [EC2Identity] EC2 registration was successful. Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5615 INFO [amazon-ssm-agent] Registration attempted. Resuming core agent startup. Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5616 INFO [CredentialRefresher] credentialRefresher has started Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5616 INFO [CredentialRefresher] Starting credentials refresher loop Oct 13 05:51:22.595834 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5951 INFO EC2RoleProvider Successfully connected with instance profile role credentials Oct 13 05:51:22.596261 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5953 INFO [CredentialRefresher] Credentials ready Oct 13 05:51:22.635835 amazon-ssm-agent[2202]: 2025-10-13 05:51:22.5958 INFO [CredentialRefresher] Next credential rotation will be in 29.999987811633332 minutes Oct 13 05:51:23.573413 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Oct 13 05:51:23.575318 systemd[1]: Started sshd@0-172.31.26.230:22-139.178.68.195:45830.service - OpenSSH per-connection server daemon (139.178.68.195:45830). Oct 13 05:51:23.607223 amazon-ssm-agent[2202]: 2025-10-13 05:51:23.6070 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Oct 13 05:51:23.707614 amazon-ssm-agent[2202]: 2025-10-13 05:51:23.6090 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2256) started Oct 13 05:51:23.797810 sshd[2253]: Accepted publickey for core from 139.178.68.195 port 45830 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:51:23.801266 sshd-session[2253]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:51:23.807922 amazon-ssm-agent[2202]: 2025-10-13 05:51:23.6090 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Oct 13 05:51:23.810670 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Oct 13 05:51:23.813320 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Oct 13 05:51:23.829710 systemd-logind[1980]: New session 1 of user core. Oct 13 05:51:23.841347 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Oct 13 05:51:23.847705 systemd[1]: Starting user@500.service - User Manager for UID 500... Oct 13 05:51:23.865815 (systemd)[2272]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Oct 13 05:51:23.869738 systemd-logind[1980]: New session c1 of user core. Oct 13 05:51:24.054934 systemd[2272]: Queued start job for default target default.target. Oct 13 05:51:24.058181 systemd[2272]: Created slice app.slice - User Application Slice. Oct 13 05:51:24.058214 systemd[2272]: Reached target paths.target - Paths. Oct 13 05:51:24.058864 systemd[2272]: Reached target timers.target - Timers. Oct 13 05:51:24.061899 systemd[2272]: Starting dbus.socket - D-Bus User Message Bus Socket... Oct 13 05:51:24.081920 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 05:51:24.085461 systemd[1]: Reached target multi-user.target - Multi-User System. Oct 13 05:51:24.087074 systemd[2272]: Listening on dbus.socket - D-Bus User Message Bus Socket. Oct 13 05:51:24.087297 systemd[2272]: Reached target sockets.target - Sockets. Oct 13 05:51:24.087419 systemd[2272]: Reached target basic.target - Basic System. Oct 13 05:51:24.087544 systemd[2272]: Reached target default.target - Main User Target. Oct 13 05:51:24.087633 systemd[2272]: Startup finished in 207ms. Oct 13 05:51:24.088440 systemd[1]: Started user@500.service - User Manager for UID 500. Oct 13 05:51:24.100488 (kubelet)[2283]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 05:51:24.100697 systemd[1]: Started session-1.scope - Session 1 of User core. Oct 13 05:51:24.102840 systemd[1]: Startup finished in 2.753s (kernel) + 6.949s (initrd) + 7.516s (userspace) = 17.219s. Oct 13 05:51:24.257869 systemd[1]: Started sshd@1-172.31.26.230:22-139.178.68.195:45844.service - OpenSSH per-connection server daemon (139.178.68.195:45844). Oct 13 05:51:24.442065 sshd[2293]: Accepted publickey for core from 139.178.68.195 port 45844 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:51:24.444168 sshd-session[2293]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:51:24.451852 systemd-logind[1980]: New session 2 of user core. Oct 13 05:51:24.455988 systemd[1]: Started session-2.scope - Session 2 of User core. Oct 13 05:51:24.578371 sshd[2300]: Connection closed by 139.178.68.195 port 45844 Oct 13 05:51:24.579085 sshd-session[2293]: pam_unix(sshd:session): session closed for user core Oct 13 05:51:24.584744 systemd[1]: sshd@1-172.31.26.230:22-139.178.68.195:45844.service: Deactivated successfully. Oct 13 05:51:24.587128 systemd[1]: session-2.scope: Deactivated successfully. Oct 13 05:51:24.591006 systemd-logind[1980]: Session 2 logged out. Waiting for processes to exit. Oct 13 05:51:24.593670 systemd-logind[1980]: Removed session 2. Oct 13 05:51:24.610005 systemd[1]: Started sshd@2-172.31.26.230:22-139.178.68.195:45856.service - OpenSSH per-connection server daemon (139.178.68.195:45856). Oct 13 05:51:24.787626 sshd[2306]: Accepted publickey for core from 139.178.68.195 port 45856 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:51:24.789158 sshd-session[2306]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:51:24.794532 systemd-logind[1980]: New session 3 of user core. Oct 13 05:51:24.800978 systemd[1]: Started session-3.scope - Session 3 of User core. Oct 13 05:51:24.913236 kubelet[2283]: E1013 05:51:24.913179 2283 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 05:51:24.915359 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 05:51:24.915671 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 05:51:24.916014 systemd[1]: kubelet.service: Consumed 1.079s CPU time, 264.1M memory peak. Oct 13 05:51:24.923131 sshd[2310]: Connection closed by 139.178.68.195 port 45856 Oct 13 05:51:24.923962 sshd-session[2306]: pam_unix(sshd:session): session closed for user core Oct 13 05:51:24.927614 systemd[1]: sshd@2-172.31.26.230:22-139.178.68.195:45856.service: Deactivated successfully. Oct 13 05:51:24.929400 systemd[1]: session-3.scope: Deactivated successfully. Oct 13 05:51:24.932019 systemd-logind[1980]: Session 3 logged out. Waiting for processes to exit. Oct 13 05:51:24.933106 systemd-logind[1980]: Removed session 3. Oct 13 05:51:24.957707 systemd[1]: Started sshd@3-172.31.26.230:22-139.178.68.195:45870.service - OpenSSH per-connection server daemon (139.178.68.195:45870). Oct 13 05:51:25.135966 sshd[2317]: Accepted publickey for core from 139.178.68.195 port 45870 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:51:25.137325 sshd-session[2317]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:51:25.143972 systemd-logind[1980]: New session 4 of user core. Oct 13 05:51:25.152356 systemd[1]: Started session-4.scope - Session 4 of User core. Oct 13 05:51:25.275228 sshd[2320]: Connection closed by 139.178.68.195 port 45870 Oct 13 05:51:25.275966 sshd-session[2317]: pam_unix(sshd:session): session closed for user core Oct 13 05:51:25.280041 systemd[1]: sshd@3-172.31.26.230:22-139.178.68.195:45870.service: Deactivated successfully. Oct 13 05:51:25.281876 systemd[1]: session-4.scope: Deactivated successfully. Oct 13 05:51:25.282703 systemd-logind[1980]: Session 4 logged out. Waiting for processes to exit. Oct 13 05:51:25.284565 systemd-logind[1980]: Removed session 4. Oct 13 05:51:25.316402 systemd[1]: Started sshd@4-172.31.26.230:22-139.178.68.195:45878.service - OpenSSH per-connection server daemon (139.178.68.195:45878). Oct 13 05:51:25.497756 sshd[2326]: Accepted publickey for core from 139.178.68.195 port 45878 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:51:25.499119 sshd-session[2326]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:51:25.506635 systemd-logind[1980]: New session 5 of user core. Oct 13 05:51:25.511027 systemd[1]: Started session-5.scope - Session 5 of User core. Oct 13 05:51:25.624860 sudo[2330]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Oct 13 05:51:25.625148 sudo[2330]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 05:51:25.642382 sudo[2330]: pam_unix(sudo:session): session closed for user root Oct 13 05:51:25.666158 sshd[2329]: Connection closed by 139.178.68.195 port 45878 Oct 13 05:51:25.666878 sshd-session[2326]: pam_unix(sshd:session): session closed for user core Oct 13 05:51:25.671504 systemd[1]: sshd@4-172.31.26.230:22-139.178.68.195:45878.service: Deactivated successfully. Oct 13 05:51:25.673718 systemd[1]: session-5.scope: Deactivated successfully. Oct 13 05:51:25.674915 systemd-logind[1980]: Session 5 logged out. Waiting for processes to exit. Oct 13 05:51:25.676611 systemd-logind[1980]: Removed session 5. Oct 13 05:51:25.701713 systemd[1]: Started sshd@5-172.31.26.230:22-139.178.68.195:45886.service - OpenSSH per-connection server daemon (139.178.68.195:45886). Oct 13 05:51:25.881913 sshd[2336]: Accepted publickey for core from 139.178.68.195 port 45886 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:51:25.883941 sshd-session[2336]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:51:25.890880 systemd-logind[1980]: New session 6 of user core. Oct 13 05:51:25.897065 systemd[1]: Started session-6.scope - Session 6 of User core. Oct 13 05:51:26.001903 sudo[2341]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Oct 13 05:51:26.002279 sudo[2341]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 05:51:26.008480 sudo[2341]: pam_unix(sudo:session): session closed for user root Oct 13 05:51:26.014382 sudo[2340]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Oct 13 05:51:26.014737 sudo[2340]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 05:51:26.025655 systemd[1]: Starting audit-rules.service - Load Audit Rules... Oct 13 05:51:26.069881 augenrules[2363]: No rules Oct 13 05:51:26.071229 systemd[1]: audit-rules.service: Deactivated successfully. Oct 13 05:51:26.071582 systemd[1]: Finished audit-rules.service - Load Audit Rules. Oct 13 05:51:26.074060 sudo[2340]: pam_unix(sudo:session): session closed for user root Oct 13 05:51:26.096994 sshd[2339]: Connection closed by 139.178.68.195 port 45886 Oct 13 05:51:26.097507 sshd-session[2336]: pam_unix(sshd:session): session closed for user core Oct 13 05:51:26.102969 systemd[1]: sshd@5-172.31.26.230:22-139.178.68.195:45886.service: Deactivated successfully. Oct 13 05:51:26.105187 systemd[1]: session-6.scope: Deactivated successfully. Oct 13 05:51:26.106277 systemd-logind[1980]: Session 6 logged out. Waiting for processes to exit. Oct 13 05:51:26.108243 systemd-logind[1980]: Removed session 6. Oct 13 05:51:26.132243 systemd[1]: Started sshd@6-172.31.26.230:22-139.178.68.195:45888.service - OpenSSH per-connection server daemon (139.178.68.195:45888). Oct 13 05:51:26.306670 sshd[2372]: Accepted publickey for core from 139.178.68.195 port 45888 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:51:26.308155 sshd-session[2372]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:51:26.314461 systemd-logind[1980]: New session 7 of user core. Oct 13 05:51:26.324114 systemd[1]: Started session-7.scope - Session 7 of User core. Oct 13 05:51:26.419997 sudo[2376]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Oct 13 05:51:26.420369 sudo[2376]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Oct 13 05:51:26.847243 systemd[1]: Starting docker.service - Docker Application Container Engine... Oct 13 05:51:26.861280 (dockerd)[2395]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Oct 13 05:51:27.176841 dockerd[2395]: time="2025-10-13T05:51:27.176555778Z" level=info msg="Starting up" Oct 13 05:51:27.181506 dockerd[2395]: time="2025-10-13T05:51:27.181467201Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Oct 13 05:51:27.194016 dockerd[2395]: time="2025-10-13T05:51:27.193971214Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Oct 13 05:51:27.404327 dockerd[2395]: time="2025-10-13T05:51:27.404264914Z" level=info msg="Loading containers: start." Oct 13 05:51:27.418897 kernel: Initializing XFRM netlink socket Oct 13 05:51:27.660603 (udev-worker)[2416]: Network interface NamePolicy= disabled on kernel command line. Oct 13 05:51:27.707515 systemd-networkd[1896]: docker0: Link UP Oct 13 05:51:27.721988 dockerd[2395]: time="2025-10-13T05:51:27.721919490Z" level=info msg="Loading containers: done." Oct 13 05:51:27.741389 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2661169753-merged.mount: Deactivated successfully. Oct 13 05:51:27.743267 dockerd[2395]: time="2025-10-13T05:51:27.743222208Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Oct 13 05:51:27.743468 dockerd[2395]: time="2025-10-13T05:51:27.743321388Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Oct 13 05:51:27.743651 dockerd[2395]: time="2025-10-13T05:51:27.743622545Z" level=info msg="Initializing buildkit" Oct 13 05:51:27.776350 dockerd[2395]: time="2025-10-13T05:51:27.775059484Z" level=info msg="Completed buildkit initialization" Oct 13 05:51:27.789896 dockerd[2395]: time="2025-10-13T05:51:27.789826203Z" level=info msg="Daemon has completed initialization" Oct 13 05:51:27.790129 dockerd[2395]: time="2025-10-13T05:51:27.790035709Z" level=info msg="API listen on /run/docker.sock" Oct 13 05:51:27.790288 systemd[1]: Started docker.service - Docker Application Container Engine. Oct 13 05:51:28.864077 containerd[2012]: time="2025-10-13T05:51:28.864031670Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\"" Oct 13 05:51:29.743172 systemd-resolved[1897]: Clock change detected. Flushing caches. Oct 13 05:51:30.051931 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount820819140.mount: Deactivated successfully. Oct 13 05:51:31.445950 containerd[2012]: time="2025-10-13T05:51:31.445880075Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:31.447173 containerd[2012]: time="2025-10-13T05:51:31.447003087Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.9: active requests=0, bytes read=28837916" Oct 13 05:51:31.448425 containerd[2012]: time="2025-10-13T05:51:31.448389681Z" level=info msg="ImageCreate event name:\"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:31.450877 containerd[2012]: time="2025-10-13T05:51:31.450830151Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:31.452446 containerd[2012]: time="2025-10-13T05:51:31.452193194Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.9\" with image id \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6df11cc2ad9679b1117be34d3a0230add88bc0a08fd7a3ebc26b680575e8de97\", size \"28834515\" in 1.957448454s" Oct 13 05:51:31.452446 containerd[2012]: time="2025-10-13T05:51:31.452234577Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.9\" returns image reference \"sha256:abd2b525baf428ffb8b8b7d1e09761dc5cdb7ed0c7896a9427e29e84f8eafc59\"" Oct 13 05:51:31.454377 containerd[2012]: time="2025-10-13T05:51:31.454343950Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\"" Oct 13 05:51:32.958545 containerd[2012]: time="2025-10-13T05:51:32.958492551Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:32.959755 containerd[2012]: time="2025-10-13T05:51:32.959623138Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.9: active requests=0, bytes read=24787027" Oct 13 05:51:32.960756 containerd[2012]: time="2025-10-13T05:51:32.960722295Z" level=info msg="ImageCreate event name:\"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:32.963512 containerd[2012]: time="2025-10-13T05:51:32.963474668Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:32.964695 containerd[2012]: time="2025-10-13T05:51:32.964510096Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.9\" with image id \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:243c4b8e3bce271fcb1b78008ab996ab6976b1a20096deac08338fcd17979922\", size \"26421706\" in 1.510024149s" Oct 13 05:51:32.964695 containerd[2012]: time="2025-10-13T05:51:32.964549491Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.9\" returns image reference \"sha256:0debe32fbb7223500fcf8c312f2a568a5abd3ed9274d8ec6780cfb30b8861e91\"" Oct 13 05:51:32.965056 containerd[2012]: time="2025-10-13T05:51:32.965033293Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\"" Oct 13 05:51:34.379179 containerd[2012]: time="2025-10-13T05:51:34.378901060Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:34.386775 containerd[2012]: time="2025-10-13T05:51:34.386719568Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.9: active requests=0, bytes read=19176289" Oct 13 05:51:34.388916 containerd[2012]: time="2025-10-13T05:51:34.388824990Z" level=info msg="ImageCreate event name:\"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:34.394489 containerd[2012]: time="2025-10-13T05:51:34.394384627Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:34.395563 containerd[2012]: time="2025-10-13T05:51:34.395365493Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.9\" with image id \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:50c49520dbd0e8b4076b6a5c77d8014df09ea3d59a73e8bafd2678d51ebb92d5\", size \"20810986\" in 1.430295478s" Oct 13 05:51:34.395563 containerd[2012]: time="2025-10-13T05:51:34.395408736Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.9\" returns image reference \"sha256:6934c23b154fcb9bf54ed5913782de746735a49f4daa4732285915050cd44ad5\"" Oct 13 05:51:34.396262 containerd[2012]: time="2025-10-13T05:51:34.396100114Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\"" Oct 13 05:51:35.444206 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount543135020.mount: Deactivated successfully. Oct 13 05:51:35.782297 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Oct 13 05:51:35.785678 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 05:51:36.097195 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 05:51:36.108980 (kubelet)[2687]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Oct 13 05:51:36.133473 containerd[2012]: time="2025-10-13T05:51:36.133411338Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:36.135266 containerd[2012]: time="2025-10-13T05:51:36.135223613Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.9: active requests=0, bytes read=30924206" Oct 13 05:51:36.136262 containerd[2012]: time="2025-10-13T05:51:36.136227080Z" level=info msg="ImageCreate event name:\"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:36.139058 containerd[2012]: time="2025-10-13T05:51:36.139001443Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:36.139629 containerd[2012]: time="2025-10-13T05:51:36.139438571Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.9\" with image id \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\", repo tag \"registry.k8s.io/kube-proxy:v1.32.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:886af02535dc34886e4618b902f8c140d89af57233a245621d29642224516064\", size \"30923225\" in 1.742986032s" Oct 13 05:51:36.139629 containerd[2012]: time="2025-10-13T05:51:36.139488848Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.9\" returns image reference \"sha256:fa3fdca615a501743d8deb39729a96e731312aac8d96accec061d5265360332f\"" Oct 13 05:51:36.141768 containerd[2012]: time="2025-10-13T05:51:36.141740668Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Oct 13 05:51:36.168621 kubelet[2687]: E1013 05:51:36.168547 2687 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Oct 13 05:51:36.172816 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Oct 13 05:51:36.173003 systemd[1]: kubelet.service: Failed with result 'exit-code'. Oct 13 05:51:36.173368 systemd[1]: kubelet.service: Consumed 211ms CPU time, 110.9M memory peak. Oct 13 05:51:36.659174 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2839641642.mount: Deactivated successfully. Oct 13 05:51:38.006436 containerd[2012]: time="2025-10-13T05:51:38.006381444Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:38.007508 containerd[2012]: time="2025-10-13T05:51:38.007375767Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Oct 13 05:51:38.008503 containerd[2012]: time="2025-10-13T05:51:38.008462336Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:38.011229 containerd[2012]: time="2025-10-13T05:51:38.011174241Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:38.013086 containerd[2012]: time="2025-10-13T05:51:38.012920589Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.871139278s" Oct 13 05:51:38.013086 containerd[2012]: time="2025-10-13T05:51:38.012964476Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Oct 13 05:51:38.013550 containerd[2012]: time="2025-10-13T05:51:38.013527424Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Oct 13 05:51:38.447929 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3066096832.mount: Deactivated successfully. Oct 13 05:51:38.456417 containerd[2012]: time="2025-10-13T05:51:38.456346252Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 13 05:51:38.457280 containerd[2012]: time="2025-10-13T05:51:38.457137002Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Oct 13 05:51:38.458601 containerd[2012]: time="2025-10-13T05:51:38.458565531Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 13 05:51:38.464012 containerd[2012]: time="2025-10-13T05:51:38.463891671Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Oct 13 05:51:38.466320 containerd[2012]: time="2025-10-13T05:51:38.466262305Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 452.605202ms" Oct 13 05:51:38.466638 containerd[2012]: time="2025-10-13T05:51:38.466510980Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Oct 13 05:51:38.468278 containerd[2012]: time="2025-10-13T05:51:38.468236439Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Oct 13 05:51:38.987621 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1977914020.mount: Deactivated successfully. Oct 13 05:51:41.004854 containerd[2012]: time="2025-10-13T05:51:41.004782000Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:41.005758 containerd[2012]: time="2025-10-13T05:51:41.005716129Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57682056" Oct 13 05:51:41.007250 containerd[2012]: time="2025-10-13T05:51:41.007193391Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:41.010757 containerd[2012]: time="2025-10-13T05:51:41.010691637Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:51:41.012130 containerd[2012]: time="2025-10-13T05:51:41.011899681Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.543627685s" Oct 13 05:51:41.012130 containerd[2012]: time="2025-10-13T05:51:41.011947773Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Oct 13 05:51:43.484265 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 05:51:43.485003 systemd[1]: kubelet.service: Consumed 211ms CPU time, 110.9M memory peak. Oct 13 05:51:43.487766 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 05:51:43.528661 systemd[1]: Reload requested from client PID 2831 ('systemctl') (unit session-7.scope)... Oct 13 05:51:43.528807 systemd[1]: Reloading... Oct 13 05:51:43.681487 zram_generator::config[2879]: No configuration found. Oct 13 05:51:43.964539 systemd[1]: Reloading finished in 435 ms. Oct 13 05:51:44.031502 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Oct 13 05:51:44.031630 systemd[1]: kubelet.service: Failed with result 'signal'. Oct 13 05:51:44.032005 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 05:51:44.032081 systemd[1]: kubelet.service: Consumed 146ms CPU time, 98.3M memory peak. Oct 13 05:51:44.034259 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 05:51:44.306359 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 05:51:44.318548 (kubelet)[2939]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 13 05:51:44.377511 kubelet[2939]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 05:51:44.377511 kubelet[2939]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 13 05:51:44.377511 kubelet[2939]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 05:51:44.377511 kubelet[2939]: I1013 05:51:44.376336 2939 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 13 05:51:44.620558 kubelet[2939]: I1013 05:51:44.619379 2939 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 13 05:51:44.620558 kubelet[2939]: I1013 05:51:44.619790 2939 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 13 05:51:44.620558 kubelet[2939]: I1013 05:51:44.620240 2939 server.go:954] "Client rotation is on, will bootstrap in background" Oct 13 05:51:44.670317 kubelet[2939]: E1013 05:51:44.670213 2939 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.26.230:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:44.673638 kubelet[2939]: I1013 05:51:44.673590 2939 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 13 05:51:44.689604 kubelet[2939]: I1013 05:51:44.689579 2939 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 13 05:51:44.697217 kubelet[2939]: I1013 05:51:44.697176 2939 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 13 05:51:44.699700 kubelet[2939]: I1013 05:51:44.699632 2939 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 13 05:51:44.699929 kubelet[2939]: I1013 05:51:44.699691 2939 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-26-230","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 13 05:51:44.702159 kubelet[2939]: I1013 05:51:44.701944 2939 topology_manager.go:138] "Creating topology manager with none policy" Oct 13 05:51:44.702159 kubelet[2939]: I1013 05:51:44.702152 2939 container_manager_linux.go:304] "Creating device plugin manager" Oct 13 05:51:44.703652 kubelet[2939]: I1013 05:51:44.703623 2939 state_mem.go:36] "Initialized new in-memory state store" Oct 13 05:51:44.715540 kubelet[2939]: I1013 05:51:44.715407 2939 kubelet.go:446] "Attempting to sync node with API server" Oct 13 05:51:44.715540 kubelet[2939]: I1013 05:51:44.715464 2939 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 13 05:51:44.717358 kubelet[2939]: I1013 05:51:44.717321 2939 kubelet.go:352] "Adding apiserver pod source" Oct 13 05:51:44.717358 kubelet[2939]: I1013 05:51:44.717351 2939 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 13 05:51:44.720843 kubelet[2939]: W1013 05:51:44.720783 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.26.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-26-230&limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:44.720959 kubelet[2939]: E1013 05:51:44.720851 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.26.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-26-230&limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:44.721236 kubelet[2939]: W1013 05:51:44.721195 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.26.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:44.721279 kubelet[2939]: E1013 05:51:44.721234 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.26.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:44.723269 kubelet[2939]: I1013 05:51:44.723147 2939 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 13 05:51:44.728138 kubelet[2939]: I1013 05:51:44.728110 2939 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 13 05:51:44.729233 kubelet[2939]: W1013 05:51:44.728302 2939 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Oct 13 05:51:44.734239 kubelet[2939]: I1013 05:51:44.733906 2939 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 13 05:51:44.734239 kubelet[2939]: I1013 05:51:44.734057 2939 server.go:1287] "Started kubelet" Oct 13 05:51:44.735902 kubelet[2939]: I1013 05:51:44.735616 2939 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 13 05:51:44.746305 kubelet[2939]: I1013 05:51:44.746206 2939 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 13 05:51:44.746803 kubelet[2939]: I1013 05:51:44.746774 2939 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 13 05:51:44.747812 kubelet[2939]: I1013 05:51:44.747790 2939 server.go:479] "Adding debug handlers to kubelet server" Oct 13 05:51:44.758679 kubelet[2939]: E1013 05:51:44.753255 2939 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.26.230:6443/api/v1/namespaces/default/events\": dial tcp 172.31.26.230:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-26-230.186df71659753e99 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-26-230,UID:ip-172-31-26-230,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-26-230,},FirstTimestamp:2025-10-13 05:51:44.733929113 +0000 UTC m=+0.410770979,LastTimestamp:2025-10-13 05:51:44.733929113 +0000 UTC m=+0.410770979,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-26-230,}" Oct 13 05:51:44.760995 kubelet[2939]: I1013 05:51:44.760863 2939 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 13 05:51:44.767082 kubelet[2939]: I1013 05:51:44.761021 2939 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 13 05:51:44.767227 kubelet[2939]: I1013 05:51:44.767210 2939 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 13 05:51:44.771478 kubelet[2939]: E1013 05:51:44.770931 2939 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-26-230\" not found" Oct 13 05:51:44.771996 kubelet[2939]: I1013 05:51:44.771967 2939 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 13 05:51:44.772091 kubelet[2939]: I1013 05:51:44.772066 2939 reconciler.go:26] "Reconciler: start to sync state" Oct 13 05:51:44.772704 kubelet[2939]: E1013 05:51:44.772658 2939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.26.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-26-230?timeout=10s\": dial tcp 172.31.26.230:6443: connect: connection refused" interval="200ms" Oct 13 05:51:44.777321 kubelet[2939]: W1013 05:51:44.772784 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.26.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:44.777441 kubelet[2939]: E1013 05:51:44.777354 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.26.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:44.782657 kubelet[2939]: I1013 05:51:44.782627 2939 factory.go:221] Registration of the systemd container factory successfully Oct 13 05:51:44.782894 kubelet[2939]: I1013 05:51:44.782875 2939 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 13 05:51:44.789483 kubelet[2939]: I1013 05:51:44.788891 2939 factory.go:221] Registration of the containerd container factory successfully Oct 13 05:51:44.794682 kubelet[2939]: I1013 05:51:44.794505 2939 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 13 05:51:44.796592 kubelet[2939]: I1013 05:51:44.796560 2939 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 13 05:51:44.796744 kubelet[2939]: I1013 05:51:44.796734 2939 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 13 05:51:44.796845 kubelet[2939]: I1013 05:51:44.796835 2939 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 13 05:51:44.796913 kubelet[2939]: I1013 05:51:44.796905 2939 kubelet.go:2382] "Starting kubelet main sync loop" Oct 13 05:51:44.797046 kubelet[2939]: E1013 05:51:44.797014 2939 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 13 05:51:44.806989 kubelet[2939]: W1013 05:51:44.806918 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.26.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:44.807251 kubelet[2939]: E1013 05:51:44.807230 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.26.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:44.812876 kubelet[2939]: E1013 05:51:44.812839 2939 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 13 05:51:44.821847 kubelet[2939]: I1013 05:51:44.821817 2939 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 13 05:51:44.821847 kubelet[2939]: I1013 05:51:44.821836 2939 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 13 05:51:44.822209 kubelet[2939]: I1013 05:51:44.821858 2939 state_mem.go:36] "Initialized new in-memory state store" Oct 13 05:51:44.824427 kubelet[2939]: I1013 05:51:44.824389 2939 policy_none.go:49] "None policy: Start" Oct 13 05:51:44.824427 kubelet[2939]: I1013 05:51:44.824414 2939 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 13 05:51:44.824427 kubelet[2939]: I1013 05:51:44.824426 2939 state_mem.go:35] "Initializing new in-memory state store" Oct 13 05:51:44.832133 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Oct 13 05:51:44.848718 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Oct 13 05:51:44.854236 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Oct 13 05:51:44.866957 kubelet[2939]: I1013 05:51:44.866928 2939 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 13 05:51:44.867351 kubelet[2939]: I1013 05:51:44.867334 2939 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 13 05:51:44.867589 kubelet[2939]: I1013 05:51:44.867548 2939 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 13 05:51:44.870926 kubelet[2939]: I1013 05:51:44.870823 2939 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 13 05:51:44.873372 kubelet[2939]: E1013 05:51:44.873347 2939 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 13 05:51:44.873592 kubelet[2939]: E1013 05:51:44.873398 2939 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-26-230\" not found" Oct 13 05:51:44.910814 systemd[1]: Created slice kubepods-burstable-podd362fb2966d13a857993160b074e18a4.slice - libcontainer container kubepods-burstable-podd362fb2966d13a857993160b074e18a4.slice. Oct 13 05:51:44.919499 kubelet[2939]: E1013 05:51:44.919417 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:44.923001 systemd[1]: Created slice kubepods-burstable-podb02086a7cb8c4cb1528bfab6382535e4.slice - libcontainer container kubepods-burstable-podb02086a7cb8c4cb1528bfab6382535e4.slice. Oct 13 05:51:44.926266 kubelet[2939]: E1013 05:51:44.926237 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:44.928193 systemd[1]: Created slice kubepods-burstable-pod22202722d7a76ed9269470f4a95719e9.slice - libcontainer container kubepods-burstable-pod22202722d7a76ed9269470f4a95719e9.slice. Oct 13 05:51:44.930885 kubelet[2939]: E1013 05:51:44.930856 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:44.969782 kubelet[2939]: I1013 05:51:44.969744 2939 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-26-230" Oct 13 05:51:44.970360 kubelet[2939]: E1013 05:51:44.970326 2939 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.26.230:6443/api/v1/nodes\": dial tcp 172.31.26.230:6443: connect: connection refused" node="ip-172-31-26-230" Oct 13 05:51:44.973440 kubelet[2939]: E1013 05:51:44.973383 2939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.26.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-26-230?timeout=10s\": dial tcp 172.31.26.230:6443: connect: connection refused" interval="400ms" Oct 13 05:51:44.973621 kubelet[2939]: I1013 05:51:44.973517 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-ca-certs\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:44.973621 kubelet[2939]: I1013 05:51:44.973547 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-k8s-certs\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:44.973621 kubelet[2939]: I1013 05:51:44.973572 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d362fb2966d13a857993160b074e18a4-k8s-certs\") pod \"kube-apiserver-ip-172-31-26-230\" (UID: \"d362fb2966d13a857993160b074e18a4\") " pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:44.973621 kubelet[2939]: I1013 05:51:44.973596 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d362fb2966d13a857993160b074e18a4-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-26-230\" (UID: \"d362fb2966d13a857993160b074e18a4\") " pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:44.973621 kubelet[2939]: I1013 05:51:44.973618 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-kubeconfig\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:44.973892 kubelet[2939]: I1013 05:51:44.973646 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:44.973892 kubelet[2939]: I1013 05:51:44.973669 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b02086a7cb8c4cb1528bfab6382535e4-kubeconfig\") pod \"kube-scheduler-ip-172-31-26-230\" (UID: \"b02086a7cb8c4cb1528bfab6382535e4\") " pod="kube-system/kube-scheduler-ip-172-31-26-230" Oct 13 05:51:44.973892 kubelet[2939]: I1013 05:51:44.973694 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d362fb2966d13a857993160b074e18a4-ca-certs\") pod \"kube-apiserver-ip-172-31-26-230\" (UID: \"d362fb2966d13a857993160b074e18a4\") " pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:44.973892 kubelet[2939]: I1013 05:51:44.973718 2939 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:45.173361 kubelet[2939]: I1013 05:51:45.173247 2939 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-26-230" Oct 13 05:51:45.174884 kubelet[2939]: E1013 05:51:45.174837 2939 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.26.230:6443/api/v1/nodes\": dial tcp 172.31.26.230:6443: connect: connection refused" node="ip-172-31-26-230" Oct 13 05:51:45.221083 containerd[2012]: time="2025-10-13T05:51:45.221044533Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-26-230,Uid:d362fb2966d13a857993160b074e18a4,Namespace:kube-system,Attempt:0,}" Oct 13 05:51:45.228209 containerd[2012]: time="2025-10-13T05:51:45.228163906Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-26-230,Uid:b02086a7cb8c4cb1528bfab6382535e4,Namespace:kube-system,Attempt:0,}" Oct 13 05:51:45.235043 containerd[2012]: time="2025-10-13T05:51:45.234977112Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-26-230,Uid:22202722d7a76ed9269470f4a95719e9,Namespace:kube-system,Attempt:0,}" Oct 13 05:51:45.376226 kubelet[2939]: E1013 05:51:45.375732 2939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.26.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-26-230?timeout=10s\": dial tcp 172.31.26.230:6443: connect: connection refused" interval="800ms" Oct 13 05:51:45.379776 containerd[2012]: time="2025-10-13T05:51:45.379405618Z" level=info msg="connecting to shim 18e12f67dcb218e04f0fe4541526b6f2728daad070e595b63724a63179809417" address="unix:///run/containerd/s/bb38417ee46fea3d9e9048d218a4df147a9b02a6919c9e3abc24f463b1340106" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:51:45.387506 containerd[2012]: time="2025-10-13T05:51:45.387061827Z" level=info msg="connecting to shim 4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff" address="unix:///run/containerd/s/07b7bad936943a8db40cd17d2d6ebe98b9d13be2944471daffd130d1d6635b8c" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:51:45.390548 containerd[2012]: time="2025-10-13T05:51:45.390490500Z" level=info msg="connecting to shim 9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937" address="unix:///run/containerd/s/03b2d2f144bfd4fc661ba908dae25b5f7804d7746f11078cfcd2afd401f76d72" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:51:45.525730 systemd[1]: Started cri-containerd-18e12f67dcb218e04f0fe4541526b6f2728daad070e595b63724a63179809417.scope - libcontainer container 18e12f67dcb218e04f0fe4541526b6f2728daad070e595b63724a63179809417. Oct 13 05:51:45.528511 systemd[1]: Started cri-containerd-4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff.scope - libcontainer container 4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff. Oct 13 05:51:45.541113 systemd[1]: Started cri-containerd-9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937.scope - libcontainer container 9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937. Oct 13 05:51:45.580510 kubelet[2939]: I1013 05:51:45.580086 2939 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-26-230" Oct 13 05:51:45.581963 kubelet[2939]: E1013 05:51:45.581414 2939 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.26.230:6443/api/v1/nodes\": dial tcp 172.31.26.230:6443: connect: connection refused" node="ip-172-31-26-230" Oct 13 05:51:45.659804 containerd[2012]: time="2025-10-13T05:51:45.659272747Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-26-230,Uid:22202722d7a76ed9269470f4a95719e9,Namespace:kube-system,Attempt:0,} returns sandbox id \"4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff\"" Oct 13 05:51:45.666551 containerd[2012]: time="2025-10-13T05:51:45.666061588Z" level=info msg="CreateContainer within sandbox \"4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Oct 13 05:51:45.681291 containerd[2012]: time="2025-10-13T05:51:45.681242578Z" level=info msg="Container 2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:51:45.688759 containerd[2012]: time="2025-10-13T05:51:45.688639930Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-26-230,Uid:d362fb2966d13a857993160b074e18a4,Namespace:kube-system,Attempt:0,} returns sandbox id \"18e12f67dcb218e04f0fe4541526b6f2728daad070e595b63724a63179809417\"" Oct 13 05:51:45.694670 containerd[2012]: time="2025-10-13T05:51:45.694623200Z" level=info msg="CreateContainer within sandbox \"18e12f67dcb218e04f0fe4541526b6f2728daad070e595b63724a63179809417\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Oct 13 05:51:45.700520 containerd[2012]: time="2025-10-13T05:51:45.694779951Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-26-230,Uid:b02086a7cb8c4cb1528bfab6382535e4,Namespace:kube-system,Attempt:0,} returns sandbox id \"9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937\"" Oct 13 05:51:45.703128 containerd[2012]: time="2025-10-13T05:51:45.702632882Z" level=info msg="CreateContainer within sandbox \"9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Oct 13 05:51:45.710883 containerd[2012]: time="2025-10-13T05:51:45.710824953Z" level=info msg="CreateContainer within sandbox \"4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd\"" Oct 13 05:51:45.711711 containerd[2012]: time="2025-10-13T05:51:45.711636630Z" level=info msg="StartContainer for \"2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd\"" Oct 13 05:51:45.712921 containerd[2012]: time="2025-10-13T05:51:45.712883309Z" level=info msg="connecting to shim 2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd" address="unix:///run/containerd/s/07b7bad936943a8db40cd17d2d6ebe98b9d13be2944471daffd130d1d6635b8c" protocol=ttrpc version=3 Oct 13 05:51:45.725300 containerd[2012]: time="2025-10-13T05:51:45.725240406Z" level=info msg="Container 5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:51:45.728819 containerd[2012]: time="2025-10-13T05:51:45.728774542Z" level=info msg="Container 775151bcadc84081018e86e83dfd93915fa2816c41ac9c0b1b4f3c8f4dc0886c: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:51:45.742971 systemd[1]: Started cri-containerd-2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd.scope - libcontainer container 2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd. Oct 13 05:51:45.750559 containerd[2012]: time="2025-10-13T05:51:45.750493744Z" level=info msg="CreateContainer within sandbox \"9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d\"" Oct 13 05:51:45.751938 containerd[2012]: time="2025-10-13T05:51:45.751904166Z" level=info msg="StartContainer for \"5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d\"" Oct 13 05:51:45.755678 containerd[2012]: time="2025-10-13T05:51:45.755635122Z" level=info msg="connecting to shim 5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d" address="unix:///run/containerd/s/03b2d2f144bfd4fc661ba908dae25b5f7804d7746f11078cfcd2afd401f76d72" protocol=ttrpc version=3 Oct 13 05:51:45.758873 containerd[2012]: time="2025-10-13T05:51:45.758830002Z" level=info msg="CreateContainer within sandbox \"18e12f67dcb218e04f0fe4541526b6f2728daad070e595b63724a63179809417\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"775151bcadc84081018e86e83dfd93915fa2816c41ac9c0b1b4f3c8f4dc0886c\"" Oct 13 05:51:45.760292 containerd[2012]: time="2025-10-13T05:51:45.760238057Z" level=info msg="StartContainer for \"775151bcadc84081018e86e83dfd93915fa2816c41ac9c0b1b4f3c8f4dc0886c\"" Oct 13 05:51:45.764614 containerd[2012]: time="2025-10-13T05:51:45.764566386Z" level=info msg="connecting to shim 775151bcadc84081018e86e83dfd93915fa2816c41ac9c0b1b4f3c8f4dc0886c" address="unix:///run/containerd/s/bb38417ee46fea3d9e9048d218a4df147a9b02a6919c9e3abc24f463b1340106" protocol=ttrpc version=3 Oct 13 05:51:45.792634 kubelet[2939]: W1013 05:51:45.792490 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.26.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:45.792634 kubelet[2939]: E1013 05:51:45.792590 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.26.230:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:45.797737 systemd[1]: Started cri-containerd-775151bcadc84081018e86e83dfd93915fa2816c41ac9c0b1b4f3c8f4dc0886c.scope - libcontainer container 775151bcadc84081018e86e83dfd93915fa2816c41ac9c0b1b4f3c8f4dc0886c. Oct 13 05:51:45.802177 kubelet[2939]: W1013 05:51:45.802113 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.26.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:45.802299 kubelet[2939]: E1013 05:51:45.802194 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.26.230:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:45.812828 systemd[1]: Started cri-containerd-5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d.scope - libcontainer container 5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d. Oct 13 05:51:45.882881 containerd[2012]: time="2025-10-13T05:51:45.882827848Z" level=info msg="StartContainer for \"2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd\" returns successfully" Oct 13 05:51:45.905578 containerd[2012]: time="2025-10-13T05:51:45.905442672Z" level=info msg="StartContainer for \"5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d\" returns successfully" Oct 13 05:51:45.939752 containerd[2012]: time="2025-10-13T05:51:45.939543318Z" level=info msg="StartContainer for \"775151bcadc84081018e86e83dfd93915fa2816c41ac9c0b1b4f3c8f4dc0886c\" returns successfully" Oct 13 05:51:46.015877 kubelet[2939]: W1013 05:51:46.015767 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.26.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:46.015877 kubelet[2939]: E1013 05:51:46.015847 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.26.230:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:46.145358 kubelet[2939]: W1013 05:51:46.145151 2939 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.26.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-26-230&limit=500&resourceVersion=0": dial tcp 172.31.26.230:6443: connect: connection refused Oct 13 05:51:46.145358 kubelet[2939]: E1013 05:51:46.145236 2939 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.26.230:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-26-230&limit=500&resourceVersion=0\": dial tcp 172.31.26.230:6443: connect: connection refused" logger="UnhandledError" Oct 13 05:51:46.179207 kubelet[2939]: E1013 05:51:46.179150 2939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.26.230:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-26-230?timeout=10s\": dial tcp 172.31.26.230:6443: connect: connection refused" interval="1.6s" Oct 13 05:51:46.387175 kubelet[2939]: I1013 05:51:46.386687 2939 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-26-230" Oct 13 05:51:46.387175 kubelet[2939]: E1013 05:51:46.387068 2939 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.26.230:6443/api/v1/nodes\": dial tcp 172.31.26.230:6443: connect: connection refused" node="ip-172-31-26-230" Oct 13 05:51:46.863358 kubelet[2939]: E1013 05:51:46.863326 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:46.869030 kubelet[2939]: E1013 05:51:46.868994 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:46.872220 kubelet[2939]: E1013 05:51:46.872190 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:47.875404 kubelet[2939]: E1013 05:51:47.875342 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:47.878101 kubelet[2939]: E1013 05:51:47.877928 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:47.878101 kubelet[2939]: E1013 05:51:47.878032 2939 kubelet.go:3190] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:47.991475 kubelet[2939]: I1013 05:51:47.990015 2939 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-26-230" Oct 13 05:51:48.650476 kubelet[2939]: E1013 05:51:48.650256 2939 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-26-230\" not found" node="ip-172-31-26-230" Oct 13 05:51:48.727107 kubelet[2939]: I1013 05:51:48.727064 2939 apiserver.go:52] "Watching apiserver" Oct 13 05:51:48.752684 kubelet[2939]: I1013 05:51:48.752560 2939 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-26-230" Oct 13 05:51:48.753141 kubelet[2939]: E1013 05:51:48.752810 2939 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ip-172-31-26-230\": node \"ip-172-31-26-230\" not found" Oct 13 05:51:48.772260 kubelet[2939]: I1013 05:51:48.771835 2939 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:48.772260 kubelet[2939]: I1013 05:51:48.772127 2939 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 13 05:51:48.792662 kubelet[2939]: E1013 05:51:48.792626 2939 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-26-230\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:48.792854 kubelet[2939]: I1013 05:51:48.792844 2939 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:48.795993 kubelet[2939]: E1013 05:51:48.795809 2939 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-26-230\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:48.795993 kubelet[2939]: I1013 05:51:48.795844 2939 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-26-230" Oct 13 05:51:48.799232 kubelet[2939]: E1013 05:51:48.799192 2939 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-26-230\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-26-230" Oct 13 05:51:48.876506 kubelet[2939]: I1013 05:51:48.875898 2939 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:48.876506 kubelet[2939]: I1013 05:51:48.876261 2939 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-26-230" Oct 13 05:51:48.881471 kubelet[2939]: E1013 05:51:48.881088 2939 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-26-230\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:48.881846 kubelet[2939]: E1013 05:51:48.881822 2939 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-26-230\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-26-230" Oct 13 05:51:49.877823 kubelet[2939]: I1013 05:51:49.877774 2939 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:51.292131 systemd[1]: Reload requested from client PID 3209 ('systemctl') (unit session-7.scope)... Oct 13 05:51:51.292153 systemd[1]: Reloading... Oct 13 05:51:51.431531 zram_generator::config[3252]: No configuration found. Oct 13 05:51:51.728376 systemd[1]: Reloading finished in 435 ms. Oct 13 05:51:51.758965 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 05:51:51.778271 systemd[1]: kubelet.service: Deactivated successfully. Oct 13 05:51:51.778517 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 05:51:51.778578 systemd[1]: kubelet.service: Consumed 860ms CPU time, 130.2M memory peak. Oct 13 05:51:51.782435 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Oct 13 05:51:52.102175 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Oct 13 05:51:52.115018 (kubelet)[3314]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Oct 13 05:51:52.197504 kubelet[3314]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 05:51:52.197919 kubelet[3314]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Oct 13 05:51:52.197998 kubelet[3314]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 05:51:52.198205 kubelet[3314]: I1013 05:51:52.198162 3314 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 13 05:51:52.207870 kubelet[3314]: I1013 05:51:52.207829 3314 server.go:520] "Kubelet version" kubeletVersion="v1.32.4" Oct 13 05:51:52.207870 kubelet[3314]: I1013 05:51:52.207866 3314 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 13 05:51:52.208576 kubelet[3314]: I1013 05:51:52.208553 3314 server.go:954] "Client rotation is on, will bootstrap in background" Oct 13 05:51:52.212925 kubelet[3314]: I1013 05:51:52.212903 3314 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 13 05:51:52.216761 kubelet[3314]: I1013 05:51:52.216736 3314 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Oct 13 05:51:52.222517 kubelet[3314]: I1013 05:51:52.221663 3314 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 13 05:51:52.224743 kubelet[3314]: I1013 05:51:52.224709 3314 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Oct 13 05:51:52.224998 kubelet[3314]: I1013 05:51:52.224961 3314 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 13 05:51:52.225202 kubelet[3314]: I1013 05:51:52.224995 3314 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-26-230","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 13 05:51:52.225341 kubelet[3314]: I1013 05:51:52.225212 3314 topology_manager.go:138] "Creating topology manager with none policy" Oct 13 05:51:52.225341 kubelet[3314]: I1013 05:51:52.225225 3314 container_manager_linux.go:304] "Creating device plugin manager" Oct 13 05:51:52.225341 kubelet[3314]: I1013 05:51:52.225284 3314 state_mem.go:36] "Initialized new in-memory state store" Oct 13 05:51:52.225618 kubelet[3314]: I1013 05:51:52.225506 3314 kubelet.go:446] "Attempting to sync node with API server" Oct 13 05:51:52.225618 kubelet[3314]: I1013 05:51:52.225532 3314 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 13 05:51:52.225618 kubelet[3314]: I1013 05:51:52.225560 3314 kubelet.go:352] "Adding apiserver pod source" Oct 13 05:51:52.225618 kubelet[3314]: I1013 05:51:52.225574 3314 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 13 05:51:52.231201 kubelet[3314]: I1013 05:51:52.231160 3314 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.5" apiVersion="v1" Oct 13 05:51:52.232427 kubelet[3314]: I1013 05:51:52.232365 3314 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 13 05:51:52.234150 kubelet[3314]: I1013 05:51:52.233483 3314 watchdog_linux.go:99] "Systemd watchdog is not enabled" Oct 13 05:51:52.234150 kubelet[3314]: I1013 05:51:52.233527 3314 server.go:1287] "Started kubelet" Oct 13 05:51:52.246473 kubelet[3314]: I1013 05:51:52.246356 3314 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Oct 13 05:51:52.250948 kubelet[3314]: I1013 05:51:52.250899 3314 server.go:479] "Adding debug handlers to kubelet server" Oct 13 05:51:52.257869 kubelet[3314]: I1013 05:51:52.257785 3314 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 13 05:51:52.258116 kubelet[3314]: I1013 05:51:52.258087 3314 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 13 05:51:52.261039 kubelet[3314]: I1013 05:51:52.260721 3314 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 13 05:51:52.269037 kubelet[3314]: I1013 05:51:52.268997 3314 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Oct 13 05:51:52.271714 kubelet[3314]: I1013 05:51:52.271691 3314 volume_manager.go:297] "Starting Kubelet Volume Manager" Oct 13 05:51:52.272096 kubelet[3314]: E1013 05:51:52.272075 3314 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-26-230\" not found" Oct 13 05:51:52.272867 kubelet[3314]: I1013 05:51:52.272837 3314 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Oct 13 05:51:52.273151 kubelet[3314]: I1013 05:51:52.273140 3314 reconciler.go:26] "Reconciler: start to sync state" Oct 13 05:51:52.283100 kubelet[3314]: I1013 05:51:52.282935 3314 factory.go:221] Registration of the systemd container factory successfully Oct 13 05:51:52.283940 kubelet[3314]: I1013 05:51:52.283801 3314 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Oct 13 05:51:52.295701 kubelet[3314]: E1013 05:51:52.295588 3314 kubelet.go:1555] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Oct 13 05:51:52.298644 kubelet[3314]: I1013 05:51:52.298619 3314 factory.go:221] Registration of the containerd container factory successfully Oct 13 05:51:52.307766 kubelet[3314]: I1013 05:51:52.307566 3314 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 13 05:51:52.312336 kubelet[3314]: I1013 05:51:52.312303 3314 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 13 05:51:52.312558 kubelet[3314]: I1013 05:51:52.312539 3314 status_manager.go:227] "Starting to sync pod status with apiserver" Oct 13 05:51:52.312636 kubelet[3314]: I1013 05:51:52.312571 3314 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Oct 13 05:51:52.312636 kubelet[3314]: I1013 05:51:52.312581 3314 kubelet.go:2382] "Starting kubelet main sync loop" Oct 13 05:51:52.312711 kubelet[3314]: E1013 05:51:52.312642 3314 kubelet.go:2406] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 13 05:51:52.392139 kubelet[3314]: I1013 05:51:52.392025 3314 cpu_manager.go:221] "Starting CPU manager" policy="none" Oct 13 05:51:52.392139 kubelet[3314]: I1013 05:51:52.392044 3314 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Oct 13 05:51:52.392139 kubelet[3314]: I1013 05:51:52.392070 3314 state_mem.go:36] "Initialized new in-memory state store" Oct 13 05:51:52.394475 kubelet[3314]: I1013 05:51:52.393753 3314 state_mem.go:88] "Updated default CPUSet" cpuSet="" Oct 13 05:51:52.394475 kubelet[3314]: I1013 05:51:52.393778 3314 state_mem.go:96] "Updated CPUSet assignments" assignments={} Oct 13 05:51:52.394475 kubelet[3314]: I1013 05:51:52.393804 3314 policy_none.go:49] "None policy: Start" Oct 13 05:51:52.394475 kubelet[3314]: I1013 05:51:52.393820 3314 memory_manager.go:186] "Starting memorymanager" policy="None" Oct 13 05:51:52.394475 kubelet[3314]: I1013 05:51:52.393835 3314 state_mem.go:35] "Initializing new in-memory state store" Oct 13 05:51:52.394475 kubelet[3314]: I1013 05:51:52.394121 3314 state_mem.go:75] "Updated machine memory state" Oct 13 05:51:52.404982 kubelet[3314]: I1013 05:51:52.404956 3314 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 13 05:51:52.405539 kubelet[3314]: I1013 05:51:52.405515 3314 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 13 05:51:52.405647 kubelet[3314]: I1013 05:51:52.405540 3314 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 13 05:51:52.407276 kubelet[3314]: I1013 05:51:52.406162 3314 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 13 05:51:52.412079 kubelet[3314]: E1013 05:51:52.412053 3314 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Oct 13 05:51:52.420428 kubelet[3314]: I1013 05:51:52.420402 3314 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-26-230" Oct 13 05:51:52.420919 kubelet[3314]: I1013 05:51:52.420903 3314 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:52.422386 kubelet[3314]: I1013 05:51:52.422354 3314 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:52.443639 kubelet[3314]: E1013 05:51:52.443594 3314 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-26-230\" already exists" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:52.475434 kubelet[3314]: I1013 05:51:52.475391 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-ca-certs\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:52.475700 kubelet[3314]: I1013 05:51:52.475492 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-k8s-certs\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:52.475700 kubelet[3314]: I1013 05:51:52.475516 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:52.475700 kubelet[3314]: I1013 05:51:52.475550 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d362fb2966d13a857993160b074e18a4-ca-certs\") pod \"kube-apiserver-ip-172-31-26-230\" (UID: \"d362fb2966d13a857993160b074e18a4\") " pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:52.475700 kubelet[3314]: I1013 05:51:52.475578 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:52.475700 kubelet[3314]: I1013 05:51:52.475602 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/22202722d7a76ed9269470f4a95719e9-kubeconfig\") pod \"kube-controller-manager-ip-172-31-26-230\" (UID: \"22202722d7a76ed9269470f4a95719e9\") " pod="kube-system/kube-controller-manager-ip-172-31-26-230" Oct 13 05:51:52.476046 kubelet[3314]: I1013 05:51:52.475623 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b02086a7cb8c4cb1528bfab6382535e4-kubeconfig\") pod \"kube-scheduler-ip-172-31-26-230\" (UID: \"b02086a7cb8c4cb1528bfab6382535e4\") " pod="kube-system/kube-scheduler-ip-172-31-26-230" Oct 13 05:51:52.476046 kubelet[3314]: I1013 05:51:52.475648 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d362fb2966d13a857993160b074e18a4-k8s-certs\") pod \"kube-apiserver-ip-172-31-26-230\" (UID: \"d362fb2966d13a857993160b074e18a4\") " pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:52.476046 kubelet[3314]: I1013 05:51:52.475671 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d362fb2966d13a857993160b074e18a4-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-26-230\" (UID: \"d362fb2966d13a857993160b074e18a4\") " pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:52.526366 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Oct 13 05:51:52.535271 kubelet[3314]: I1013 05:51:52.534798 3314 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-26-230" Oct 13 05:51:52.547209 kubelet[3314]: I1013 05:51:52.547141 3314 kubelet_node_status.go:124] "Node was previously registered" node="ip-172-31-26-230" Oct 13 05:51:52.547640 kubelet[3314]: I1013 05:51:52.547559 3314 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-26-230" Oct 13 05:51:53.237847 kubelet[3314]: I1013 05:51:53.237615 3314 apiserver.go:52] "Watching apiserver" Oct 13 05:51:53.273851 kubelet[3314]: I1013 05:51:53.273750 3314 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Oct 13 05:51:53.355737 kubelet[3314]: I1013 05:51:53.355694 3314 kubelet.go:3194] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:53.367475 kubelet[3314]: E1013 05:51:53.366310 3314 kubelet.go:3196] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-26-230\" already exists" pod="kube-system/kube-apiserver-ip-172-31-26-230" Oct 13 05:51:53.390873 kubelet[3314]: I1013 05:51:53.390801 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-26-230" podStartSLOduration=1.390784575 podStartE2EDuration="1.390784575s" podCreationTimestamp="2025-10-13 05:51:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 05:51:53.390422616 +0000 UTC m=+1.267136812" watchObservedRunningTime="2025-10-13 05:51:53.390784575 +0000 UTC m=+1.267498766" Oct 13 05:51:53.406002 kubelet[3314]: I1013 05:51:53.405941 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-26-230" podStartSLOduration=1.405909071 podStartE2EDuration="1.405909071s" podCreationTimestamp="2025-10-13 05:51:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 05:51:53.40427084 +0000 UTC m=+1.280985033" watchObservedRunningTime="2025-10-13 05:51:53.405909071 +0000 UTC m=+1.282623266" Oct 13 05:51:53.424655 kubelet[3314]: I1013 05:51:53.424566 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-26-230" podStartSLOduration=4.424525383 podStartE2EDuration="4.424525383s" podCreationTimestamp="2025-10-13 05:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 05:51:53.423165063 +0000 UTC m=+1.299879257" watchObservedRunningTime="2025-10-13 05:51:53.424525383 +0000 UTC m=+1.301239579" Oct 13 05:51:57.650322 kubelet[3314]: I1013 05:51:57.650285 3314 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Oct 13 05:51:57.651176 containerd[2012]: time="2025-10-13T05:51:57.651130225Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Oct 13 05:51:57.651574 kubelet[3314]: I1013 05:51:57.651355 3314 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Oct 13 05:51:58.477516 systemd[1]: Created slice kubepods-besteffort-pode6dba650_9bcc_4888_91e5_d257cac7474d.slice - libcontainer container kubepods-besteffort-pode6dba650_9bcc_4888_91e5_d257cac7474d.slice. Oct 13 05:51:58.514949 kubelet[3314]: I1013 05:51:58.514606 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/e6dba650-9bcc-4888-91e5-d257cac7474d-kube-proxy\") pod \"kube-proxy-gwtv5\" (UID: \"e6dba650-9bcc-4888-91e5-d257cac7474d\") " pod="kube-system/kube-proxy-gwtv5" Oct 13 05:51:58.514949 kubelet[3314]: I1013 05:51:58.514829 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e6dba650-9bcc-4888-91e5-d257cac7474d-xtables-lock\") pod \"kube-proxy-gwtv5\" (UID: \"e6dba650-9bcc-4888-91e5-d257cac7474d\") " pod="kube-system/kube-proxy-gwtv5" Oct 13 05:51:58.616055 kubelet[3314]: I1013 05:51:58.615982 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e6dba650-9bcc-4888-91e5-d257cac7474d-lib-modules\") pod \"kube-proxy-gwtv5\" (UID: \"e6dba650-9bcc-4888-91e5-d257cac7474d\") " pod="kube-system/kube-proxy-gwtv5" Oct 13 05:51:58.616421 kubelet[3314]: I1013 05:51:58.616316 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-786t2\" (UniqueName: \"kubernetes.io/projected/e6dba650-9bcc-4888-91e5-d257cac7474d-kube-api-access-786t2\") pod \"kube-proxy-gwtv5\" (UID: \"e6dba650-9bcc-4888-91e5-d257cac7474d\") " pod="kube-system/kube-proxy-gwtv5" Oct 13 05:51:58.709014 systemd[1]: Created slice kubepods-besteffort-pod3c3b8ffa_9009_49d3_ab08_90f2b4168b4e.slice - libcontainer container kubepods-besteffort-pod3c3b8ffa_9009_49d3_ab08_90f2b4168b4e.slice. Oct 13 05:51:58.717145 kubelet[3314]: I1013 05:51:58.717105 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3c3b8ffa-9009-49d3-ab08-90f2b4168b4e-var-lib-calico\") pod \"tigera-operator-755d956888-s49wj\" (UID: \"3c3b8ffa-9009-49d3-ab08-90f2b4168b4e\") " pod="tigera-operator/tigera-operator-755d956888-s49wj" Oct 13 05:51:58.717642 kubelet[3314]: I1013 05:51:58.717157 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwqk4\" (UniqueName: \"kubernetes.io/projected/3c3b8ffa-9009-49d3-ab08-90f2b4168b4e-kube-api-access-xwqk4\") pod \"tigera-operator-755d956888-s49wj\" (UID: \"3c3b8ffa-9009-49d3-ab08-90f2b4168b4e\") " pod="tigera-operator/tigera-operator-755d956888-s49wj" Oct 13 05:51:58.787805 containerd[2012]: time="2025-10-13T05:51:58.787559125Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gwtv5,Uid:e6dba650-9bcc-4888-91e5-d257cac7474d,Namespace:kube-system,Attempt:0,}" Oct 13 05:51:58.831055 containerd[2012]: time="2025-10-13T05:51:58.830588260Z" level=info msg="connecting to shim 3e5b8d3447b530188968a69d330a42756f42cf8edf5db65badae846eba1d6f46" address="unix:///run/containerd/s/00cea2798c0ab79bdae6079242f89c9ccba1911ab55340ab418476c85fbb1150" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:51:58.870739 systemd[1]: Started cri-containerd-3e5b8d3447b530188968a69d330a42756f42cf8edf5db65badae846eba1d6f46.scope - libcontainer container 3e5b8d3447b530188968a69d330a42756f42cf8edf5db65badae846eba1d6f46. Oct 13 05:51:58.909431 containerd[2012]: time="2025-10-13T05:51:58.909301929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gwtv5,Uid:e6dba650-9bcc-4888-91e5-d257cac7474d,Namespace:kube-system,Attempt:0,} returns sandbox id \"3e5b8d3447b530188968a69d330a42756f42cf8edf5db65badae846eba1d6f46\"" Oct 13 05:51:58.915168 containerd[2012]: time="2025-10-13T05:51:58.915108384Z" level=info msg="CreateContainer within sandbox \"3e5b8d3447b530188968a69d330a42756f42cf8edf5db65badae846eba1d6f46\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Oct 13 05:51:58.934692 containerd[2012]: time="2025-10-13T05:51:58.934647587Z" level=info msg="Container ab1c166c541692798034cc6aaac64ebff001af4a10db7d55492f4b86a7c145a7: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:51:58.953811 containerd[2012]: time="2025-10-13T05:51:58.953754639Z" level=info msg="CreateContainer within sandbox \"3e5b8d3447b530188968a69d330a42756f42cf8edf5db65badae846eba1d6f46\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"ab1c166c541692798034cc6aaac64ebff001af4a10db7d55492f4b86a7c145a7\"" Oct 13 05:51:58.954837 containerd[2012]: time="2025-10-13T05:51:58.954590691Z" level=info msg="StartContainer for \"ab1c166c541692798034cc6aaac64ebff001af4a10db7d55492f4b86a7c145a7\"" Oct 13 05:51:58.957542 containerd[2012]: time="2025-10-13T05:51:58.957501255Z" level=info msg="connecting to shim ab1c166c541692798034cc6aaac64ebff001af4a10db7d55492f4b86a7c145a7" address="unix:///run/containerd/s/00cea2798c0ab79bdae6079242f89c9ccba1911ab55340ab418476c85fbb1150" protocol=ttrpc version=3 Oct 13 05:51:58.979764 systemd[1]: Started cri-containerd-ab1c166c541692798034cc6aaac64ebff001af4a10db7d55492f4b86a7c145a7.scope - libcontainer container ab1c166c541692798034cc6aaac64ebff001af4a10db7d55492f4b86a7c145a7. Oct 13 05:51:59.015477 containerd[2012]: time="2025-10-13T05:51:59.015404189Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-s49wj,Uid:3c3b8ffa-9009-49d3-ab08-90f2b4168b4e,Namespace:tigera-operator,Attempt:0,}" Oct 13 05:51:59.055514 containerd[2012]: time="2025-10-13T05:51:59.053982805Z" level=info msg="StartContainer for \"ab1c166c541692798034cc6aaac64ebff001af4a10db7d55492f4b86a7c145a7\" returns successfully" Oct 13 05:51:59.067384 containerd[2012]: time="2025-10-13T05:51:59.067210439Z" level=info msg="connecting to shim 884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa" address="unix:///run/containerd/s/4bf00479a77d1b0129eb2cdd185e3d2f5f530a71abce3f40d4f2ccf71dc003a3" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:51:59.096660 systemd[1]: Started cri-containerd-884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa.scope - libcontainer container 884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa. Oct 13 05:51:59.172047 containerd[2012]: time="2025-10-13T05:51:59.171991238Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-755d956888-s49wj,Uid:3c3b8ffa-9009-49d3-ab08-90f2b4168b4e,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa\"" Oct 13 05:51:59.176353 containerd[2012]: time="2025-10-13T05:51:59.176217203Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\"" Oct 13 05:51:59.388841 kubelet[3314]: I1013 05:51:59.388643 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-gwtv5" podStartSLOduration=1.388628681 podStartE2EDuration="1.388628681s" podCreationTimestamp="2025-10-13 05:51:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 05:51:59.388189302 +0000 UTC m=+7.264903495" watchObservedRunningTime="2025-10-13 05:51:59.388628681 +0000 UTC m=+7.265342872" Oct 13 05:51:59.758046 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2066153059.mount: Deactivated successfully. Oct 13 05:52:02.138878 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1697705357.mount: Deactivated successfully. Oct 13 05:52:04.232947 containerd[2012]: time="2025-10-13T05:52:04.231792819Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:04.236926 containerd[2012]: time="2025-10-13T05:52:04.236023213Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.6: active requests=0, bytes read=25062609" Oct 13 05:52:04.243726 containerd[2012]: time="2025-10-13T05:52:04.238525892Z" level=info msg="ImageCreate event name:\"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:04.250109 containerd[2012]: time="2025-10-13T05:52:04.250046424Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:04.253473 containerd[2012]: time="2025-10-13T05:52:04.253304112Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.6\" with image id \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\", repo tag \"quay.io/tigera/operator:v1.38.6\", repo digest \"quay.io/tigera/operator@sha256:00a7a9b62f9b9a4e0856128b078539783b8352b07f707bff595cb604cc580f6e\", size \"25058604\" in 5.076995089s" Oct 13 05:52:04.253473 containerd[2012]: time="2025-10-13T05:52:04.253358475Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.6\" returns image reference \"sha256:1911afdd8478c6ca3036ff85614050d5d19acc0f0c3f6a5a7b3e34b38dd309c9\"" Oct 13 05:52:04.268965 containerd[2012]: time="2025-10-13T05:52:04.268683758Z" level=info msg="CreateContainer within sandbox \"884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Oct 13 05:52:04.322402 containerd[2012]: time="2025-10-13T05:52:04.322332607Z" level=info msg="Container 7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:04.328519 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3981453905.mount: Deactivated successfully. Oct 13 05:52:04.339679 containerd[2012]: time="2025-10-13T05:52:04.339634130Z" level=info msg="CreateContainer within sandbox \"884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e\"" Oct 13 05:52:04.340289 containerd[2012]: time="2025-10-13T05:52:04.340234961Z" level=info msg="StartContainer for \"7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e\"" Oct 13 05:52:04.342004 containerd[2012]: time="2025-10-13T05:52:04.341972350Z" level=info msg="connecting to shim 7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e" address="unix:///run/containerd/s/4bf00479a77d1b0129eb2cdd185e3d2f5f530a71abce3f40d4f2ccf71dc003a3" protocol=ttrpc version=3 Oct 13 05:52:04.367900 systemd[1]: Started cri-containerd-7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e.scope - libcontainer container 7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e. Oct 13 05:52:04.424934 containerd[2012]: time="2025-10-13T05:52:04.424869443Z" level=info msg="StartContainer for \"7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e\" returns successfully" Oct 13 05:52:06.518137 update_engine[1984]: I20251013 05:52:06.516501 1984 update_attempter.cc:509] Updating boot flags... Oct 13 05:52:12.241069 sudo[2376]: pam_unix(sudo:session): session closed for user root Oct 13 05:52:12.263768 sshd[2375]: Connection closed by 139.178.68.195 port 45888 Oct 13 05:52:12.264880 sshd-session[2372]: pam_unix(sshd:session): session closed for user core Oct 13 05:52:12.275763 systemd-logind[1980]: Session 7 logged out. Waiting for processes to exit. Oct 13 05:52:12.276607 systemd[1]: sshd@6-172.31.26.230:22-139.178.68.195:45888.service: Deactivated successfully. Oct 13 05:52:12.285368 systemd[1]: session-7.scope: Deactivated successfully. Oct 13 05:52:12.286280 systemd[1]: session-7.scope: Consumed 4.753s CPU time, 150.7M memory peak. Oct 13 05:52:12.292777 systemd-logind[1980]: Removed session 7. Oct 13 05:52:17.509071 kubelet[3314]: I1013 05:52:17.508941 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-755d956888-s49wj" podStartSLOduration=14.420958972 podStartE2EDuration="19.508921387s" podCreationTimestamp="2025-10-13 05:51:58 +0000 UTC" firstStartedPulling="2025-10-13 05:51:59.174863411 +0000 UTC m=+7.051577584" lastFinishedPulling="2025-10-13 05:52:04.262825814 +0000 UTC m=+12.139539999" observedRunningTime="2025-10-13 05:52:05.435507821 +0000 UTC m=+13.312222014" watchObservedRunningTime="2025-10-13 05:52:17.508921387 +0000 UTC m=+25.385635581" Oct 13 05:52:17.522471 systemd[1]: Created slice kubepods-besteffort-pod2b64a7cc_eb23_4424_a23a_1e6c08b2cb04.slice - libcontainer container kubepods-besteffort-pod2b64a7cc_eb23_4424_a23a_1e6c08b2cb04.slice. Oct 13 05:52:17.594905 kubelet[3314]: I1013 05:52:17.594799 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/2b64a7cc-eb23-4424-a23a-1e6c08b2cb04-typha-certs\") pod \"calico-typha-b4c4db68c-f758s\" (UID: \"2b64a7cc-eb23-4424-a23a-1e6c08b2cb04\") " pod="calico-system/calico-typha-b4c4db68c-f758s" Oct 13 05:52:17.595208 kubelet[3314]: I1013 05:52:17.595147 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcbbz\" (UniqueName: \"kubernetes.io/projected/2b64a7cc-eb23-4424-a23a-1e6c08b2cb04-kube-api-access-mcbbz\") pod \"calico-typha-b4c4db68c-f758s\" (UID: \"2b64a7cc-eb23-4424-a23a-1e6c08b2cb04\") " pod="calico-system/calico-typha-b4c4db68c-f758s" Oct 13 05:52:17.595419 kubelet[3314]: I1013 05:52:17.595375 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b64a7cc-eb23-4424-a23a-1e6c08b2cb04-tigera-ca-bundle\") pod \"calico-typha-b4c4db68c-f758s\" (UID: \"2b64a7cc-eb23-4424-a23a-1e6c08b2cb04\") " pod="calico-system/calico-typha-b4c4db68c-f758s" Oct 13 05:52:17.729157 systemd[1]: Created slice kubepods-besteffort-pod8bef18b5_9912_4eda_9397_362fc4a4fce3.slice - libcontainer container kubepods-besteffort-pod8bef18b5_9912_4eda_9397_362fc4a4fce3.slice. Oct 13 05:52:17.797352 kubelet[3314]: I1013 05:52:17.797309 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-var-lib-calico\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.797529 kubelet[3314]: I1013 05:52:17.797362 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-xtables-lock\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.797529 kubelet[3314]: I1013 05:52:17.797387 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-cni-log-dir\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.797529 kubelet[3314]: I1013 05:52:17.797409 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-lib-modules\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.797529 kubelet[3314]: I1013 05:52:17.797438 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dslv9\" (UniqueName: \"kubernetes.io/projected/8bef18b5-9912-4eda-9397-362fc4a4fce3-kube-api-access-dslv9\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.798175 kubelet[3314]: I1013 05:52:17.798140 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-cni-bin-dir\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.798279 kubelet[3314]: I1013 05:52:17.798183 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-var-run-calico\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.798279 kubelet[3314]: I1013 05:52:17.798223 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-cni-net-dir\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.798279 kubelet[3314]: I1013 05:52:17.798244 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-policysync\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.798279 kubelet[3314]: I1013 05:52:17.798268 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bef18b5-9912-4eda-9397-362fc4a4fce3-tigera-ca-bundle\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.798472 kubelet[3314]: I1013 05:52:17.798294 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/8bef18b5-9912-4eda-9397-362fc4a4fce3-flexvol-driver-host\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.798472 kubelet[3314]: I1013 05:52:17.798325 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/8bef18b5-9912-4eda-9397-362fc4a4fce3-node-certs\") pod \"calico-node-6spdp\" (UID: \"8bef18b5-9912-4eda-9397-362fc4a4fce3\") " pod="calico-system/calico-node-6spdp" Oct 13 05:52:17.832733 containerd[2012]: time="2025-10-13T05:52:17.832676485Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-b4c4db68c-f758s,Uid:2b64a7cc-eb23-4424-a23a-1e6c08b2cb04,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:17.886892 containerd[2012]: time="2025-10-13T05:52:17.886683094Z" level=info msg="connecting to shim 2e2859ac810c2ae31bfe446792c3c9013f5384b377e9144ef28b88aad6a8b402" address="unix:///run/containerd/s/fb993e426595fc957c0f5ce631073860c557e500f3922bc34da212a89df4242d" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:17.913684 kubelet[3314]: E1013 05:52:17.913506 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:17.913684 kubelet[3314]: W1013 05:52:17.913538 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:17.914676 kubelet[3314]: E1013 05:52:17.914642 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:17.923158 kubelet[3314]: E1013 05:52:17.923128 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:17.923158 kubelet[3314]: W1013 05:52:17.923156 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:17.923442 kubelet[3314]: E1013 05:52:17.923183 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:17.943701 kubelet[3314]: E1013 05:52:17.943666 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:17.943701 kubelet[3314]: W1013 05:52:17.943697 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:17.943884 kubelet[3314]: E1013 05:52:17.943724 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:17.961843 systemd[1]: Started cri-containerd-2e2859ac810c2ae31bfe446792c3c9013f5384b377e9144ef28b88aad6a8b402.scope - libcontainer container 2e2859ac810c2ae31bfe446792c3c9013f5384b377e9144ef28b88aad6a8b402. Oct 13 05:52:18.034761 containerd[2012]: time="2025-10-13T05:52:18.034716216Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6spdp,Uid:8bef18b5-9912-4eda-9397-362fc4a4fce3,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:18.118085 kubelet[3314]: E1013 05:52:18.117328 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-rf94f" podUID="f4330ab9-5b8d-4f8a-b823-ce3fad0229e2" Oct 13 05:52:18.132379 containerd[2012]: time="2025-10-13T05:52:18.132077183Z" level=info msg="connecting to shim 7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0" address="unix:///run/containerd/s/8db6dadd6bb34363fc9877b654142734983e226c16acdef61e52e84878f5f97e" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:18.156561 kubelet[3314]: E1013 05:52:18.156512 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.156561 kubelet[3314]: W1013 05:52:18.156543 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.156561 kubelet[3314]: E1013 05:52:18.156567 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.157761 kubelet[3314]: E1013 05:52:18.157388 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.157761 kubelet[3314]: W1013 05:52:18.157406 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.157761 kubelet[3314]: E1013 05:52:18.157427 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.157761 kubelet[3314]: E1013 05:52:18.157672 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.157761 kubelet[3314]: W1013 05:52:18.157684 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.157761 kubelet[3314]: E1013 05:52:18.157698 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.158035 kubelet[3314]: E1013 05:52:18.157979 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.158035 kubelet[3314]: W1013 05:52:18.157990 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.158035 kubelet[3314]: E1013 05:52:18.158003 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.160533 kubelet[3314]: E1013 05:52:18.158551 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.160533 kubelet[3314]: W1013 05:52:18.158568 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.160533 kubelet[3314]: E1013 05:52:18.158582 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.160533 kubelet[3314]: E1013 05:52:18.159670 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.160533 kubelet[3314]: W1013 05:52:18.159684 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.160533 kubelet[3314]: E1013 05:52:18.159699 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.160533 kubelet[3314]: E1013 05:52:18.160045 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.160533 kubelet[3314]: W1013 05:52:18.160057 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.160533 kubelet[3314]: E1013 05:52:18.160071 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.163129 kubelet[3314]: E1013 05:52:18.162932 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.163129 kubelet[3314]: W1013 05:52:18.162951 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.163129 kubelet[3314]: E1013 05:52:18.162983 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.163686 kubelet[3314]: E1013 05:52:18.163657 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.163893 kubelet[3314]: W1013 05:52:18.163799 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.163893 kubelet[3314]: E1013 05:52:18.163818 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.164342 kubelet[3314]: E1013 05:52:18.164319 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.164829 kubelet[3314]: W1013 05:52:18.164559 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.164829 kubelet[3314]: E1013 05:52:18.164580 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.166287 kubelet[3314]: E1013 05:52:18.166173 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.166600 kubelet[3314]: W1013 05:52:18.166583 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.167124 kubelet[3314]: E1013 05:52:18.166958 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.168483 kubelet[3314]: E1013 05:52:18.168330 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.168483 kubelet[3314]: W1013 05:52:18.168346 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.168483 kubelet[3314]: E1013 05:52:18.168362 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.169142 kubelet[3314]: E1013 05:52:18.169028 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.169142 kubelet[3314]: W1013 05:52:18.169052 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.169142 kubelet[3314]: E1013 05:52:18.169067 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.170289 kubelet[3314]: E1013 05:52:18.170198 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.170289 kubelet[3314]: W1013 05:52:18.170213 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.170289 kubelet[3314]: E1013 05:52:18.170226 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.170921 kubelet[3314]: E1013 05:52:18.170908 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.171144 kubelet[3314]: W1013 05:52:18.170998 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.171144 kubelet[3314]: E1013 05:52:18.171015 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.171628 kubelet[3314]: E1013 05:52:18.171571 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.172186 kubelet[3314]: W1013 05:52:18.171980 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.172186 kubelet[3314]: E1013 05:52:18.172003 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.172360 kubelet[3314]: E1013 05:52:18.172348 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.172567 kubelet[3314]: W1013 05:52:18.172417 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.172567 kubelet[3314]: E1013 05:52:18.172434 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.173015 kubelet[3314]: E1013 05:52:18.172983 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.173422 kubelet[3314]: W1013 05:52:18.173148 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.173422 kubelet[3314]: E1013 05:52:18.173169 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.173882 kubelet[3314]: E1013 05:52:18.173794 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.173882 kubelet[3314]: W1013 05:52:18.173818 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.173882 kubelet[3314]: E1013 05:52:18.173831 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.174633 kubelet[3314]: E1013 05:52:18.174526 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.174633 kubelet[3314]: W1013 05:52:18.174540 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.174633 kubelet[3314]: E1013 05:52:18.174552 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.188114 systemd[1]: Started cri-containerd-7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0.scope - libcontainer container 7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0. Oct 13 05:52:18.205328 kubelet[3314]: E1013 05:52:18.203694 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.205328 kubelet[3314]: W1013 05:52:18.203801 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.205328 kubelet[3314]: E1013 05:52:18.203831 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.206707 kubelet[3314]: I1013 05:52:18.206672 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f4330ab9-5b8d-4f8a-b823-ce3fad0229e2-socket-dir\") pod \"csi-node-driver-rf94f\" (UID: \"f4330ab9-5b8d-4f8a-b823-ce3fad0229e2\") " pod="calico-system/csi-node-driver-rf94f" Oct 13 05:52:18.207371 kubelet[3314]: E1013 05:52:18.207159 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.207371 kubelet[3314]: W1013 05:52:18.207183 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.207371 kubelet[3314]: E1013 05:52:18.207209 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.207371 kubelet[3314]: I1013 05:52:18.207240 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbwvn\" (UniqueName: \"kubernetes.io/projected/f4330ab9-5b8d-4f8a-b823-ce3fad0229e2-kube-api-access-rbwvn\") pod \"csi-node-driver-rf94f\" (UID: \"f4330ab9-5b8d-4f8a-b823-ce3fad0229e2\") " pod="calico-system/csi-node-driver-rf94f" Oct 13 05:52:18.210601 kubelet[3314]: E1013 05:52:18.209046 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.210601 kubelet[3314]: W1013 05:52:18.209068 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.210601 kubelet[3314]: E1013 05:52:18.209114 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.210601 kubelet[3314]: E1013 05:52:18.209610 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.210601 kubelet[3314]: W1013 05:52:18.209623 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.210601 kubelet[3314]: E1013 05:52:18.209710 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.211101 kubelet[3314]: E1013 05:52:18.211003 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.211101 kubelet[3314]: W1013 05:52:18.211019 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.211321 kubelet[3314]: E1013 05:52:18.211230 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.211321 kubelet[3314]: I1013 05:52:18.211271 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f4330ab9-5b8d-4f8a-b823-ce3fad0229e2-kubelet-dir\") pod \"csi-node-driver-rf94f\" (UID: \"f4330ab9-5b8d-4f8a-b823-ce3fad0229e2\") " pod="calico-system/csi-node-driver-rf94f" Oct 13 05:52:18.213302 kubelet[3314]: E1013 05:52:18.212589 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.213302 kubelet[3314]: W1013 05:52:18.213208 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.213302 kubelet[3314]: E1013 05:52:18.213251 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.215479 kubelet[3314]: E1013 05:52:18.215263 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.217602 kubelet[3314]: W1013 05:52:18.216487 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.217602 kubelet[3314]: E1013 05:52:18.216653 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.217602 kubelet[3314]: E1013 05:52:18.217506 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.217602 kubelet[3314]: W1013 05:52:18.217517 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.217899 kubelet[3314]: E1013 05:52:18.217798 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.217899 kubelet[3314]: I1013 05:52:18.217850 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/f4330ab9-5b8d-4f8a-b823-ce3fad0229e2-varrun\") pod \"csi-node-driver-rf94f\" (UID: \"f4330ab9-5b8d-4f8a-b823-ce3fad0229e2\") " pod="calico-system/csi-node-driver-rf94f" Oct 13 05:52:18.218238 kubelet[3314]: E1013 05:52:18.218226 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.218539 kubelet[3314]: W1013 05:52:18.218522 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.218890 kubelet[3314]: E1013 05:52:18.218849 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.220123 kubelet[3314]: E1013 05:52:18.219953 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.220278 kubelet[3314]: W1013 05:52:18.220232 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.221557 kubelet[3314]: E1013 05:52:18.220579 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.223583 kubelet[3314]: E1013 05:52:18.222052 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.223583 kubelet[3314]: W1013 05:52:18.222073 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.223583 kubelet[3314]: E1013 05:52:18.222093 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.224191 kubelet[3314]: I1013 05:52:18.224085 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f4330ab9-5b8d-4f8a-b823-ce3fad0229e2-registration-dir\") pod \"csi-node-driver-rf94f\" (UID: \"f4330ab9-5b8d-4f8a-b823-ce3fad0229e2\") " pod="calico-system/csi-node-driver-rf94f" Oct 13 05:52:18.224511 kubelet[3314]: E1013 05:52:18.224442 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.224694 kubelet[3314]: W1013 05:52:18.224494 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.224694 kubelet[3314]: E1013 05:52:18.224593 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.225486 kubelet[3314]: E1013 05:52:18.225272 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.225486 kubelet[3314]: W1013 05:52:18.225284 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.226053 kubelet[3314]: E1013 05:52:18.225672 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.228532 kubelet[3314]: E1013 05:52:18.226998 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.228532 kubelet[3314]: W1013 05:52:18.228303 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.228532 kubelet[3314]: E1013 05:52:18.228326 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.228965 kubelet[3314]: E1013 05:52:18.228802 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.228965 kubelet[3314]: W1013 05:52:18.228817 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.228965 kubelet[3314]: E1013 05:52:18.228835 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.262242 containerd[2012]: time="2025-10-13T05:52:18.262165034Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-b4c4db68c-f758s,Uid:2b64a7cc-eb23-4424-a23a-1e6c08b2cb04,Namespace:calico-system,Attempt:0,} returns sandbox id \"2e2859ac810c2ae31bfe446792c3c9013f5384b377e9144ef28b88aad6a8b402\"" Oct 13 05:52:18.267543 containerd[2012]: time="2025-10-13T05:52:18.266229879Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\"" Oct 13 05:52:18.331523 kubelet[3314]: E1013 05:52:18.331497 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.332025 kubelet[3314]: W1013 05:52:18.331738 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.332025 kubelet[3314]: E1013 05:52:18.331765 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.333232 kubelet[3314]: E1013 05:52:18.333144 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.333232 kubelet[3314]: W1013 05:52:18.333167 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.335482 kubelet[3314]: E1013 05:52:18.334623 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.335482 kubelet[3314]: W1013 05:52:18.334646 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.335482 kubelet[3314]: E1013 05:52:18.334669 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.335937 kubelet[3314]: E1013 05:52:18.335887 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.335937 kubelet[3314]: W1013 05:52:18.335905 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.337490 kubelet[3314]: E1013 05:52:18.333444 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.337490 kubelet[3314]: E1013 05:52:18.336972 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.341499 kubelet[3314]: E1013 05:52:18.340528 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.341499 kubelet[3314]: W1013 05:52:18.340551 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.341499 kubelet[3314]: E1013 05:52:18.340573 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.347534 kubelet[3314]: E1013 05:52:18.347472 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.347534 kubelet[3314]: W1013 05:52:18.347501 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.348190 kubelet[3314]: E1013 05:52:18.347830 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.348190 kubelet[3314]: E1013 05:52:18.348095 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.348190 kubelet[3314]: W1013 05:52:18.348112 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.349465 kubelet[3314]: E1013 05:52:18.348472 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.349824 kubelet[3314]: E1013 05:52:18.349800 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.349981 kubelet[3314]: W1013 05:52:18.349957 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.350194 kubelet[3314]: E1013 05:52:18.350144 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.350562 kubelet[3314]: E1013 05:52:18.350549 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.350686 kubelet[3314]: W1013 05:52:18.350670 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.351546 kubelet[3314]: E1013 05:52:18.351381 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.351797 kubelet[3314]: E1013 05:52:18.351787 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.352504 kubelet[3314]: W1013 05:52:18.352485 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.353058 kubelet[3314]: E1013 05:52:18.352994 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.353602 kubelet[3314]: E1013 05:52:18.353531 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.353602 kubelet[3314]: W1013 05:52:18.353546 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.354554 kubelet[3314]: E1013 05:52:18.353737 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.354726 kubelet[3314]: E1013 05:52:18.354667 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.354726 kubelet[3314]: W1013 05:52:18.354679 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.355113 kubelet[3314]: E1013 05:52:18.355068 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.356551 kubelet[3314]: E1013 05:52:18.356438 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.356551 kubelet[3314]: W1013 05:52:18.356474 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.356877 kubelet[3314]: E1013 05:52:18.356803 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.357595 kubelet[3314]: E1013 05:52:18.357447 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.357595 kubelet[3314]: W1013 05:52:18.357572 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.358632 kubelet[3314]: E1013 05:52:18.358532 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.359073 kubelet[3314]: E1013 05:52:18.358970 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.359185 kubelet[3314]: W1013 05:52:18.359161 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.360752 kubelet[3314]: E1013 05:52:18.359440 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.362507 kubelet[3314]: E1013 05:52:18.362423 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.362507 kubelet[3314]: W1013 05:52:18.362483 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.363538 kubelet[3314]: E1013 05:52:18.363427 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.364128 kubelet[3314]: E1013 05:52:18.363954 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.364128 kubelet[3314]: W1013 05:52:18.363989 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.364715 kubelet[3314]: E1013 05:52:18.364510 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.365623 kubelet[3314]: E1013 05:52:18.365607 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.367622 kubelet[3314]: W1013 05:52:18.365708 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.367622 kubelet[3314]: E1013 05:52:18.367522 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.367777 kubelet[3314]: E1013 05:52:18.367753 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.367777 kubelet[3314]: W1013 05:52:18.367768 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.367890 kubelet[3314]: E1013 05:52:18.367876 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.368162 kubelet[3314]: E1013 05:52:18.368071 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.368162 kubelet[3314]: W1013 05:52:18.368089 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.368271 kubelet[3314]: E1013 05:52:18.368187 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.369662 kubelet[3314]: E1013 05:52:18.369636 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.369662 kubelet[3314]: W1013 05:52:18.369654 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.369956 kubelet[3314]: E1013 05:52:18.369681 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.371273 kubelet[3314]: E1013 05:52:18.370125 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.371273 kubelet[3314]: W1013 05:52:18.370140 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.371273 kubelet[3314]: E1013 05:52:18.370156 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.372163 kubelet[3314]: E1013 05:52:18.372015 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.373511 kubelet[3314]: W1013 05:52:18.372478 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.373511 kubelet[3314]: E1013 05:52:18.372511 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.374472 kubelet[3314]: E1013 05:52:18.374261 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.375086 kubelet[3314]: W1013 05:52:18.375064 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.375788 kubelet[3314]: E1013 05:52:18.375210 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.377009 kubelet[3314]: E1013 05:52:18.376581 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.377009 kubelet[3314]: W1013 05:52:18.376601 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.377009 kubelet[3314]: E1013 05:52:18.376619 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:18.389057 containerd[2012]: time="2025-10-13T05:52:18.389010487Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6spdp,Uid:8bef18b5-9912-4eda-9397-362fc4a4fce3,Namespace:calico-system,Attempt:0,} returns sandbox id \"7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0\"" Oct 13 05:52:18.411625 kubelet[3314]: E1013 05:52:18.411513 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:18.411625 kubelet[3314]: W1013 05:52:18.411540 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:18.411625 kubelet[3314]: E1013 05:52:18.411563 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:19.891596 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount744408722.mount: Deactivated successfully. Oct 13 05:52:20.313423 kubelet[3314]: E1013 05:52:20.313371 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-rf94f" podUID="f4330ab9-5b8d-4f8a-b823-ce3fad0229e2" Oct 13 05:52:20.952231 containerd[2012]: time="2025-10-13T05:52:20.952060646Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:20.959647 containerd[2012]: time="2025-10-13T05:52:20.954409488Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.3: active requests=0, bytes read=35237389" Oct 13 05:52:20.959910 containerd[2012]: time="2025-10-13T05:52:20.959654176Z" level=info msg="ImageCreate event name:\"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:20.963610 containerd[2012]: time="2025-10-13T05:52:20.963554921Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:20.964487 containerd[2012]: time="2025-10-13T05:52:20.964271042Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.3\" with image id \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f4a3d61ffda9c98a53adeb412c5af404ca3727a3cc2d0b4ef28d197bdd47ecaa\", size \"35237243\" in 2.697997617s" Oct 13 05:52:20.964487 containerd[2012]: time="2025-10-13T05:52:20.964319156Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.3\" returns image reference \"sha256:1d7bb7b0cce2924d35c7c26f6b6600409ea7c9535074c3d2e517ffbb3a0e0b36\"" Oct 13 05:52:20.966038 containerd[2012]: time="2025-10-13T05:52:20.965981080Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\"" Oct 13 05:52:20.997354 containerd[2012]: time="2025-10-13T05:52:20.997316221Z" level=info msg="CreateContainer within sandbox \"2e2859ac810c2ae31bfe446792c3c9013f5384b377e9144ef28b88aad6a8b402\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Oct 13 05:52:21.007622 containerd[2012]: time="2025-10-13T05:52:21.007579541Z" level=info msg="Container b899dd09cc78487cd8f2651f8d7148b3c987a5cde5793c46d72b4c7db37abfe6: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:21.016773 containerd[2012]: time="2025-10-13T05:52:21.016717802Z" level=info msg="CreateContainer within sandbox \"2e2859ac810c2ae31bfe446792c3c9013f5384b377e9144ef28b88aad6a8b402\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"b899dd09cc78487cd8f2651f8d7148b3c987a5cde5793c46d72b4c7db37abfe6\"" Oct 13 05:52:21.018034 containerd[2012]: time="2025-10-13T05:52:21.017668032Z" level=info msg="StartContainer for \"b899dd09cc78487cd8f2651f8d7148b3c987a5cde5793c46d72b4c7db37abfe6\"" Oct 13 05:52:21.019752 containerd[2012]: time="2025-10-13T05:52:21.019707530Z" level=info msg="connecting to shim b899dd09cc78487cd8f2651f8d7148b3c987a5cde5793c46d72b4c7db37abfe6" address="unix:///run/containerd/s/fb993e426595fc957c0f5ce631073860c557e500f3922bc34da212a89df4242d" protocol=ttrpc version=3 Oct 13 05:52:21.074749 systemd[1]: Started cri-containerd-b899dd09cc78487cd8f2651f8d7148b3c987a5cde5793c46d72b4c7db37abfe6.scope - libcontainer container b899dd09cc78487cd8f2651f8d7148b3c987a5cde5793c46d72b4c7db37abfe6. Oct 13 05:52:21.184632 containerd[2012]: time="2025-10-13T05:52:21.184587671Z" level=info msg="StartContainer for \"b899dd09cc78487cd8f2651f8d7148b3c987a5cde5793c46d72b4c7db37abfe6\" returns successfully" Oct 13 05:52:21.601127 kubelet[3314]: E1013 05:52:21.601099 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.602696 kubelet[3314]: W1013 05:52:21.602669 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.613025 kubelet[3314]: E1013 05:52:21.612979 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.613687 kubelet[3314]: E1013 05:52:21.613539 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.613687 kubelet[3314]: W1013 05:52:21.613563 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.613687 kubelet[3314]: E1013 05:52:21.613588 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.614166 kubelet[3314]: E1013 05:52:21.614077 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.614166 kubelet[3314]: W1013 05:52:21.614094 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.614166 kubelet[3314]: E1013 05:52:21.614109 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.615636 kubelet[3314]: E1013 05:52:21.615531 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.615636 kubelet[3314]: W1013 05:52:21.615549 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.615636 kubelet[3314]: E1013 05:52:21.615567 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.616094 kubelet[3314]: E1013 05:52:21.616015 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.616094 kubelet[3314]: W1013 05:52:21.616030 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.616094 kubelet[3314]: E1013 05:52:21.616046 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.616470 kubelet[3314]: E1013 05:52:21.616415 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.616470 kubelet[3314]: W1013 05:52:21.616430 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.616635 kubelet[3314]: E1013 05:52:21.616444 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.617654 kubelet[3314]: E1013 05:52:21.617566 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.617654 kubelet[3314]: W1013 05:52:21.617581 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.617654 kubelet[3314]: E1013 05:52:21.617598 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.618097 kubelet[3314]: E1013 05:52:21.617999 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.618097 kubelet[3314]: W1013 05:52:21.618014 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.618097 kubelet[3314]: E1013 05:52:21.618028 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.618492 kubelet[3314]: E1013 05:52:21.618407 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.618492 kubelet[3314]: W1013 05:52:21.618422 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.618492 kubelet[3314]: E1013 05:52:21.618435 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.619690 kubelet[3314]: E1013 05:52:21.619610 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.619690 kubelet[3314]: W1013 05:52:21.619625 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.619690 kubelet[3314]: E1013 05:52:21.619640 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.620088 kubelet[3314]: E1013 05:52:21.620013 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.620088 kubelet[3314]: W1013 05:52:21.620027 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.620088 kubelet[3314]: E1013 05:52:21.620039 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.620473 kubelet[3314]: E1013 05:52:21.620394 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.620473 kubelet[3314]: W1013 05:52:21.620407 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.620473 kubelet[3314]: E1013 05:52:21.620419 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.621664 kubelet[3314]: E1013 05:52:21.621580 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.621664 kubelet[3314]: W1013 05:52:21.621595 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.621664 kubelet[3314]: E1013 05:52:21.621609 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.622074 kubelet[3314]: E1013 05:52:21.621999 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.622074 kubelet[3314]: W1013 05:52:21.622014 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.622074 kubelet[3314]: E1013 05:52:21.622027 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.622466 kubelet[3314]: E1013 05:52:21.622374 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.622466 kubelet[3314]: W1013 05:52:21.622387 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.622466 kubelet[3314]: E1013 05:52:21.622399 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.690379 kubelet[3314]: E1013 05:52:21.690292 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.690379 kubelet[3314]: W1013 05:52:21.690321 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.690379 kubelet[3314]: E1013 05:52:21.690348 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.691140 kubelet[3314]: E1013 05:52:21.691098 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.691140 kubelet[3314]: W1013 05:52:21.691118 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.691389 kubelet[3314]: E1013 05:52:21.691305 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.691820 kubelet[3314]: E1013 05:52:21.691657 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.691820 kubelet[3314]: W1013 05:52:21.691673 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.691820 kubelet[3314]: E1013 05:52:21.691689 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.692334 kubelet[3314]: E1013 05:52:21.692077 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.692334 kubelet[3314]: W1013 05:52:21.692091 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.692334 kubelet[3314]: E1013 05:52:21.692105 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.692942 kubelet[3314]: E1013 05:52:21.692842 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.694069 kubelet[3314]: W1013 05:52:21.693098 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.694069 kubelet[3314]: E1013 05:52:21.693138 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.694545 kubelet[3314]: E1013 05:52:21.694511 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.694545 kubelet[3314]: W1013 05:52:21.694526 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.694888 kubelet[3314]: E1013 05:52:21.694871 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.695342 kubelet[3314]: E1013 05:52:21.695308 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.695342 kubelet[3314]: W1013 05:52:21.695324 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.695776 kubelet[3314]: E1013 05:52:21.695749 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.696025 kubelet[3314]: E1013 05:52:21.695991 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.696025 kubelet[3314]: W1013 05:52:21.696007 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.696256 kubelet[3314]: E1013 05:52:21.696207 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.696718 kubelet[3314]: E1013 05:52:21.696704 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.696925 kubelet[3314]: W1013 05:52:21.696764 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.697058 kubelet[3314]: E1013 05:52:21.697010 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.697318 kubelet[3314]: E1013 05:52:21.697290 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.697510 kubelet[3314]: W1013 05:52:21.697404 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.697510 kubelet[3314]: E1013 05:52:21.697435 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.698173 kubelet[3314]: E1013 05:52:21.698159 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.698345 kubelet[3314]: W1013 05:52:21.698276 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.698345 kubelet[3314]: E1013 05:52:21.698316 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.698863 kubelet[3314]: E1013 05:52:21.698807 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.698863 kubelet[3314]: W1013 05:52:21.698835 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.699368 kubelet[3314]: E1013 05:52:21.699322 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.701117 kubelet[3314]: E1013 05:52:21.700166 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.701117 kubelet[3314]: W1013 05:52:21.700280 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.701117 kubelet[3314]: E1013 05:52:21.700303 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.702787 kubelet[3314]: E1013 05:52:21.702759 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.704157 kubelet[3314]: W1013 05:52:21.702862 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.704157 kubelet[3314]: E1013 05:52:21.702878 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.704977 kubelet[3314]: E1013 05:52:21.704959 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.705081 kubelet[3314]: W1013 05:52:21.705068 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.705223 kubelet[3314]: E1013 05:52:21.705179 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.705690 kubelet[3314]: E1013 05:52:21.705591 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.705690 kubelet[3314]: W1013 05:52:21.705603 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.706892 kubelet[3314]: E1013 05:52:21.706255 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.706892 kubelet[3314]: W1013 05:52:21.706270 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.706892 kubelet[3314]: E1013 05:52:21.706286 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.706892 kubelet[3314]: E1013 05:52:21.706316 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:21.707941 kubelet[3314]: E1013 05:52:21.707496 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:21.707941 kubelet[3314]: W1013 05:52:21.707509 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:21.707941 kubelet[3314]: E1013 05:52:21.707524 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.316859 kubelet[3314]: E1013 05:52:22.316677 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-rf94f" podUID="f4330ab9-5b8d-4f8a-b823-ce3fad0229e2" Oct 13 05:52:22.372904 containerd[2012]: time="2025-10-13T05:52:22.372841968Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:22.376666 containerd[2012]: time="2025-10-13T05:52:22.376402027Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3: active requests=0, bytes read=4446660" Oct 13 05:52:22.379570 containerd[2012]: time="2025-10-13T05:52:22.379499974Z" level=info msg="ImageCreate event name:\"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:22.386660 containerd[2012]: time="2025-10-13T05:52:22.386615291Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:22.407627 containerd[2012]: time="2025-10-13T05:52:22.407586357Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" with image id \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:81bdfcd9dbd36624dc35354e8c181c75631ba40e6c7df5820f5f56cea36f0ef9\", size \"5939323\" in 1.441568014s" Oct 13 05:52:22.407876 containerd[2012]: time="2025-10-13T05:52:22.407770712Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.3\" returns image reference \"sha256:4f2b088ed6fdfc6a97ac0650a4ba8171107d6656ce265c592e4c8423fd10e5c4\"" Oct 13 05:52:22.410244 containerd[2012]: time="2025-10-13T05:52:22.410169323Z" level=info msg="CreateContainer within sandbox \"7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Oct 13 05:52:22.430356 containerd[2012]: time="2025-10-13T05:52:22.429542518Z" level=info msg="Container 3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:22.452992 containerd[2012]: time="2025-10-13T05:52:22.450929036Z" level=info msg="CreateContainer within sandbox \"7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8\"" Oct 13 05:52:22.452992 containerd[2012]: time="2025-10-13T05:52:22.452274953Z" level=info msg="StartContainer for \"3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8\"" Oct 13 05:52:22.454891 containerd[2012]: time="2025-10-13T05:52:22.454842589Z" level=info msg="connecting to shim 3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8" address="unix:///run/containerd/s/8db6dadd6bb34363fc9877b654142734983e226c16acdef61e52e84878f5f97e" protocol=ttrpc version=3 Oct 13 05:52:22.482751 systemd[1]: Started cri-containerd-3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8.scope - libcontainer container 3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8. Oct 13 05:52:22.508986 kubelet[3314]: I1013 05:52:22.508906 3314 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 05:52:22.530741 kubelet[3314]: E1013 05:52:22.530441 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.530741 kubelet[3314]: W1013 05:52:22.530686 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.531232 kubelet[3314]: E1013 05:52:22.530712 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.531532 kubelet[3314]: E1013 05:52:22.531513 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.531721 kubelet[3314]: W1013 05:52:22.531595 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.531721 kubelet[3314]: E1013 05:52:22.531623 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.532223 kubelet[3314]: E1013 05:52:22.532190 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.532386 kubelet[3314]: W1013 05:52:22.532204 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.532386 kubelet[3314]: E1013 05:52:22.532344 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.532878 kubelet[3314]: E1013 05:52:22.532865 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.533077 kubelet[3314]: W1013 05:52:22.532981 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.533077 kubelet[3314]: E1013 05:52:22.533001 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.533632 kubelet[3314]: E1013 05:52:22.533531 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.533632 kubelet[3314]: W1013 05:52:22.533557 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.533632 kubelet[3314]: E1013 05:52:22.533571 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.534576 kubelet[3314]: E1013 05:52:22.534495 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.534576 kubelet[3314]: W1013 05:52:22.534511 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.534576 kubelet[3314]: E1013 05:52:22.534525 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.534989 kubelet[3314]: E1013 05:52:22.534975 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.535222 kubelet[3314]: W1013 05:52:22.535149 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.535222 kubelet[3314]: E1013 05:52:22.535169 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.535774 kubelet[3314]: E1013 05:52:22.535759 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.535999 kubelet[3314]: W1013 05:52:22.535850 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.535999 kubelet[3314]: E1013 05:52:22.535870 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.536389 kubelet[3314]: E1013 05:52:22.536376 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.536540 kubelet[3314]: W1013 05:52:22.536482 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.536733 kubelet[3314]: E1013 05:52:22.536577 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.537190 kubelet[3314]: E1013 05:52:22.537177 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.537540 kubelet[3314]: W1013 05:52:22.537270 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.537540 kubelet[3314]: E1013 05:52:22.537288 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.538305 kubelet[3314]: E1013 05:52:22.538266 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.538305 kubelet[3314]: W1013 05:52:22.538285 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.538305 kubelet[3314]: E1013 05:52:22.538301 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.538705 kubelet[3314]: E1013 05:52:22.538690 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.538765 kubelet[3314]: W1013 05:52:22.538705 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.538765 kubelet[3314]: E1013 05:52:22.538720 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.539954 kubelet[3314]: E1013 05:52:22.539932 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.539954 kubelet[3314]: W1013 05:52:22.539949 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.540180 kubelet[3314]: E1013 05:52:22.539965 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.540508 kubelet[3314]: E1013 05:52:22.540488 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.541577 kubelet[3314]: W1013 05:52:22.541556 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.541655 kubelet[3314]: E1013 05:52:22.541586 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.542023 kubelet[3314]: E1013 05:52:22.541995 3314 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Oct 13 05:52:22.542090 kubelet[3314]: W1013 05:52:22.542027 3314 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Oct 13 05:52:22.542090 kubelet[3314]: E1013 05:52:22.542044 3314 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Oct 13 05:52:22.550688 containerd[2012]: time="2025-10-13T05:52:22.550596629Z" level=info msg="StartContainer for \"3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8\" returns successfully" Oct 13 05:52:22.562020 systemd[1]: cri-containerd-3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8.scope: Deactivated successfully. Oct 13 05:52:22.582740 containerd[2012]: time="2025-10-13T05:52:22.582160597Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8\" id:\"3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8\" pid:4174 exited_at:{seconds:1760334742 nanos:566514163}" Oct 13 05:52:22.583565 containerd[2012]: time="2025-10-13T05:52:22.583483704Z" level=info msg="received exit event container_id:\"3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8\" id:\"3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8\" pid:4174 exited_at:{seconds:1760334742 nanos:566514163}" Oct 13 05:52:22.625269 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3d665bea74f29974f57f376a0db08fa56da07c09e637411446ee634dbddfcdf8-rootfs.mount: Deactivated successfully. Oct 13 05:52:23.515071 containerd[2012]: time="2025-10-13T05:52:23.515030294Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\"" Oct 13 05:52:23.538358 kubelet[3314]: I1013 05:52:23.538298 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-b4c4db68c-f758s" podStartSLOduration=3.838334586 podStartE2EDuration="6.538281912s" podCreationTimestamp="2025-10-13 05:52:17 +0000 UTC" firstStartedPulling="2025-10-13 05:52:18.265530179 +0000 UTC m=+26.142244366" lastFinishedPulling="2025-10-13 05:52:20.965477508 +0000 UTC m=+28.842191692" observedRunningTime="2025-10-13 05:52:21.549578386 +0000 UTC m=+29.426292580" watchObservedRunningTime="2025-10-13 05:52:23.538281912 +0000 UTC m=+31.414996131" Oct 13 05:52:24.314496 kubelet[3314]: E1013 05:52:24.314338 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-rf94f" podUID="f4330ab9-5b8d-4f8a-b823-ce3fad0229e2" Oct 13 05:52:26.313746 kubelet[3314]: E1013 05:52:26.313126 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-rf94f" podUID="f4330ab9-5b8d-4f8a-b823-ce3fad0229e2" Oct 13 05:52:28.040808 containerd[2012]: time="2025-10-13T05:52:28.040682261Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:28.042727 containerd[2012]: time="2025-10-13T05:52:28.042669964Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.3: active requests=0, bytes read=70440613" Oct 13 05:52:28.043953 containerd[2012]: time="2025-10-13T05:52:28.043892228Z" level=info msg="ImageCreate event name:\"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:28.047381 containerd[2012]: time="2025-10-13T05:52:28.047070171Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:28.047381 containerd[2012]: time="2025-10-13T05:52:28.047101282Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.3\" with image id \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:73d1e391050490d54e5bee8ff2b1a50a8be1746c98dc530361b00e8c0ab63f87\", size \"71933316\" in 4.532029074s" Oct 13 05:52:28.047381 containerd[2012]: time="2025-10-13T05:52:28.047126533Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.3\" returns image reference \"sha256:034822460c2f667e1f4a7679c843cc35ce1bf2c25dec86f04e07fb403df7e458\"" Oct 13 05:52:28.050785 containerd[2012]: time="2025-10-13T05:52:28.050745713Z" level=info msg="CreateContainer within sandbox \"7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Oct 13 05:52:28.063530 containerd[2012]: time="2025-10-13T05:52:28.061633415Z" level=info msg="Container 8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:28.077216 containerd[2012]: time="2025-10-13T05:52:28.077141499Z" level=info msg="CreateContainer within sandbox \"7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787\"" Oct 13 05:52:28.078253 containerd[2012]: time="2025-10-13T05:52:28.078174454Z" level=info msg="StartContainer for \"8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787\"" Oct 13 05:52:28.080625 containerd[2012]: time="2025-10-13T05:52:28.080577396Z" level=info msg="connecting to shim 8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787" address="unix:///run/containerd/s/8db6dadd6bb34363fc9877b654142734983e226c16acdef61e52e84878f5f97e" protocol=ttrpc version=3 Oct 13 05:52:28.120162 systemd[1]: Started cri-containerd-8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787.scope - libcontainer container 8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787. Oct 13 05:52:28.174783 containerd[2012]: time="2025-10-13T05:52:28.174596215Z" level=info msg="StartContainer for \"8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787\" returns successfully" Oct 13 05:52:28.313745 kubelet[3314]: E1013 05:52:28.313471 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-rf94f" podUID="f4330ab9-5b8d-4f8a-b823-ce3fad0229e2" Oct 13 05:52:29.020976 systemd[1]: cri-containerd-8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787.scope: Deactivated successfully. Oct 13 05:52:29.021356 systemd[1]: cri-containerd-8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787.scope: Consumed 631ms CPU time, 168M memory peak, 7.5M read from disk, 171.3M written to disk. Oct 13 05:52:29.028265 containerd[2012]: time="2025-10-13T05:52:29.028215530Z" level=info msg="TaskExit event in podsandbox handler container_id:\"8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787\" id:\"8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787\" pid:4246 exited_at:{seconds:1760334749 nanos:26545680}" Oct 13 05:52:29.028606 containerd[2012]: time="2025-10-13T05:52:29.028313308Z" level=info msg="received exit event container_id:\"8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787\" id:\"8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787\" pid:4246 exited_at:{seconds:1760334749 nanos:26545680}" Oct 13 05:52:29.073009 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8818b6aabfcca9d5f15b35a2b3031521f895b644123d6ea215118a43c99e9787-rootfs.mount: Deactivated successfully. Oct 13 05:52:29.093574 kubelet[3314]: I1013 05:52:29.093508 3314 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Oct 13 05:52:29.230727 systemd[1]: Created slice kubepods-besteffort-pod53bab1dc_a70e_4691_b0d4_1d4a4f4897f4.slice - libcontainer container kubepods-besteffort-pod53bab1dc_a70e_4691_b0d4_1d4a4f4897f4.slice. Oct 13 05:52:29.239225 systemd[1]: Created slice kubepods-burstable-podc873bf5f_0799_4b45_804e_35c4259dab04.slice - libcontainer container kubepods-burstable-podc873bf5f_0799_4b45_804e_35c4259dab04.slice. Oct 13 05:52:29.261917 kubelet[3314]: I1013 05:52:29.261876 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/53bab1dc-a70e-4691-b0d4-1d4a4f4897f4-calico-apiserver-certs\") pod \"calico-apiserver-6987b98758-9cvbg\" (UID: \"53bab1dc-a70e-4691-b0d4-1d4a4f4897f4\") " pod="calico-apiserver/calico-apiserver-6987b98758-9cvbg" Oct 13 05:52:29.262122 kubelet[3314]: I1013 05:52:29.261933 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6rdx\" (UniqueName: \"kubernetes.io/projected/c873bf5f-0799-4b45-804e-35c4259dab04-kube-api-access-m6rdx\") pod \"coredns-668d6bf9bc-hb8kw\" (UID: \"c873bf5f-0799-4b45-804e-35c4259dab04\") " pod="kube-system/coredns-668d6bf9bc-hb8kw" Oct 13 05:52:29.262122 kubelet[3314]: I1013 05:52:29.261968 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbrlc\" (UniqueName: \"kubernetes.io/projected/53bab1dc-a70e-4691-b0d4-1d4a4f4897f4-kube-api-access-fbrlc\") pod \"calico-apiserver-6987b98758-9cvbg\" (UID: \"53bab1dc-a70e-4691-b0d4-1d4a4f4897f4\") " pod="calico-apiserver/calico-apiserver-6987b98758-9cvbg" Oct 13 05:52:29.262122 kubelet[3314]: I1013 05:52:29.262015 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c873bf5f-0799-4b45-804e-35c4259dab04-config-volume\") pod \"coredns-668d6bf9bc-hb8kw\" (UID: \"c873bf5f-0799-4b45-804e-35c4259dab04\") " pod="kube-system/coredns-668d6bf9bc-hb8kw" Oct 13 05:52:29.278773 systemd[1]: Created slice kubepods-besteffort-pod44b0ad04_1318_4494_bf50_babcc1e76aee.slice - libcontainer container kubepods-besteffort-pod44b0ad04_1318_4494_bf50_babcc1e76aee.slice. Oct 13 05:52:29.289221 systemd[1]: Created slice kubepods-besteffort-podecc13646_e2a1_4792_8aea_d851b5ac2a4b.slice - libcontainer container kubepods-besteffort-podecc13646_e2a1_4792_8aea_d851b5ac2a4b.slice. Oct 13 05:52:29.302377 systemd[1]: Created slice kubepods-besteffort-pod0cf457ad_7845_4a8a_b8e6_90371ccb10b0.slice - libcontainer container kubepods-besteffort-pod0cf457ad_7845_4a8a_b8e6_90371ccb10b0.slice. Oct 13 05:52:29.315507 systemd[1]: Created slice kubepods-besteffort-pod144d2d91_145c_4286_ba51_668f0c8a5336.slice - libcontainer container kubepods-besteffort-pod144d2d91_145c_4286_ba51_668f0c8a5336.slice. Oct 13 05:52:29.326072 systemd[1]: Created slice kubepods-burstable-poda5655587_90e0_4916_ac75_cff211f43863.slice - libcontainer container kubepods-burstable-poda5655587_90e0_4916_ac75_cff211f43863.slice. Oct 13 05:52:29.363067 kubelet[3314]: I1013 05:52:29.363004 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecc13646-e2a1-4792-8aea-d851b5ac2a4b-goldmane-ca-bundle\") pod \"goldmane-54d579b49d-6f9jm\" (UID: \"ecc13646-e2a1-4792-8aea-d851b5ac2a4b\") " pod="calico-system/goldmane-54d579b49d-6f9jm" Oct 13 05:52:29.363984 kubelet[3314]: I1013 05:52:29.363586 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-backend-key-pair\") pod \"whisker-7989ddf9bb-zgnmf\" (UID: \"44b0ad04-1318-4494-bf50-babcc1e76aee\") " pod="calico-system/whisker-7989ddf9bb-zgnmf" Oct 13 05:52:29.363984 kubelet[3314]: I1013 05:52:29.363669 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdgn\" (UniqueName: \"kubernetes.io/projected/a5655587-90e0-4916-ac75-cff211f43863-kube-api-access-bqdgn\") pod \"coredns-668d6bf9bc-jts9b\" (UID: \"a5655587-90e0-4916-ac75-cff211f43863\") " pod="kube-system/coredns-668d6bf9bc-jts9b" Oct 13 05:52:29.363984 kubelet[3314]: I1013 05:52:29.363756 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdqp\" (UniqueName: \"kubernetes.io/projected/ecc13646-e2a1-4792-8aea-d851b5ac2a4b-kube-api-access-bqdqp\") pod \"goldmane-54d579b49d-6f9jm\" (UID: \"ecc13646-e2a1-4792-8aea-d851b5ac2a4b\") " pod="calico-system/goldmane-54d579b49d-6f9jm" Oct 13 05:52:29.363984 kubelet[3314]: I1013 05:52:29.363785 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/0cf457ad-7845-4a8a-b8e6-90371ccb10b0-calico-apiserver-certs\") pod \"calico-apiserver-6987b98758-wc9wm\" (UID: \"0cf457ad-7845-4a8a-b8e6-90371ccb10b0\") " pod="calico-apiserver/calico-apiserver-6987b98758-wc9wm" Oct 13 05:52:29.363984 kubelet[3314]: I1013 05:52:29.363932 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-ca-bundle\") pod \"whisker-7989ddf9bb-zgnmf\" (UID: \"44b0ad04-1318-4494-bf50-babcc1e76aee\") " pod="calico-system/whisker-7989ddf9bb-zgnmf" Oct 13 05:52:29.364645 kubelet[3314]: I1013 05:52:29.364188 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68fbd\" (UniqueName: \"kubernetes.io/projected/44b0ad04-1318-4494-bf50-babcc1e76aee-kube-api-access-68fbd\") pod \"whisker-7989ddf9bb-zgnmf\" (UID: \"44b0ad04-1318-4494-bf50-babcc1e76aee\") " pod="calico-system/whisker-7989ddf9bb-zgnmf" Oct 13 05:52:29.364871 kubelet[3314]: I1013 05:52:29.364581 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdmch\" (UniqueName: \"kubernetes.io/projected/144d2d91-145c-4286-ba51-668f0c8a5336-kube-api-access-rdmch\") pod \"calico-kube-controllers-96476c8b9-95bm6\" (UID: \"144d2d91-145c-4286-ba51-668f0c8a5336\") " pod="calico-system/calico-kube-controllers-96476c8b9-95bm6" Oct 13 05:52:29.365076 kubelet[3314]: I1013 05:52:29.365048 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/ecc13646-e2a1-4792-8aea-d851b5ac2a4b-goldmane-key-pair\") pod \"goldmane-54d579b49d-6f9jm\" (UID: \"ecc13646-e2a1-4792-8aea-d851b5ac2a4b\") " pod="calico-system/goldmane-54d579b49d-6f9jm" Oct 13 05:52:29.365186 kubelet[3314]: I1013 05:52:29.365087 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5655587-90e0-4916-ac75-cff211f43863-config-volume\") pod \"coredns-668d6bf9bc-jts9b\" (UID: \"a5655587-90e0-4916-ac75-cff211f43863\") " pod="kube-system/coredns-668d6bf9bc-jts9b" Oct 13 05:52:29.365186 kubelet[3314]: I1013 05:52:29.365137 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc13646-e2a1-4792-8aea-d851b5ac2a4b-config\") pod \"goldmane-54d579b49d-6f9jm\" (UID: \"ecc13646-e2a1-4792-8aea-d851b5ac2a4b\") " pod="calico-system/goldmane-54d579b49d-6f9jm" Oct 13 05:52:29.365186 kubelet[3314]: I1013 05:52:29.365166 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l27zn\" (UniqueName: \"kubernetes.io/projected/0cf457ad-7845-4a8a-b8e6-90371ccb10b0-kube-api-access-l27zn\") pod \"calico-apiserver-6987b98758-wc9wm\" (UID: \"0cf457ad-7845-4a8a-b8e6-90371ccb10b0\") " pod="calico-apiserver/calico-apiserver-6987b98758-wc9wm" Oct 13 05:52:29.365574 kubelet[3314]: I1013 05:52:29.365190 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/144d2d91-145c-4286-ba51-668f0c8a5336-tigera-ca-bundle\") pod \"calico-kube-controllers-96476c8b9-95bm6\" (UID: \"144d2d91-145c-4286-ba51-668f0c8a5336\") " pod="calico-system/calico-kube-controllers-96476c8b9-95bm6" Oct 13 05:52:29.542578 containerd[2012]: time="2025-10-13T05:52:29.542275587Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\"" Oct 13 05:52:29.559352 containerd[2012]: time="2025-10-13T05:52:29.559302767Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-9cvbg,Uid:53bab1dc-a70e-4691-b0d4-1d4a4f4897f4,Namespace:calico-apiserver,Attempt:0,}" Oct 13 05:52:29.563911 containerd[2012]: time="2025-10-13T05:52:29.563060838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hb8kw,Uid:c873bf5f-0799-4b45-804e-35c4259dab04,Namespace:kube-system,Attempt:0,}" Oct 13 05:52:29.589200 containerd[2012]: time="2025-10-13T05:52:29.588194152Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7989ddf9bb-zgnmf,Uid:44b0ad04-1318-4494-bf50-babcc1e76aee,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:29.603509 containerd[2012]: time="2025-10-13T05:52:29.603160200Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-6f9jm,Uid:ecc13646-e2a1-4792-8aea-d851b5ac2a4b,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:29.611267 containerd[2012]: time="2025-10-13T05:52:29.610477419Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-wc9wm,Uid:0cf457ad-7845-4a8a-b8e6-90371ccb10b0,Namespace:calico-apiserver,Attempt:0,}" Oct 13 05:52:29.642597 containerd[2012]: time="2025-10-13T05:52:29.642544386Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-jts9b,Uid:a5655587-90e0-4916-ac75-cff211f43863,Namespace:kube-system,Attempt:0,}" Oct 13 05:52:29.642793 containerd[2012]: time="2025-10-13T05:52:29.642762923Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-96476c8b9-95bm6,Uid:144d2d91-145c-4286-ba51-668f0c8a5336,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:30.089576 containerd[2012]: time="2025-10-13T05:52:30.085936123Z" level=error msg="Failed to destroy network for sandbox \"599b1646f1860fe22ccef50ed60b3c3920c22ba623f71710943f4511538fa28a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.089576 containerd[2012]: time="2025-10-13T05:52:30.089167432Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-wc9wm,Uid:0cf457ad-7845-4a8a-b8e6-90371ccb10b0,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"599b1646f1860fe22ccef50ed60b3c3920c22ba623f71710943f4511538fa28a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.122760 containerd[2012]: time="2025-10-13T05:52:30.122706819Z" level=error msg="Failed to destroy network for sandbox \"507271f2eb6fd34b441781a9be00a7a031d073dcafb5d67d6d47fee101275f5e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.124757 kubelet[3314]: E1013 05:52:30.123395 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"599b1646f1860fe22ccef50ed60b3c3920c22ba623f71710943f4511538fa28a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.131565 systemd[1]: run-netns-cni\x2d405ba7f9\x2dd5c1\x2dede2\x2db386\x2da50434fc53a4.mount: Deactivated successfully. Oct 13 05:52:30.139183 systemd[1]: run-netns-cni\x2dffba6c59\x2d4b23\x2dae15\x2d0c9e\x2d80a56cf0eb91.mount: Deactivated successfully. Oct 13 05:52:30.144173 kubelet[3314]: E1013 05:52:30.143481 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"599b1646f1860fe22ccef50ed60b3c3920c22ba623f71710943f4511538fa28a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6987b98758-wc9wm" Oct 13 05:52:30.150672 containerd[2012]: time="2025-10-13T05:52:30.150614712Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-6f9jm,Uid:ecc13646-e2a1-4792-8aea-d851b5ac2a4b,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"507271f2eb6fd34b441781a9be00a7a031d073dcafb5d67d6d47fee101275f5e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.155867 kubelet[3314]: E1013 05:52:30.155804 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"599b1646f1860fe22ccef50ed60b3c3920c22ba623f71710943f4511538fa28a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6987b98758-wc9wm" Oct 13 05:52:30.157115 kubelet[3314]: E1013 05:52:30.157028 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"507271f2eb6fd34b441781a9be00a7a031d073dcafb5d67d6d47fee101275f5e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.157115 kubelet[3314]: E1013 05:52:30.157114 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"507271f2eb6fd34b441781a9be00a7a031d073dcafb5d67d6d47fee101275f5e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-6f9jm" Oct 13 05:52:30.158441 kubelet[3314]: E1013 05:52:30.157140 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"507271f2eb6fd34b441781a9be00a7a031d073dcafb5d67d6d47fee101275f5e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-54d579b49d-6f9jm" Oct 13 05:52:30.168017 containerd[2012]: time="2025-10-13T05:52:30.167958784Z" level=error msg="Failed to destroy network for sandbox \"31bb516f05f2b6081c984c62fdb5ab7cd72ba6bd7f1c6247b48e86cb9734ff3c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.173076 systemd[1]: run-netns-cni\x2d5182cf37\x2daf20\x2da0f0\x2dda5e\x2d3e36bac41f84.mount: Deactivated successfully. Oct 13 05:52:30.175104 kubelet[3314]: E1013 05:52:30.174076 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6987b98758-wc9wm_calico-apiserver(0cf457ad-7845-4a8a-b8e6-90371ccb10b0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6987b98758-wc9wm_calico-apiserver(0cf457ad-7845-4a8a-b8e6-90371ccb10b0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"599b1646f1860fe22ccef50ed60b3c3920c22ba623f71710943f4511538fa28a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6987b98758-wc9wm" podUID="0cf457ad-7845-4a8a-b8e6-90371ccb10b0" Oct 13 05:52:30.178163 kubelet[3314]: E1013 05:52:30.177567 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-54d579b49d-6f9jm_calico-system(ecc13646-e2a1-4792-8aea-d851b5ac2a4b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-54d579b49d-6f9jm_calico-system(ecc13646-e2a1-4792-8aea-d851b5ac2a4b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"507271f2eb6fd34b441781a9be00a7a031d073dcafb5d67d6d47fee101275f5e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-54d579b49d-6f9jm" podUID="ecc13646-e2a1-4792-8aea-d851b5ac2a4b" Oct 13 05:52:30.179831 containerd[2012]: time="2025-10-13T05:52:30.179784903Z" level=error msg="Failed to destroy network for sandbox \"0efb0ac7e5e28b1395aef5274af4d45b17abf133788697250094d932f588c3dd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.183057 containerd[2012]: time="2025-10-13T05:52:30.181751041Z" level=error msg="Failed to destroy network for sandbox \"5ecc7ae38f29a6feb3611aba76525332cd5ecd12a8c134e3cdfe5b5db8a3dd03\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.186839 containerd[2012]: time="2025-10-13T05:52:30.186680597Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-jts9b,Uid:a5655587-90e0-4916-ac75-cff211f43863,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"31bb516f05f2b6081c984c62fdb5ab7cd72ba6bd7f1c6247b48e86cb9734ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.187044 systemd[1]: run-netns-cni\x2d5c3e6c67\x2df4bc\x2d61f6\x2d8ecb\x2d9ba8ffe8c82a.mount: Deactivated successfully. Oct 13 05:52:30.188651 kubelet[3314]: E1013 05:52:30.188604 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"31bb516f05f2b6081c984c62fdb5ab7cd72ba6bd7f1c6247b48e86cb9734ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.189046 kubelet[3314]: E1013 05:52:30.188832 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"31bb516f05f2b6081c984c62fdb5ab7cd72ba6bd7f1c6247b48e86cb9734ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-jts9b" Oct 13 05:52:30.189046 kubelet[3314]: E1013 05:52:30.188887 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"31bb516f05f2b6081c984c62fdb5ab7cd72ba6bd7f1c6247b48e86cb9734ff3c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-jts9b" Oct 13 05:52:30.189046 kubelet[3314]: E1013 05:52:30.188964 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-jts9b_kube-system(a5655587-90e0-4916-ac75-cff211f43863)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-jts9b_kube-system(a5655587-90e0-4916-ac75-cff211f43863)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"31bb516f05f2b6081c984c62fdb5ab7cd72ba6bd7f1c6247b48e86cb9734ff3c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-jts9b" podUID="a5655587-90e0-4916-ac75-cff211f43863" Oct 13 05:52:30.194147 containerd[2012]: time="2025-10-13T05:52:30.194090882Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-96476c8b9-95bm6,Uid:144d2d91-145c-4286-ba51-668f0c8a5336,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0efb0ac7e5e28b1395aef5274af4d45b17abf133788697250094d932f588c3dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.194719 kubelet[3314]: E1013 05:52:30.194346 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0efb0ac7e5e28b1395aef5274af4d45b17abf133788697250094d932f588c3dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.194719 kubelet[3314]: E1013 05:52:30.194409 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0efb0ac7e5e28b1395aef5274af4d45b17abf133788697250094d932f588c3dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-96476c8b9-95bm6" Oct 13 05:52:30.194719 kubelet[3314]: E1013 05:52:30.194433 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0efb0ac7e5e28b1395aef5274af4d45b17abf133788697250094d932f588c3dd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-96476c8b9-95bm6" Oct 13 05:52:30.195162 kubelet[3314]: E1013 05:52:30.194513 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-96476c8b9-95bm6_calico-system(144d2d91-145c-4286-ba51-668f0c8a5336)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-96476c8b9-95bm6_calico-system(144d2d91-145c-4286-ba51-668f0c8a5336)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0efb0ac7e5e28b1395aef5274af4d45b17abf133788697250094d932f588c3dd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-96476c8b9-95bm6" podUID="144d2d91-145c-4286-ba51-668f0c8a5336" Oct 13 05:52:30.197625 containerd[2012]: time="2025-10-13T05:52:30.197571056Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7989ddf9bb-zgnmf,Uid:44b0ad04-1318-4494-bf50-babcc1e76aee,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"5ecc7ae38f29a6feb3611aba76525332cd5ecd12a8c134e3cdfe5b5db8a3dd03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.198477 kubelet[3314]: E1013 05:52:30.198402 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5ecc7ae38f29a6feb3611aba76525332cd5ecd12a8c134e3cdfe5b5db8a3dd03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.198742 kubelet[3314]: E1013 05:52:30.198618 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5ecc7ae38f29a6feb3611aba76525332cd5ecd12a8c134e3cdfe5b5db8a3dd03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7989ddf9bb-zgnmf" Oct 13 05:52:30.198742 kubelet[3314]: E1013 05:52:30.198654 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5ecc7ae38f29a6feb3611aba76525332cd5ecd12a8c134e3cdfe5b5db8a3dd03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-7989ddf9bb-zgnmf" Oct 13 05:52:30.199480 kubelet[3314]: E1013 05:52:30.199303 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-7989ddf9bb-zgnmf_calico-system(44b0ad04-1318-4494-bf50-babcc1e76aee)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-7989ddf9bb-zgnmf_calico-system(44b0ad04-1318-4494-bf50-babcc1e76aee)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5ecc7ae38f29a6feb3611aba76525332cd5ecd12a8c134e3cdfe5b5db8a3dd03\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-7989ddf9bb-zgnmf" podUID="44b0ad04-1318-4494-bf50-babcc1e76aee" Oct 13 05:52:30.200722 containerd[2012]: time="2025-10-13T05:52:30.200611961Z" level=error msg="Failed to destroy network for sandbox \"fbb2b30c3f9a205096fba4072e6769f38801780410c8caa885276c746dd89514\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.203064 containerd[2012]: time="2025-10-13T05:52:30.202977243Z" level=error msg="Failed to destroy network for sandbox \"cbffb37683045e68660f358400df256f716eb9bac0dd2d2f98a3fa0215f5294b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.203506 containerd[2012]: time="2025-10-13T05:52:30.203438835Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-9cvbg,Uid:53bab1dc-a70e-4691-b0d4-1d4a4f4897f4,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"fbb2b30c3f9a205096fba4072e6769f38801780410c8caa885276c746dd89514\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.203737 kubelet[3314]: E1013 05:52:30.203701 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fbb2b30c3f9a205096fba4072e6769f38801780410c8caa885276c746dd89514\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.203790 kubelet[3314]: E1013 05:52:30.203764 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fbb2b30c3f9a205096fba4072e6769f38801780410c8caa885276c746dd89514\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6987b98758-9cvbg" Oct 13 05:52:30.203790 kubelet[3314]: E1013 05:52:30.203784 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fbb2b30c3f9a205096fba4072e6769f38801780410c8caa885276c746dd89514\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6987b98758-9cvbg" Oct 13 05:52:30.203967 kubelet[3314]: E1013 05:52:30.203848 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6987b98758-9cvbg_calico-apiserver(53bab1dc-a70e-4691-b0d4-1d4a4f4897f4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6987b98758-9cvbg_calico-apiserver(53bab1dc-a70e-4691-b0d4-1d4a4f4897f4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fbb2b30c3f9a205096fba4072e6769f38801780410c8caa885276c746dd89514\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6987b98758-9cvbg" podUID="53bab1dc-a70e-4691-b0d4-1d4a4f4897f4" Oct 13 05:52:30.210577 containerd[2012]: time="2025-10-13T05:52:30.210362452Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hb8kw,Uid:c873bf5f-0799-4b45-804e-35c4259dab04,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"cbffb37683045e68660f358400df256f716eb9bac0dd2d2f98a3fa0215f5294b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.212181 kubelet[3314]: E1013 05:52:30.212120 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cbffb37683045e68660f358400df256f716eb9bac0dd2d2f98a3fa0215f5294b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.212443 kubelet[3314]: E1013 05:52:30.212217 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cbffb37683045e68660f358400df256f716eb9bac0dd2d2f98a3fa0215f5294b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-hb8kw" Oct 13 05:52:30.212443 kubelet[3314]: E1013 05:52:30.212247 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cbffb37683045e68660f358400df256f716eb9bac0dd2d2f98a3fa0215f5294b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-hb8kw" Oct 13 05:52:30.212443 kubelet[3314]: E1013 05:52:30.212377 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-hb8kw_kube-system(c873bf5f-0799-4b45-804e-35c4259dab04)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-hb8kw_kube-system(c873bf5f-0799-4b45-804e-35c4259dab04)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cbffb37683045e68660f358400df256f716eb9bac0dd2d2f98a3fa0215f5294b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-hb8kw" podUID="c873bf5f-0799-4b45-804e-35c4259dab04" Oct 13 05:52:30.341361 systemd[1]: Created slice kubepods-besteffort-podf4330ab9_5b8d_4f8a_b823_ce3fad0229e2.slice - libcontainer container kubepods-besteffort-podf4330ab9_5b8d_4f8a_b823_ce3fad0229e2.slice. Oct 13 05:52:30.348639 containerd[2012]: time="2025-10-13T05:52:30.348590594Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-rf94f,Uid:f4330ab9-5b8d-4f8a-b823-ce3fad0229e2,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:30.431000 containerd[2012]: time="2025-10-13T05:52:30.430937777Z" level=error msg="Failed to destroy network for sandbox \"455685341bef87abb47ae9688e1f5760f4a82389e294afcf035d8ef1d3f313a3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.433473 containerd[2012]: time="2025-10-13T05:52:30.433399194Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-rf94f,Uid:f4330ab9-5b8d-4f8a-b823-ce3fad0229e2,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"455685341bef87abb47ae9688e1f5760f4a82389e294afcf035d8ef1d3f313a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.433740 kubelet[3314]: E1013 05:52:30.433696 3314 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"455685341bef87abb47ae9688e1f5760f4a82389e294afcf035d8ef1d3f313a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Oct 13 05:52:30.434386 kubelet[3314]: E1013 05:52:30.433877 3314 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"455685341bef87abb47ae9688e1f5760f4a82389e294afcf035d8ef1d3f313a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-rf94f" Oct 13 05:52:30.434386 kubelet[3314]: E1013 05:52:30.433910 3314 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"455685341bef87abb47ae9688e1f5760f4a82389e294afcf035d8ef1d3f313a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-rf94f" Oct 13 05:52:30.434386 kubelet[3314]: E1013 05:52:30.433986 3314 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-rf94f_calico-system(f4330ab9-5b8d-4f8a-b823-ce3fad0229e2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-rf94f_calico-system(f4330ab9-5b8d-4f8a-b823-ce3fad0229e2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"455685341bef87abb47ae9688e1f5760f4a82389e294afcf035d8ef1d3f313a3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-rf94f" podUID="f4330ab9-5b8d-4f8a-b823-ce3fad0229e2" Oct 13 05:52:31.071218 systemd[1]: run-netns-cni\x2db98e2368\x2d3e56\x2d440d\x2d5414\x2d57d5f2dde79e.mount: Deactivated successfully. Oct 13 05:52:31.071320 systemd[1]: run-netns-cni\x2d59cb22a6\x2d631f\x2dd856\x2ddfa5\x2d457be08a3b97.mount: Deactivated successfully. Oct 13 05:52:31.071386 systemd[1]: run-netns-cni\x2daa436507\x2de1f7\x2ddf4f\x2dd9c9\x2d6962b2a687eb.mount: Deactivated successfully. Oct 13 05:52:37.192406 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4030970144.mount: Deactivated successfully. Oct 13 05:52:37.308714 containerd[2012]: time="2025-10-13T05:52:37.307521781Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:37.321205 containerd[2012]: time="2025-10-13T05:52:37.298841659Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.3: active requests=0, bytes read=157078339" Oct 13 05:52:37.326788 containerd[2012]: time="2025-10-13T05:52:37.326739983Z" level=info msg="ImageCreate event name:\"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:37.327431 containerd[2012]: time="2025-10-13T05:52:37.327401012Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:37.346529 containerd[2012]: time="2025-10-13T05:52:37.346446949Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.3\" with image id \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:bcb8146fcaeced1e1c88fad3eaa697f1680746bd23c3e7e8d4535bc484c6f2a1\", size \"157078201\" in 7.801587197s" Oct 13 05:52:37.346529 containerd[2012]: time="2025-10-13T05:52:37.346515927Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.3\" returns image reference \"sha256:ce9c4ac0f175f22c56e80844e65379d9ebe1d8a4e2bbb38dc1db0f53a8826f0f\"" Oct 13 05:52:37.406339 containerd[2012]: time="2025-10-13T05:52:37.406295299Z" level=info msg="CreateContainer within sandbox \"7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Oct 13 05:52:37.438523 containerd[2012]: time="2025-10-13T05:52:37.438311562Z" level=info msg="Container 4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:37.442773 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1490170098.mount: Deactivated successfully. Oct 13 05:52:37.522440 containerd[2012]: time="2025-10-13T05:52:37.522375908Z" level=info msg="CreateContainer within sandbox \"7b3b72cb8df736efee2daa447559f827416055a862a5af9d53983a57407068d0\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\"" Oct 13 05:52:37.523344 containerd[2012]: time="2025-10-13T05:52:37.523003905Z" level=info msg="StartContainer for \"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\"" Oct 13 05:52:37.528631 containerd[2012]: time="2025-10-13T05:52:37.528586872Z" level=info msg="connecting to shim 4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb" address="unix:///run/containerd/s/8db6dadd6bb34363fc9877b654142734983e226c16acdef61e52e84878f5f97e" protocol=ttrpc version=3 Oct 13 05:52:37.673715 systemd[1]: Started cri-containerd-4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb.scope - libcontainer container 4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb. Oct 13 05:52:37.726472 containerd[2012]: time="2025-10-13T05:52:37.726339829Z" level=info msg="StartContainer for \"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\" returns successfully" Oct 13 05:52:37.980710 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Oct 13 05:52:37.981911 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Oct 13 05:52:38.458511 kubelet[3314]: I1013 05:52:38.458024 3314 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-backend-key-pair\") pod \"44b0ad04-1318-4494-bf50-babcc1e76aee\" (UID: \"44b0ad04-1318-4494-bf50-babcc1e76aee\") " Oct 13 05:52:38.458511 kubelet[3314]: I1013 05:52:38.458063 3314 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-ca-bundle\") pod \"44b0ad04-1318-4494-bf50-babcc1e76aee\" (UID: \"44b0ad04-1318-4494-bf50-babcc1e76aee\") " Oct 13 05:52:38.458511 kubelet[3314]: I1013 05:52:38.458102 3314 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68fbd\" (UniqueName: \"kubernetes.io/projected/44b0ad04-1318-4494-bf50-babcc1e76aee-kube-api-access-68fbd\") pod \"44b0ad04-1318-4494-bf50-babcc1e76aee\" (UID: \"44b0ad04-1318-4494-bf50-babcc1e76aee\") " Oct 13 05:52:38.461203 kubelet[3314]: I1013 05:52:38.460924 3314 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "44b0ad04-1318-4494-bf50-babcc1e76aee" (UID: "44b0ad04-1318-4494-bf50-babcc1e76aee"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Oct 13 05:52:38.467197 kubelet[3314]: I1013 05:52:38.467150 3314 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "44b0ad04-1318-4494-bf50-babcc1e76aee" (UID: "44b0ad04-1318-4494-bf50-babcc1e76aee"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Oct 13 05:52:38.467983 systemd[1]: var-lib-kubelet-pods-44b0ad04\x2d1318\x2d4494\x2dbf50\x2dbabcc1e76aee-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Oct 13 05:52:38.469606 kubelet[3314]: I1013 05:52:38.468324 3314 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b0ad04-1318-4494-bf50-babcc1e76aee-kube-api-access-68fbd" (OuterVolumeSpecName: "kube-api-access-68fbd") pod "44b0ad04-1318-4494-bf50-babcc1e76aee" (UID: "44b0ad04-1318-4494-bf50-babcc1e76aee"). InnerVolumeSpecName "kube-api-access-68fbd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Oct 13 05:52:38.474177 systemd[1]: var-lib-kubelet-pods-44b0ad04\x2d1318\x2d4494\x2dbf50\x2dbabcc1e76aee-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d68fbd.mount: Deactivated successfully. Oct 13 05:52:38.558725 kubelet[3314]: I1013 05:52:38.558658 3314 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-68fbd\" (UniqueName: \"kubernetes.io/projected/44b0ad04-1318-4494-bf50-babcc1e76aee-kube-api-access-68fbd\") on node \"ip-172-31-26-230\" DevicePath \"\"" Oct 13 05:52:38.558725 kubelet[3314]: I1013 05:52:38.558697 3314 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-backend-key-pair\") on node \"ip-172-31-26-230\" DevicePath \"\"" Oct 13 05:52:38.558725 kubelet[3314]: I1013 05:52:38.558710 3314 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/44b0ad04-1318-4494-bf50-babcc1e76aee-whisker-ca-bundle\") on node \"ip-172-31-26-230\" DevicePath \"\"" Oct 13 05:52:38.613267 systemd[1]: Removed slice kubepods-besteffort-pod44b0ad04_1318_4494_bf50_babcc1e76aee.slice - libcontainer container kubepods-besteffort-pod44b0ad04_1318_4494_bf50_babcc1e76aee.slice. Oct 13 05:52:38.641901 kubelet[3314]: I1013 05:52:38.641818 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-6spdp" podStartSLOduration=2.664205222 podStartE2EDuration="21.641794066s" podCreationTimestamp="2025-10-13 05:52:17 +0000 UTC" firstStartedPulling="2025-10-13 05:52:18.391827877 +0000 UTC m=+26.268542062" lastFinishedPulling="2025-10-13 05:52:37.369416715 +0000 UTC m=+45.246130906" observedRunningTime="2025-10-13 05:52:38.640126591 +0000 UTC m=+46.516840788" watchObservedRunningTime="2025-10-13 05:52:38.641794066 +0000 UTC m=+46.518508258" Oct 13 05:52:38.740978 systemd[1]: Created slice kubepods-besteffort-pode1fb1f26_74f8_4aba_b400_d91f3f04a929.slice - libcontainer container kubepods-besteffort-pode1fb1f26_74f8_4aba_b400_d91f3f04a929.slice. Oct 13 05:52:38.860604 kubelet[3314]: I1013 05:52:38.860517 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e1fb1f26-74f8-4aba-b400-d91f3f04a929-whisker-backend-key-pair\") pod \"whisker-794f9d9bfc-8g6v5\" (UID: \"e1fb1f26-74f8-4aba-b400-d91f3f04a929\") " pod="calico-system/whisker-794f9d9bfc-8g6v5" Oct 13 05:52:38.860779 kubelet[3314]: I1013 05:52:38.860618 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dvzw\" (UniqueName: \"kubernetes.io/projected/e1fb1f26-74f8-4aba-b400-d91f3f04a929-kube-api-access-5dvzw\") pod \"whisker-794f9d9bfc-8g6v5\" (UID: \"e1fb1f26-74f8-4aba-b400-d91f3f04a929\") " pod="calico-system/whisker-794f9d9bfc-8g6v5" Oct 13 05:52:38.860779 kubelet[3314]: I1013 05:52:38.860676 3314 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1fb1f26-74f8-4aba-b400-d91f3f04a929-whisker-ca-bundle\") pod \"whisker-794f9d9bfc-8g6v5\" (UID: \"e1fb1f26-74f8-4aba-b400-d91f3f04a929\") " pod="calico-system/whisker-794f9d9bfc-8g6v5" Oct 13 05:52:39.046073 containerd[2012]: time="2025-10-13T05:52:39.045917508Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-794f9d9bfc-8g6v5,Uid:e1fb1f26-74f8-4aba-b400-d91f3f04a929,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:39.564905 (udev-worker)[4537]: Network interface NamePolicy= disabled on kernel command line. Oct 13 05:52:39.575566 systemd-networkd[1896]: calicd6bb65ccdd: Link UP Oct 13 05:52:39.576138 systemd-networkd[1896]: calicd6bb65ccdd: Gained carrier Oct 13 05:52:39.598342 containerd[2012]: 2025-10-13 05:52:39.087 [INFO][4566] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 13 05:52:39.598342 containerd[2012]: 2025-10-13 05:52:39.141 [INFO][4566] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0 whisker-794f9d9bfc- calico-system e1fb1f26-74f8-4aba-b400-d91f3f04a929 869 0 2025-10-13 05:52:38 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:794f9d9bfc projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ip-172-31-26-230 whisker-794f9d9bfc-8g6v5 eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calicd6bb65ccdd [] [] }} ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-" Oct 13 05:52:39.598342 containerd[2012]: 2025-10-13 05:52:39.141 [INFO][4566] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" Oct 13 05:52:39.598342 containerd[2012]: 2025-10-13 05:52:39.477 [INFO][4578] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" HandleID="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Workload="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.482 [INFO][4578] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" HandleID="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Workload="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002f5720), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-26-230", "pod":"whisker-794f9d9bfc-8g6v5", "timestamp":"2025-10-13 05:52:39.477221354 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.482 [INFO][4578] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.484 [INFO][4578] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.484 [INFO][4578] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.502 [INFO][4578] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" host="ip-172-31-26-230" Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.521 [INFO][4578] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.528 [INFO][4578] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.531 [INFO][4578] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:39.599058 containerd[2012]: 2025-10-13 05:52:39.534 [INFO][4578] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:39.600724 containerd[2012]: 2025-10-13 05:52:39.534 [INFO][4578] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" host="ip-172-31-26-230" Oct 13 05:52:39.600724 containerd[2012]: 2025-10-13 05:52:39.536 [INFO][4578] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd Oct 13 05:52:39.600724 containerd[2012]: 2025-10-13 05:52:39.541 [INFO][4578] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" host="ip-172-31-26-230" Oct 13 05:52:39.600724 containerd[2012]: 2025-10-13 05:52:39.549 [INFO][4578] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.193/26] block=192.168.91.192/26 handle="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" host="ip-172-31-26-230" Oct 13 05:52:39.600724 containerd[2012]: 2025-10-13 05:52:39.550 [INFO][4578] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.193/26] handle="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" host="ip-172-31-26-230" Oct 13 05:52:39.600724 containerd[2012]: 2025-10-13 05:52:39.550 [INFO][4578] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:39.600724 containerd[2012]: 2025-10-13 05:52:39.550 [INFO][4578] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.193/26] IPv6=[] ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" HandleID="k8s-pod-network.71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Workload="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" Oct 13 05:52:39.601009 containerd[2012]: 2025-10-13 05:52:39.552 [INFO][4566] cni-plugin/k8s.go 418: Populated endpoint ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0", GenerateName:"whisker-794f9d9bfc-", Namespace:"calico-system", SelfLink:"", UID:"e1fb1f26-74f8-4aba-b400-d91f3f04a929", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"794f9d9bfc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"whisker-794f9d9bfc-8g6v5", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.91.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calicd6bb65ccdd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:39.601009 containerd[2012]: 2025-10-13 05:52:39.552 [INFO][4566] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.193/32] ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" Oct 13 05:52:39.601174 containerd[2012]: 2025-10-13 05:52:39.552 [INFO][4566] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicd6bb65ccdd ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" Oct 13 05:52:39.601174 containerd[2012]: 2025-10-13 05:52:39.577 [INFO][4566] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" Oct 13 05:52:39.601261 containerd[2012]: 2025-10-13 05:52:39.577 [INFO][4566] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0", GenerateName:"whisker-794f9d9bfc-", Namespace:"calico-system", SelfLink:"", UID:"e1fb1f26-74f8-4aba-b400-d91f3f04a929", ResourceVersion:"869", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 38, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"794f9d9bfc", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd", Pod:"whisker-794f9d9bfc-8g6v5", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.91.193/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calicd6bb65ccdd", MAC:"f6:93:5a:9e:fa:6e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:39.601366 containerd[2012]: 2025-10-13 05:52:39.590 [INFO][4566] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" Namespace="calico-system" Pod="whisker-794f9d9bfc-8g6v5" WorkloadEndpoint="ip--172--31--26--230-k8s-whisker--794f9d9bfc--8g6v5-eth0" Oct 13 05:52:40.166933 containerd[2012]: time="2025-10-13T05:52:40.166817103Z" level=info msg="connecting to shim 71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd" address="unix:///run/containerd/s/174e7239be406219fb2018d9ad6fe167426e5a595b8b96da5f9e67d62bd77f0a" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:40.230831 systemd[1]: Started cri-containerd-71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd.scope - libcontainer container 71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd. Oct 13 05:52:40.337390 kubelet[3314]: I1013 05:52:40.337344 3314 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b0ad04-1318-4494-bf50-babcc1e76aee" path="/var/lib/kubelet/pods/44b0ad04-1318-4494-bf50-babcc1e76aee/volumes" Oct 13 05:52:40.418783 containerd[2012]: time="2025-10-13T05:52:40.418658001Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-794f9d9bfc-8g6v5,Uid:e1fb1f26-74f8-4aba-b400-d91f3f04a929,Namespace:calico-system,Attempt:0,} returns sandbox id \"71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd\"" Oct 13 05:52:40.433407 containerd[2012]: time="2025-10-13T05:52:40.432648828Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\"" Oct 13 05:52:40.497032 containerd[2012]: time="2025-10-13T05:52:40.496987780Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\" id:\"d48216b3e08d13819c9c61befbaa7ac5b3d592f85714634a23b6e94e3b3eca46\" pid:4692 exit_status:1 exited_at:{seconds:1760334760 nanos:493940214}" Oct 13 05:52:40.719333 containerd[2012]: time="2025-10-13T05:52:40.719118159Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\" id:\"654bf43fd6f81661ddd297942593f48ef4c1614744dd8556c2919443f44cad11\" pid:4768 exit_status:1 exited_at:{seconds:1760334760 nanos:718564293}" Oct 13 05:52:41.321565 containerd[2012]: time="2025-10-13T05:52:41.321067666Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hb8kw,Uid:c873bf5f-0799-4b45-804e-35c4259dab04,Namespace:kube-system,Attempt:0,}" Oct 13 05:52:41.321565 containerd[2012]: time="2025-10-13T05:52:41.321476540Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-9cvbg,Uid:53bab1dc-a70e-4691-b0d4-1d4a4f4897f4,Namespace:calico-apiserver,Attempt:0,}" Oct 13 05:52:41.325409 containerd[2012]: time="2025-10-13T05:52:41.321748474Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-rf94f,Uid:f4330ab9-5b8d-4f8a-b823-ce3fad0229e2,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:41.529505 systemd-networkd[1896]: calicd6bb65ccdd: Gained IPv6LL Oct 13 05:52:41.669205 systemd-networkd[1896]: calid184fe4c4a5: Link UP Oct 13 05:52:41.670910 systemd-networkd[1896]: calid184fe4c4a5: Gained carrier Oct 13 05:52:41.705245 containerd[2012]: 2025-10-13 05:52:41.412 [INFO][4782] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 13 05:52:41.705245 containerd[2012]: 2025-10-13 05:52:41.446 [INFO][4782] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0 coredns-668d6bf9bc- kube-system c873bf5f-0799-4b45-804e-35c4259dab04 799 0 2025-10-13 05:51:58 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-26-230 coredns-668d6bf9bc-hb8kw eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calid184fe4c4a5 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-" Oct 13 05:52:41.705245 containerd[2012]: 2025-10-13 05:52:41.447 [INFO][4782] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" Oct 13 05:52:41.705245 containerd[2012]: 2025-10-13 05:52:41.561 [INFO][4822] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" HandleID="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Workload="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.561 [INFO][4822] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" HandleID="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Workload="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000310490), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-26-230", "pod":"coredns-668d6bf9bc-hb8kw", "timestamp":"2025-10-13 05:52:41.561124579 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.561 [INFO][4822] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.561 [INFO][4822] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.561 [INFO][4822] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.576 [INFO][4822] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" host="ip-172-31-26-230" Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.590 [INFO][4822] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.611 [INFO][4822] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.619 [INFO][4822] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:41.705632 containerd[2012]: 2025-10-13 05:52:41.628 [INFO][4822] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:41.706584 containerd[2012]: 2025-10-13 05:52:41.628 [INFO][4822] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" host="ip-172-31-26-230" Oct 13 05:52:41.706584 containerd[2012]: 2025-10-13 05:52:41.633 [INFO][4822] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131 Oct 13 05:52:41.706584 containerd[2012]: 2025-10-13 05:52:41.639 [INFO][4822] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" host="ip-172-31-26-230" Oct 13 05:52:41.706584 containerd[2012]: 2025-10-13 05:52:41.651 [INFO][4822] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.194/26] block=192.168.91.192/26 handle="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" host="ip-172-31-26-230" Oct 13 05:52:41.706584 containerd[2012]: 2025-10-13 05:52:41.651 [INFO][4822] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.194/26] handle="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" host="ip-172-31-26-230" Oct 13 05:52:41.706584 containerd[2012]: 2025-10-13 05:52:41.651 [INFO][4822] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:41.706584 containerd[2012]: 2025-10-13 05:52:41.652 [INFO][4822] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.194/26] IPv6=[] ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" HandleID="k8s-pod-network.e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Workload="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" Oct 13 05:52:41.707260 containerd[2012]: 2025-10-13 05:52:41.665 [INFO][4782] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"c873bf5f-0799-4b45-804e-35c4259dab04", ResourceVersion:"799", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 51, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"coredns-668d6bf9bc-hb8kw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.91.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid184fe4c4a5", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:41.707260 containerd[2012]: 2025-10-13 05:52:41.666 [INFO][4782] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.194/32] ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" Oct 13 05:52:41.707260 containerd[2012]: 2025-10-13 05:52:41.666 [INFO][4782] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid184fe4c4a5 ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" Oct 13 05:52:41.707260 containerd[2012]: 2025-10-13 05:52:41.671 [INFO][4782] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" Oct 13 05:52:41.707260 containerd[2012]: 2025-10-13 05:52:41.672 [INFO][4782] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"c873bf5f-0799-4b45-804e-35c4259dab04", ResourceVersion:"799", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 51, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131", Pod:"coredns-668d6bf9bc-hb8kw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.91.194/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid184fe4c4a5", MAC:"22:49:3b:1f:4d:92", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:41.707260 containerd[2012]: 2025-10-13 05:52:41.697 [INFO][4782] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" Namespace="kube-system" Pod="coredns-668d6bf9bc-hb8kw" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--hb8kw-eth0" Oct 13 05:52:41.818493 systemd-networkd[1896]: cali858d622e5fd: Link UP Oct 13 05:52:41.822820 systemd-networkd[1896]: cali858d622e5fd: Gained carrier Oct 13 05:52:41.865298 containerd[2012]: time="2025-10-13T05:52:41.865221124Z" level=info msg="connecting to shim e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131" address="unix:///run/containerd/s/9af576cc06fee3d14aa5af8ab7e2878b6daa51b1d05443f3147e4289a72c76d2" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.418 [INFO][4786] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.444 [INFO][4786] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0 calico-apiserver-6987b98758- calico-apiserver 53bab1dc-a70e-4691-b0d4-1d4a4f4897f4 796 0 2025-10-13 05:52:13 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6987b98758 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-26-230 calico-apiserver-6987b98758-9cvbg eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali858d622e5fd [] [] }} ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.445 [INFO][4786] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.564 [INFO][4823] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" HandleID="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Workload="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.564 [INFO][4823] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" HandleID="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Workload="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fea0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-26-230", "pod":"calico-apiserver-6987b98758-9cvbg", "timestamp":"2025-10-13 05:52:41.560990281 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.565 [INFO][4823] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.652 [INFO][4823] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.652 [INFO][4823] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.674 [INFO][4823] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.689 [INFO][4823] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.716 [INFO][4823] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.725 [INFO][4823] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.743 [INFO][4823] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.743 [INFO][4823] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.750 [INFO][4823] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374 Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.759 [INFO][4823] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.776 [INFO][4823] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.195/26] block=192.168.91.192/26 handle="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.776 [INFO][4823] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.195/26] handle="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" host="ip-172-31-26-230" Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.777 [INFO][4823] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:41.897478 containerd[2012]: 2025-10-13 05:52:41.778 [INFO][4823] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.195/26] IPv6=[] ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" HandleID="k8s-pod-network.8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Workload="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" Oct 13 05:52:41.900123 containerd[2012]: 2025-10-13 05:52:41.798 [INFO][4786] cni-plugin/k8s.go 418: Populated endpoint ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0", GenerateName:"calico-apiserver-6987b98758-", Namespace:"calico-apiserver", SelfLink:"", UID:"53bab1dc-a70e-4691-b0d4-1d4a4f4897f4", ResourceVersion:"796", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6987b98758", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"calico-apiserver-6987b98758-9cvbg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.91.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali858d622e5fd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:41.900123 containerd[2012]: 2025-10-13 05:52:41.801 [INFO][4786] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.195/32] ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" Oct 13 05:52:41.900123 containerd[2012]: 2025-10-13 05:52:41.801 [INFO][4786] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali858d622e5fd ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" Oct 13 05:52:41.900123 containerd[2012]: 2025-10-13 05:52:41.830 [INFO][4786] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" Oct 13 05:52:41.900123 containerd[2012]: 2025-10-13 05:52:41.843 [INFO][4786] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0", GenerateName:"calico-apiserver-6987b98758-", Namespace:"calico-apiserver", SelfLink:"", UID:"53bab1dc-a70e-4691-b0d4-1d4a4f4897f4", ResourceVersion:"796", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6987b98758", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374", Pod:"calico-apiserver-6987b98758-9cvbg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.91.195/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali858d622e5fd", MAC:"9a:a2:30:dc:92:64", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:41.900123 containerd[2012]: 2025-10-13 05:52:41.876 [INFO][4786] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-9cvbg" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--9cvbg-eth0" Oct 13 05:52:41.939115 kubelet[3314]: I1013 05:52:41.938522 3314 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 05:52:41.967913 systemd[1]: Started cri-containerd-e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131.scope - libcontainer container e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131. Oct 13 05:52:42.000559 systemd-networkd[1896]: califd096bb4ac7: Link UP Oct 13 05:52:42.002485 systemd-networkd[1896]: califd096bb4ac7: Gained carrier Oct 13 05:52:42.089324 containerd[2012]: time="2025-10-13T05:52:42.089024865Z" level=info msg="connecting to shim 8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374" address="unix:///run/containerd/s/be777ccb68c16d575e2c9ecd05bee31ac0b82ceb07a2522679cbf0519e0800e3" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.469 [INFO][4797] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.502 [INFO][4797] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0 csi-node-driver- calico-system f4330ab9-5b8d-4f8a-b823-ce3fad0229e2 670 0 2025-10-13 05:52:18 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:6c96d95cc7 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ip-172-31-26-230 csi-node-driver-rf94f eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] califd096bb4ac7 [] [] }} ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.503 [INFO][4797] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.622 [INFO][4838] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" HandleID="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Workload="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.623 [INFO][4838] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" HandleID="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Workload="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000384a60), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-26-230", "pod":"csi-node-driver-rf94f", "timestamp":"2025-10-13 05:52:41.621389515 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.623 [INFO][4838] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.778 [INFO][4838] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.778 [INFO][4838] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.804 [INFO][4838] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.853 [INFO][4838] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.876 [INFO][4838] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.884 [INFO][4838] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.901 [INFO][4838] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.902 [INFO][4838] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.910 [INFO][4838] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.931 [INFO][4838] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.958 [INFO][4838] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.196/26] block=192.168.91.192/26 handle="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.959 [INFO][4838] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.196/26] handle="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" host="ip-172-31-26-230" Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.960 [INFO][4838] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:42.106970 containerd[2012]: 2025-10-13 05:52:41.960 [INFO][4838] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.196/26] IPv6=[] ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" HandleID="k8s-pod-network.e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Workload="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" Oct 13 05:52:42.109149 containerd[2012]: 2025-10-13 05:52:41.993 [INFO][4797] cni-plugin/k8s.go 418: Populated endpoint ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f4330ab9-5b8d-4f8a-b823-ce3fad0229e2", ResourceVersion:"670", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"csi-node-driver-rf94f", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.91.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"califd096bb4ac7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:42.109149 containerd[2012]: 2025-10-13 05:52:41.994 [INFO][4797] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.196/32] ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" Oct 13 05:52:42.109149 containerd[2012]: 2025-10-13 05:52:41.994 [INFO][4797] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to califd096bb4ac7 ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" Oct 13 05:52:42.109149 containerd[2012]: 2025-10-13 05:52:42.021 [INFO][4797] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" Oct 13 05:52:42.109149 containerd[2012]: 2025-10-13 05:52:42.023 [INFO][4797] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f4330ab9-5b8d-4f8a-b823-ce3fad0229e2", ResourceVersion:"670", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"6c96d95cc7", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b", Pod:"csi-node-driver-rf94f", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.91.196/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"califd096bb4ac7", MAC:"1e:db:24:d2:13:85", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:42.109149 containerd[2012]: 2025-10-13 05:52:42.085 [INFO][4797] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" Namespace="calico-system" Pod="csi-node-driver-rf94f" WorkloadEndpoint="ip--172--31--26--230-k8s-csi--node--driver--rf94f-eth0" Oct 13 05:52:42.171726 systemd[1]: Started cri-containerd-8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374.scope - libcontainer container 8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374. Oct 13 05:52:42.251127 containerd[2012]: time="2025-10-13T05:52:42.250702077Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-hb8kw,Uid:c873bf5f-0799-4b45-804e-35c4259dab04,Namespace:kube-system,Attempt:0,} returns sandbox id \"e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131\"" Oct 13 05:52:42.296086 containerd[2012]: time="2025-10-13T05:52:42.295533619Z" level=info msg="connecting to shim e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b" address="unix:///run/containerd/s/cf5167d35e739b0137ba751280f7418bd48f70e846ef6e36179157082e408da0" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:42.321225 containerd[2012]: time="2025-10-13T05:52:42.321084677Z" level=info msg="CreateContainer within sandbox \"e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 13 05:52:42.326073 containerd[2012]: time="2025-10-13T05:52:42.326009332Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-96476c8b9-95bm6,Uid:144d2d91-145c-4286-ba51-668f0c8a5336,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:42.351648 containerd[2012]: time="2025-10-13T05:52:42.347684574Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-jts9b,Uid:a5655587-90e0-4916-ac75-cff211f43863,Namespace:kube-system,Attempt:0,}" Oct 13 05:52:42.500849 systemd[1]: Started cri-containerd-e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b.scope - libcontainer container e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b. Oct 13 05:52:42.514085 containerd[2012]: time="2025-10-13T05:52:42.513992643Z" level=info msg="Container f0e2f4779ae79e4e44a228b9b6e87c67c5ec884197439173139605f5931833f5: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:42.589123 containerd[2012]: time="2025-10-13T05:52:42.589015321Z" level=info msg="CreateContainer within sandbox \"e61ff3658b1bd54123ac26054825d751d063134cf070c5f58e657ae637532131\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"f0e2f4779ae79e4e44a228b9b6e87c67c5ec884197439173139605f5931833f5\"" Oct 13 05:52:42.594507 containerd[2012]: time="2025-10-13T05:52:42.594444529Z" level=info msg="StartContainer for \"f0e2f4779ae79e4e44a228b9b6e87c67c5ec884197439173139605f5931833f5\"" Oct 13 05:52:42.604491 containerd[2012]: time="2025-10-13T05:52:42.604412213Z" level=info msg="connecting to shim f0e2f4779ae79e4e44a228b9b6e87c67c5ec884197439173139605f5931833f5" address="unix:///run/containerd/s/9af576cc06fee3d14aa5af8ab7e2878b6daa51b1d05443f3147e4289a72c76d2" protocol=ttrpc version=3 Oct 13 05:52:42.702935 systemd[1]: Started cri-containerd-f0e2f4779ae79e4e44a228b9b6e87c67c5ec884197439173139605f5931833f5.scope - libcontainer container f0e2f4779ae79e4e44a228b9b6e87c67c5ec884197439173139605f5931833f5. Oct 13 05:52:42.866876 containerd[2012]: time="2025-10-13T05:52:42.866823592Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-rf94f,Uid:f4330ab9-5b8d-4f8a-b823-ce3fad0229e2,Namespace:calico-system,Attempt:0,} returns sandbox id \"e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b\"" Oct 13 05:52:42.920668 containerd[2012]: time="2025-10-13T05:52:42.920581385Z" level=info msg="StartContainer for \"f0e2f4779ae79e4e44a228b9b6e87c67c5ec884197439173139605f5931833f5\" returns successfully" Oct 13 05:52:42.941807 containerd[2012]: time="2025-10-13T05:52:42.941756689Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-9cvbg,Uid:53bab1dc-a70e-4691-b0d4-1d4a4f4897f4,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374\"" Oct 13 05:52:43.015652 containerd[2012]: time="2025-10-13T05:52:43.015033661Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:43.018000 containerd[2012]: time="2025-10-13T05:52:43.017953243Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.3: active requests=0, bytes read=4661291" Oct 13 05:52:43.024160 containerd[2012]: time="2025-10-13T05:52:43.023056650Z" level=info msg="ImageCreate event name:\"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:43.030890 containerd[2012]: time="2025-10-13T05:52:43.030583465Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:43.034315 containerd[2012]: time="2025-10-13T05:52:43.033373430Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.3\" with image id \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:e7113761fc7633d515882f0d48b5c8d0b8e62f3f9d34823f2ee194bb16d2ec44\", size \"6153986\" in 2.599717514s" Oct 13 05:52:43.034531 containerd[2012]: time="2025-10-13T05:52:43.034509703Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.3\" returns image reference \"sha256:9a4eedeed4a531acefb7f5d0a1b7e3856b1a9a24d9e7d25deef2134d7a734c2d\"" Oct 13 05:52:43.037032 containerd[2012]: time="2025-10-13T05:52:43.037000558Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\"" Oct 13 05:52:43.050481 containerd[2012]: time="2025-10-13T05:52:43.050350939Z" level=info msg="CreateContainer within sandbox \"71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Oct 13 05:52:43.069498 systemd-networkd[1896]: calid184fe4c4a5: Gained IPv6LL Oct 13 05:52:43.096476 containerd[2012]: time="2025-10-13T05:52:43.092598333Z" level=info msg="Container bbdc124e366308604917dff6dbbaab7630389e713c5c0ff15e1004fc1b92b4f5: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:43.132795 systemd-networkd[1896]: cali53925284b94: Link UP Oct 13 05:52:43.134848 systemd-networkd[1896]: cali53925284b94: Gained carrier Oct 13 05:52:43.136477 containerd[2012]: time="2025-10-13T05:52:43.129430920Z" level=info msg="CreateContainer within sandbox \"71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"bbdc124e366308604917dff6dbbaab7630389e713c5c0ff15e1004fc1b92b4f5\"" Oct 13 05:52:43.148279 containerd[2012]: time="2025-10-13T05:52:43.147101047Z" level=info msg="StartContainer for \"bbdc124e366308604917dff6dbbaab7630389e713c5c0ff15e1004fc1b92b4f5\"" Oct 13 05:52:43.151913 containerd[2012]: time="2025-10-13T05:52:43.151866705Z" level=info msg="connecting to shim bbdc124e366308604917dff6dbbaab7630389e713c5c0ff15e1004fc1b92b4f5" address="unix:///run/containerd/s/174e7239be406219fb2018d9ad6fe167426e5a595b8b96da5f9e67d62bd77f0a" protocol=ttrpc version=3 Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.717 [INFO][4990] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.767 [INFO][4990] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0 coredns-668d6bf9bc- kube-system a5655587-90e0-4916-ac75-cff211f43863 798 0 2025-10-13 05:51:58 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-26-230 coredns-668d6bf9bc-jts9b eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali53925284b94 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.768 [INFO][4990] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.981 [INFO][5057] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" HandleID="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Workload="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.982 [INFO][5057] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" HandleID="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Workload="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004fe70), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-26-230", "pod":"coredns-668d6bf9bc-jts9b", "timestamp":"2025-10-13 05:52:42.973422165 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.982 [INFO][5057] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.982 [INFO][5057] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:42.982 [INFO][5057] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.006 [INFO][5057] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.022 [INFO][5057] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.040 [INFO][5057] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.048 [INFO][5057] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.054 [INFO][5057] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.054 [INFO][5057] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.058 [INFO][5057] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.069 [INFO][5057] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.093 [INFO][5057] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.197/26] block=192.168.91.192/26 handle="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.093 [INFO][5057] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.197/26] handle="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" host="ip-172-31-26-230" Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.093 [INFO][5057] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:43.209494 containerd[2012]: 2025-10-13 05:52:43.093 [INFO][5057] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.197/26] IPv6=[] ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" HandleID="k8s-pod-network.237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Workload="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" Oct 13 05:52:43.213586 containerd[2012]: 2025-10-13 05:52:43.103 [INFO][4990] cni-plugin/k8s.go 418: Populated endpoint ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"a5655587-90e0-4916-ac75-cff211f43863", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 51, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"coredns-668d6bf9bc-jts9b", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.91.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali53925284b94", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:43.213586 containerd[2012]: 2025-10-13 05:52:43.104 [INFO][4990] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.197/32] ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" Oct 13 05:52:43.213586 containerd[2012]: 2025-10-13 05:52:43.104 [INFO][4990] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali53925284b94 ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" Oct 13 05:52:43.213586 containerd[2012]: 2025-10-13 05:52:43.134 [INFO][4990] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" Oct 13 05:52:43.213586 containerd[2012]: 2025-10-13 05:52:43.137 [INFO][4990] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"a5655587-90e0-4916-ac75-cff211f43863", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 51, 58, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e", Pod:"coredns-668d6bf9bc-jts9b", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.91.197/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali53925284b94", MAC:"92:04:a0:a6:48:07", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:43.213586 containerd[2012]: 2025-10-13 05:52:43.200 [INFO][4990] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" Namespace="kube-system" Pod="coredns-668d6bf9bc-jts9b" WorkloadEndpoint="ip--172--31--26--230-k8s-coredns--668d6bf9bc--jts9b-eth0" Oct 13 05:52:43.256739 systemd-networkd[1896]: califd096bb4ac7: Gained IPv6LL Oct 13 05:52:43.284655 containerd[2012]: time="2025-10-13T05:52:43.284600665Z" level=info msg="connecting to shim 237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e" address="unix:///run/containerd/s/4ec9e70f56ff82798619bd0ce515d7521e441cbdb5906a011bfa1614bea8090a" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:43.300606 systemd[1]: Started cri-containerd-bbdc124e366308604917dff6dbbaab7630389e713c5c0ff15e1004fc1b92b4f5.scope - libcontainer container bbdc124e366308604917dff6dbbaab7630389e713c5c0ff15e1004fc1b92b4f5. Oct 13 05:52:43.338329 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3094679453.mount: Deactivated successfully. Oct 13 05:52:43.378126 systemd-networkd[1896]: calieae54ae2f99: Link UP Oct 13 05:52:43.379823 systemd-networkd[1896]: calieae54ae2f99: Gained carrier Oct 13 05:52:43.412266 systemd[1]: Started cri-containerd-237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e.scope - libcontainer container 237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e. Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:42.726 [INFO][4983] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:42.805 [INFO][4983] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0 calico-kube-controllers-96476c8b9- calico-system 144d2d91-145c-4286-ba51-668f0c8a5336 803 0 2025-10-13 05:52:18 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:96476c8b9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ip-172-31-26-230 calico-kube-controllers-96476c8b9-95bm6 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calieae54ae2f99 [] [] }} ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:42.805 [INFO][4983] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:42.988 [INFO][5067] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" HandleID="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Workload="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:42.990 [INFO][5067] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" HandleID="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Workload="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000123970), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-26-230", "pod":"calico-kube-controllers-96476c8b9-95bm6", "timestamp":"2025-10-13 05:52:42.98870827 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:42.990 [INFO][5067] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.096 [INFO][5067] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.096 [INFO][5067] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.144 [INFO][5067] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.167 [INFO][5067] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.211 [INFO][5067] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.219 [INFO][5067] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.230 [INFO][5067] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.232 [INFO][5067] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.245 [INFO][5067] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.268 [INFO][5067] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.329 [INFO][5067] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.198/26] block=192.168.91.192/26 handle="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.331 [INFO][5067] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.198/26] handle="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" host="ip-172-31-26-230" Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.332 [INFO][5067] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:43.430426 containerd[2012]: 2025-10-13 05:52:43.332 [INFO][5067] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.198/26] IPv6=[] ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" HandleID="k8s-pod-network.7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Workload="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" Oct 13 05:52:43.434977 containerd[2012]: 2025-10-13 05:52:43.367 [INFO][4983] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0", GenerateName:"calico-kube-controllers-96476c8b9-", Namespace:"calico-system", SelfLink:"", UID:"144d2d91-145c-4286-ba51-668f0c8a5336", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"96476c8b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"calico-kube-controllers-96476c8b9-95bm6", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.91.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calieae54ae2f99", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:43.434977 containerd[2012]: 2025-10-13 05:52:43.368 [INFO][4983] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.198/32] ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" Oct 13 05:52:43.434977 containerd[2012]: 2025-10-13 05:52:43.368 [INFO][4983] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calieae54ae2f99 ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" Oct 13 05:52:43.434977 containerd[2012]: 2025-10-13 05:52:43.383 [INFO][4983] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" Oct 13 05:52:43.434977 containerd[2012]: 2025-10-13 05:52:43.390 [INFO][4983] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0", GenerateName:"calico-kube-controllers-96476c8b9-", Namespace:"calico-system", SelfLink:"", UID:"144d2d91-145c-4286-ba51-668f0c8a5336", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"96476c8b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b", Pod:"calico-kube-controllers-96476c8b9-95bm6", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.91.198/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calieae54ae2f99", MAC:"c6:aa:d7:ab:0c:57", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:43.434977 containerd[2012]: 2025-10-13 05:52:43.420 [INFO][4983] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" Namespace="calico-system" Pod="calico-kube-controllers-96476c8b9-95bm6" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--kube--controllers--96476c8b9--95bm6-eth0" Oct 13 05:52:43.543217 containerd[2012]: time="2025-10-13T05:52:43.543046286Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-jts9b,Uid:a5655587-90e0-4916-ac75-cff211f43863,Namespace:kube-system,Attempt:0,} returns sandbox id \"237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e\"" Oct 13 05:52:43.559645 containerd[2012]: time="2025-10-13T05:52:43.558618868Z" level=info msg="CreateContainer within sandbox \"237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Oct 13 05:52:43.575387 containerd[2012]: time="2025-10-13T05:52:43.575332822Z" level=info msg="connecting to shim 7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b" address="unix:///run/containerd/s/5ca700ea6e3e05f75b256e5e0dca71c5e8f8a521559d21f11ca93931ca7c998b" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:43.576629 systemd-networkd[1896]: cali858d622e5fd: Gained IPv6LL Oct 13 05:52:43.590714 containerd[2012]: time="2025-10-13T05:52:43.590678747Z" level=info msg="Container 881fa1fc95e35c73f8d281f71f703f8e009ede2ab130a8e14abe4ac55de3b53d: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:43.614473 containerd[2012]: time="2025-10-13T05:52:43.613995395Z" level=info msg="CreateContainer within sandbox \"237bf5f9bdaf4e33ca9791008ca060857112991864d491237d90ce91b141536e\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"881fa1fc95e35c73f8d281f71f703f8e009ede2ab130a8e14abe4ac55de3b53d\"" Oct 13 05:52:43.630871 containerd[2012]: time="2025-10-13T05:52:43.630823471Z" level=info msg="StartContainer for \"881fa1fc95e35c73f8d281f71f703f8e009ede2ab130a8e14abe4ac55de3b53d\"" Oct 13 05:52:43.643034 containerd[2012]: time="2025-10-13T05:52:43.642969120Z" level=info msg="connecting to shim 881fa1fc95e35c73f8d281f71f703f8e009ede2ab130a8e14abe4ac55de3b53d" address="unix:///run/containerd/s/4ec9e70f56ff82798619bd0ce515d7521e441cbdb5906a011bfa1614bea8090a" protocol=ttrpc version=3 Oct 13 05:52:43.687080 containerd[2012]: time="2025-10-13T05:52:43.686819249Z" level=info msg="StartContainer for \"bbdc124e366308604917dff6dbbaab7630389e713c5c0ff15e1004fc1b92b4f5\" returns successfully" Oct 13 05:52:43.695864 systemd[1]: Started cri-containerd-7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b.scope - libcontainer container 7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b. Oct 13 05:52:43.724954 systemd[1]: Started cri-containerd-881fa1fc95e35c73f8d281f71f703f8e009ede2ab130a8e14abe4ac55de3b53d.scope - libcontainer container 881fa1fc95e35c73f8d281f71f703f8e009ede2ab130a8e14abe4ac55de3b53d. Oct 13 05:52:43.816875 containerd[2012]: time="2025-10-13T05:52:43.816748365Z" level=info msg="StartContainer for \"881fa1fc95e35c73f8d281f71f703f8e009ede2ab130a8e14abe4ac55de3b53d\" returns successfully" Oct 13 05:52:43.824159 kubelet[3314]: I1013 05:52:43.815010 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-hb8kw" podStartSLOduration=45.802289276 podStartE2EDuration="45.802289276s" podCreationTimestamp="2025-10-13 05:51:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 05:52:43.80161462 +0000 UTC m=+51.678328816" watchObservedRunningTime="2025-10-13 05:52:43.802289276 +0000 UTC m=+51.679003470" Oct 13 05:52:44.192659 containerd[2012]: time="2025-10-13T05:52:44.192002931Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-96476c8b9-95bm6,Uid:144d2d91-145c-4286-ba51-668f0c8a5336,Namespace:calico-system,Attempt:0,} returns sandbox id \"7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b\"" Oct 13 05:52:44.592778 systemd[1]: Started sshd@7-172.31.26.230:22-139.178.68.195:51064.service - OpenSSH per-connection server daemon (139.178.68.195:51064). Oct 13 05:52:44.664821 systemd-networkd[1896]: calieae54ae2f99: Gained IPv6LL Oct 13 05:52:44.852349 kubelet[3314]: I1013 05:52:44.851645 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-jts9b" podStartSLOduration=46.851616678 podStartE2EDuration="46.851616678s" podCreationTimestamp="2025-10-13 05:51:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 05:52:44.850904315 +0000 UTC m=+52.727618508" watchObservedRunningTime="2025-10-13 05:52:44.851616678 +0000 UTC m=+52.728330870" Oct 13 05:52:44.856659 systemd-networkd[1896]: cali53925284b94: Gained IPv6LL Oct 13 05:52:44.947959 sshd[5327]: Accepted publickey for core from 139.178.68.195 port 51064 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:52:44.961975 sshd-session[5327]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:52:44.973043 systemd-logind[1980]: New session 8 of user core. Oct 13 05:52:44.978664 systemd[1]: Started session-8.scope - Session 8 of User core. Oct 13 05:52:45.006542 containerd[2012]: time="2025-10-13T05:52:45.006502778Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:45.010040 containerd[2012]: time="2025-10-13T05:52:45.009971075Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.3: active requests=0, bytes read=8760527" Oct 13 05:52:45.011811 containerd[2012]: time="2025-10-13T05:52:45.011745719Z" level=info msg="ImageCreate event name:\"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:45.018712 containerd[2012]: time="2025-10-13T05:52:45.018553047Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:45.021150 containerd[2012]: time="2025-10-13T05:52:45.021070865Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.3\" with image id \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:f22c88018d8b58c4ef0052f594b216a13bd6852166ac131a538c5ab2fba23bb2\", size \"10253230\" in 1.983372154s" Oct 13 05:52:45.021567 containerd[2012]: time="2025-10-13T05:52:45.021120116Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.3\" returns image reference \"sha256:666f4e02e75c30547109a06ed75b415a990a970811173aa741379cfaac4d9dd7\"" Oct 13 05:52:45.023056 containerd[2012]: time="2025-10-13T05:52:45.022704215Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Oct 13 05:52:45.025342 containerd[2012]: time="2025-10-13T05:52:45.025301576Z" level=info msg="CreateContainer within sandbox \"e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Oct 13 05:52:45.047124 containerd[2012]: time="2025-10-13T05:52:45.046107110Z" level=info msg="Container fcf7797515946c234720810c673c239751777f7f092188aefde7a1f1992d2685: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:45.058025 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3958096831.mount: Deactivated successfully. Oct 13 05:52:45.069313 containerd[2012]: time="2025-10-13T05:52:45.069267225Z" level=info msg="CreateContainer within sandbox \"e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"fcf7797515946c234720810c673c239751777f7f092188aefde7a1f1992d2685\"" Oct 13 05:52:45.071609 containerd[2012]: time="2025-10-13T05:52:45.070631402Z" level=info msg="StartContainer for \"fcf7797515946c234720810c673c239751777f7f092188aefde7a1f1992d2685\"" Oct 13 05:52:45.078532 containerd[2012]: time="2025-10-13T05:52:45.078482129Z" level=info msg="connecting to shim fcf7797515946c234720810c673c239751777f7f092188aefde7a1f1992d2685" address="unix:///run/containerd/s/cf5167d35e739b0137ba751280f7418bd48f70e846ef6e36179157082e408da0" protocol=ttrpc version=3 Oct 13 05:52:45.132911 systemd[1]: Started cri-containerd-fcf7797515946c234720810c673c239751777f7f092188aefde7a1f1992d2685.scope - libcontainer container fcf7797515946c234720810c673c239751777f7f092188aefde7a1f1992d2685. Oct 13 05:52:45.324902 containerd[2012]: time="2025-10-13T05:52:45.323898039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-6f9jm,Uid:ecc13646-e2a1-4792-8aea-d851b5ac2a4b,Namespace:calico-system,Attempt:0,}" Oct 13 05:52:45.326784 containerd[2012]: time="2025-10-13T05:52:45.326339062Z" level=info msg="StartContainer for \"fcf7797515946c234720810c673c239751777f7f092188aefde7a1f1992d2685\" returns successfully" Oct 13 05:52:45.328617 containerd[2012]: time="2025-10-13T05:52:45.327368938Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-wc9wm,Uid:0cf457ad-7845-4a8a-b8e6-90371ccb10b0,Namespace:calico-apiserver,Attempt:0,}" Oct 13 05:52:45.503683 systemd-networkd[1896]: vxlan.calico: Link UP Oct 13 05:52:45.503692 systemd-networkd[1896]: vxlan.calico: Gained carrier Oct 13 05:52:45.673656 (udev-worker)[4539]: Network interface NamePolicy= disabled on kernel command line. Oct 13 05:52:45.922295 systemd-networkd[1896]: cali0a35678af53: Link UP Oct 13 05:52:45.938927 systemd-networkd[1896]: cali0a35678af53: Gained carrier Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.624 [INFO][5383] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0 calico-apiserver-6987b98758- calico-apiserver 0cf457ad-7845-4a8a-b8e6-90371ccb10b0 802 0 2025-10-13 05:52:13 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6987b98758 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-26-230 calico-apiserver-6987b98758-wc9wm eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali0a35678af53 [] [] }} ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.626 [INFO][5383] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.760 [INFO][5421] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" HandleID="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Workload="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.760 [INFO][5421] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" HandleID="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Workload="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003979a0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-26-230", "pod":"calico-apiserver-6987b98758-wc9wm", "timestamp":"2025-10-13 05:52:45.760310804 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.760 [INFO][5421] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.762 [INFO][5421] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.762 [INFO][5421] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.783 [INFO][5421] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.795 [INFO][5421] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.816 [INFO][5421] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.825 [INFO][5421] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.836 [INFO][5421] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.836 [INFO][5421] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.843 [INFO][5421] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32 Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.851 [INFO][5421] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.895 [INFO][5421] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.199/26] block=192.168.91.192/26 handle="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.897 [INFO][5421] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.199/26] handle="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" host="ip-172-31-26-230" Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.897 [INFO][5421] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:46.024712 containerd[2012]: 2025-10-13 05:52:45.898 [INFO][5421] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.199/26] IPv6=[] ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" HandleID="k8s-pod-network.07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Workload="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" Oct 13 05:52:46.027388 containerd[2012]: 2025-10-13 05:52:45.910 [INFO][5383] cni-plugin/k8s.go 418: Populated endpoint ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0", GenerateName:"calico-apiserver-6987b98758-", Namespace:"calico-apiserver", SelfLink:"", UID:"0cf457ad-7845-4a8a-b8e6-90371ccb10b0", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6987b98758", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"calico-apiserver-6987b98758-wc9wm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.91.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0a35678af53", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:46.027388 containerd[2012]: 2025-10-13 05:52:45.912 [INFO][5383] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.199/32] ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" Oct 13 05:52:46.027388 containerd[2012]: 2025-10-13 05:52:45.912 [INFO][5383] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0a35678af53 ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" Oct 13 05:52:46.027388 containerd[2012]: 2025-10-13 05:52:45.950 [INFO][5383] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" Oct 13 05:52:46.027388 containerd[2012]: 2025-10-13 05:52:45.950 [INFO][5383] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0", GenerateName:"calico-apiserver-6987b98758-", Namespace:"calico-apiserver", SelfLink:"", UID:"0cf457ad-7845-4a8a-b8e6-90371ccb10b0", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 13, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6987b98758", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32", Pod:"calico-apiserver-6987b98758-wc9wm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.91.199/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0a35678af53", MAC:"4e:18:ff:5d:31:b3", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:46.027388 containerd[2012]: 2025-10-13 05:52:46.016 [INFO][5383] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" Namespace="calico-apiserver" Pod="calico-apiserver-6987b98758-wc9wm" WorkloadEndpoint="ip--172--31--26--230-k8s-calico--apiserver--6987b98758--wc9wm-eth0" Oct 13 05:52:46.144130 systemd-networkd[1896]: cali0d6f0fe4fc8: Link UP Oct 13 05:52:46.145801 systemd-networkd[1896]: cali0d6f0fe4fc8: Gained carrier Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.624 [INFO][5379] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0 goldmane-54d579b49d- calico-system ecc13646-e2a1-4792-8aea-d851b5ac2a4b 797 0 2025-10-13 05:52:18 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:54d579b49d projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ip-172-31-26-230 goldmane-54d579b49d-6f9jm eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] cali0d6f0fe4fc8 [] [] }} ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.629 [INFO][5379] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.794 [INFO][5426] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" HandleID="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Workload="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.794 [INFO][5426] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" HandleID="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Workload="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002e35e0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-26-230", "pod":"goldmane-54d579b49d-6f9jm", "timestamp":"2025-10-13 05:52:45.793539723 +0000 UTC"}, Hostname:"ip-172-31-26-230", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.795 [INFO][5426] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.897 [INFO][5426] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.899 [INFO][5426] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-26-230' Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:45.988 [INFO][5426] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.028 [INFO][5426] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.044 [INFO][5426] ipam/ipam.go 511: Trying affinity for 192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.051 [INFO][5426] ipam/ipam.go 158: Attempting to load block cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.057 [INFO][5426] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.91.192/26 host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.058 [INFO][5426] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.91.192/26 handle="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.061 [INFO][5426] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8 Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.078 [INFO][5426] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.91.192/26 handle="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.094 [INFO][5426] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.91.200/26] block=192.168.91.192/26 handle="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.094 [INFO][5426] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.91.200/26] handle="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" host="ip-172-31-26-230" Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.094 [INFO][5426] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Oct 13 05:52:46.197999 containerd[2012]: 2025-10-13 05:52:46.094 [INFO][5426] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.91.200/26] IPv6=[] ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" HandleID="k8s-pod-network.04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Workload="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" Oct 13 05:52:46.198922 containerd[2012]: 2025-10-13 05:52:46.113 [INFO][5379] cni-plugin/k8s.go 418: Populated endpoint ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"ecc13646-e2a1-4792-8aea-d851b5ac2a4b", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"", Pod:"goldmane-54d579b49d-6f9jm", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.91.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali0d6f0fe4fc8", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:46.198922 containerd[2012]: 2025-10-13 05:52:46.113 [INFO][5379] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.91.200/32] ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" Oct 13 05:52:46.198922 containerd[2012]: 2025-10-13 05:52:46.113 [INFO][5379] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0d6f0fe4fc8 ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" Oct 13 05:52:46.198922 containerd[2012]: 2025-10-13 05:52:46.144 [INFO][5379] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" Oct 13 05:52:46.198922 containerd[2012]: 2025-10-13 05:52:46.146 [INFO][5379] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0", GenerateName:"goldmane-54d579b49d-", Namespace:"calico-system", SelfLink:"", UID:"ecc13646-e2a1-4792-8aea-d851b5ac2a4b", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.October, 13, 5, 52, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"54d579b49d", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-26-230", ContainerID:"04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8", Pod:"goldmane-54d579b49d-6f9jm", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.91.200/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"cali0d6f0fe4fc8", MAC:"32:bf:87:a5:43:10", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Oct 13 05:52:46.198922 containerd[2012]: 2025-10-13 05:52:46.182 [INFO][5379] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" Namespace="calico-system" Pod="goldmane-54d579b49d-6f9jm" WorkloadEndpoint="ip--172--31--26--230-k8s-goldmane--54d579b49d--6f9jm-eth0" Oct 13 05:52:46.229215 containerd[2012]: time="2025-10-13T05:52:46.228570065Z" level=info msg="connecting to shim 07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32" address="unix:///run/containerd/s/347a41a53b1d10eb41968fa68bacfb665fca75d93f501ee08c2a17c46c3f3644" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:46.337946 containerd[2012]: time="2025-10-13T05:52:46.337893725Z" level=info msg="connecting to shim 04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8" address="unix:///run/containerd/s/d103e206b8c5f184d6ddf75140e4e5c1067f6d00a5be9195ecdac1b2df72b23d" namespace=k8s.io protocol=ttrpc version=3 Oct 13 05:52:46.361612 systemd[1]: Started cri-containerd-07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32.scope - libcontainer container 07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32. Oct 13 05:52:46.462696 systemd[1]: Started cri-containerd-04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8.scope - libcontainer container 04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8. Oct 13 05:52:46.596280 sshd[5334]: Connection closed by 139.178.68.195 port 51064 Oct 13 05:52:46.596936 sshd-session[5327]: pam_unix(sshd:session): session closed for user core Oct 13 05:52:46.623496 systemd[1]: sshd@7-172.31.26.230:22-139.178.68.195:51064.service: Deactivated successfully. Oct 13 05:52:46.632842 systemd[1]: session-8.scope: Deactivated successfully. Oct 13 05:52:46.644353 systemd-logind[1980]: Session 8 logged out. Waiting for processes to exit. Oct 13 05:52:46.648572 systemd-networkd[1896]: vxlan.calico: Gained IPv6LL Oct 13 05:52:46.656103 systemd-logind[1980]: Removed session 8. Oct 13 05:52:46.802057 containerd[2012]: time="2025-10-13T05:52:46.802013591Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6987b98758-wc9wm,Uid:0cf457ad-7845-4a8a-b8e6-90371ccb10b0,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32\"" Oct 13 05:52:46.867488 containerd[2012]: time="2025-10-13T05:52:46.867267475Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-54d579b49d-6f9jm,Uid:ecc13646-e2a1-4792-8aea-d851b5ac2a4b,Namespace:calico-system,Attempt:0,} returns sandbox id \"04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8\"" Oct 13 05:52:47.352887 systemd-networkd[1896]: cali0d6f0fe4fc8: Gained IPv6LL Oct 13 05:52:47.544617 systemd-networkd[1896]: cali0a35678af53: Gained IPv6LL Oct 13 05:52:48.512961 containerd[2012]: time="2025-10-13T05:52:48.512653744Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:48.515168 containerd[2012]: time="2025-10-13T05:52:48.515057273Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=47333864" Oct 13 05:52:48.517052 containerd[2012]: time="2025-10-13T05:52:48.516552130Z" level=info msg="ImageCreate event name:\"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:48.519106 containerd[2012]: time="2025-10-13T05:52:48.519067706Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:48.519601 containerd[2012]: time="2025-10-13T05:52:48.519548125Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 3.496811983s" Oct 13 05:52:48.519601 containerd[2012]: time="2025-10-13T05:52:48.519602487Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Oct 13 05:52:48.521599 containerd[2012]: time="2025-10-13T05:52:48.521403316Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\"" Oct 13 05:52:48.525263 containerd[2012]: time="2025-10-13T05:52:48.525225320Z" level=info msg="CreateContainer within sandbox \"8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Oct 13 05:52:48.540964 containerd[2012]: time="2025-10-13T05:52:48.540917781Z" level=info msg="Container b518a1f3fbf18f06e0be618b3b36fa309864ada2ed541593151d89c91d0f7844: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:48.553542 containerd[2012]: time="2025-10-13T05:52:48.553492831Z" level=info msg="CreateContainer within sandbox \"8684a7e13144387faecba983f45d1d22f6550998ad63e84a378aee579e2a1374\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"b518a1f3fbf18f06e0be618b3b36fa309864ada2ed541593151d89c91d0f7844\"" Oct 13 05:52:48.556563 containerd[2012]: time="2025-10-13T05:52:48.556522997Z" level=info msg="StartContainer for \"b518a1f3fbf18f06e0be618b3b36fa309864ada2ed541593151d89c91d0f7844\"" Oct 13 05:52:48.557908 containerd[2012]: time="2025-10-13T05:52:48.557868307Z" level=info msg="connecting to shim b518a1f3fbf18f06e0be618b3b36fa309864ada2ed541593151d89c91d0f7844" address="unix:///run/containerd/s/be777ccb68c16d575e2c9ecd05bee31ac0b82ceb07a2522679cbf0519e0800e3" protocol=ttrpc version=3 Oct 13 05:52:48.590259 systemd[1]: Started cri-containerd-b518a1f3fbf18f06e0be618b3b36fa309864ada2ed541593151d89c91d0f7844.scope - libcontainer container b518a1f3fbf18f06e0be618b3b36fa309864ada2ed541593151d89c91d0f7844. Oct 13 05:52:48.678061 containerd[2012]: time="2025-10-13T05:52:48.677018009Z" level=info msg="StartContainer for \"b518a1f3fbf18f06e0be618b3b36fa309864ada2ed541593151d89c91d0f7844\" returns successfully" Oct 13 05:52:48.920063 kubelet[3314]: I1013 05:52:48.919756 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6987b98758-9cvbg" podStartSLOduration=30.347094956 podStartE2EDuration="35.919726811s" podCreationTimestamp="2025-10-13 05:52:13 +0000 UTC" firstStartedPulling="2025-10-13 05:52:42.948498662 +0000 UTC m=+50.825212845" lastFinishedPulling="2025-10-13 05:52:48.521130497 +0000 UTC m=+56.397844700" observedRunningTime="2025-10-13 05:52:48.909602988 +0000 UTC m=+56.786317183" watchObservedRunningTime="2025-10-13 05:52:48.919726811 +0000 UTC m=+56.796441003" Oct 13 05:52:49.742654 ntpd[2236]: Listen normally on 6 vxlan.calico 192.168.91.192:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 6 vxlan.calico 192.168.91.192:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 7 calicd6bb65ccdd [fe80::ecee:eeff:feee:eeee%4]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 8 calid184fe4c4a5 [fe80::ecee:eeff:feee:eeee%5]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 9 cali858d622e5fd [fe80::ecee:eeff:feee:eeee%6]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 10 califd096bb4ac7 [fe80::ecee:eeff:feee:eeee%7]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 11 cali53925284b94 [fe80::ecee:eeff:feee:eeee%8]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 12 calieae54ae2f99 [fe80::ecee:eeff:feee:eeee%9]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 13 vxlan.calico [fe80::64c7:3aff:fee6:159e%10]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 14 cali0a35678af53 [fe80::ecee:eeff:feee:eeee%13]:123 Oct 13 05:52:49.744944 ntpd[2236]: 13 Oct 05:52:49 ntpd[2236]: Listen normally on 15 cali0d6f0fe4fc8 [fe80::ecee:eeff:feee:eeee%14]:123 Oct 13 05:52:49.742732 ntpd[2236]: Listen normally on 7 calicd6bb65ccdd [fe80::ecee:eeff:feee:eeee%4]:123 Oct 13 05:52:49.742763 ntpd[2236]: Listen normally on 8 calid184fe4c4a5 [fe80::ecee:eeff:feee:eeee%5]:123 Oct 13 05:52:49.742789 ntpd[2236]: Listen normally on 9 cali858d622e5fd [fe80::ecee:eeff:feee:eeee%6]:123 Oct 13 05:52:49.742812 ntpd[2236]: Listen normally on 10 califd096bb4ac7 [fe80::ecee:eeff:feee:eeee%7]:123 Oct 13 05:52:49.742838 ntpd[2236]: Listen normally on 11 cali53925284b94 [fe80::ecee:eeff:feee:eeee%8]:123 Oct 13 05:52:49.742866 ntpd[2236]: Listen normally on 12 calieae54ae2f99 [fe80::ecee:eeff:feee:eeee%9]:123 Oct 13 05:52:49.742894 ntpd[2236]: Listen normally on 13 vxlan.calico [fe80::64c7:3aff:fee6:159e%10]:123 Oct 13 05:52:49.742919 ntpd[2236]: Listen normally on 14 cali0a35678af53 [fe80::ecee:eeff:feee:eeee%13]:123 Oct 13 05:52:49.742948 ntpd[2236]: Listen normally on 15 cali0d6f0fe4fc8 [fe80::ecee:eeff:feee:eeee%14]:123 Oct 13 05:52:49.894747 kubelet[3314]: I1013 05:52:49.894466 3314 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 05:52:50.950108 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount369310921.mount: Deactivated successfully. Oct 13 05:52:50.993443 containerd[2012]: time="2025-10-13T05:52:50.993318858Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:50.995353 containerd[2012]: time="2025-10-13T05:52:50.995173519Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.3: active requests=0, bytes read=33085545" Oct 13 05:52:50.998274 containerd[2012]: time="2025-10-13T05:52:50.998232895Z" level=info msg="ImageCreate event name:\"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:51.001999 containerd[2012]: time="2025-10-13T05:52:51.001820428Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:51.003170 containerd[2012]: time="2025-10-13T05:52:51.003127084Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" with image id \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:29becebc47401da9997a2a30f4c25c511a5f379d17275680b048224829af71a5\", size \"33085375\" in 2.481690195s" Oct 13 05:52:51.003887 containerd[2012]: time="2025-10-13T05:52:51.003286614Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.3\" returns image reference \"sha256:7e29b0984d517678aab6ca138482c318989f6f28daf9d3b5dd6e4a5a3115ac16\"" Oct 13 05:52:51.004923 containerd[2012]: time="2025-10-13T05:52:51.004891156Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\"" Oct 13 05:52:51.025209 containerd[2012]: time="2025-10-13T05:52:51.025153339Z" level=info msg="CreateContainer within sandbox \"71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Oct 13 05:52:51.048656 containerd[2012]: time="2025-10-13T05:52:51.048613339Z" level=info msg="Container 40be793faa6c900340d01b6ed370be4aac89ef307c6ff24fc77a7dae031a2f2a: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:51.160369 containerd[2012]: time="2025-10-13T05:52:51.160311284Z" level=info msg="CreateContainer within sandbox \"71c8ed637af1160c419da05a39ce3498ad38bc5986246cec408de670540fe8cd\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"40be793faa6c900340d01b6ed370be4aac89ef307c6ff24fc77a7dae031a2f2a\"" Oct 13 05:52:51.177187 containerd[2012]: time="2025-10-13T05:52:51.177130219Z" level=info msg="StartContainer for \"40be793faa6c900340d01b6ed370be4aac89ef307c6ff24fc77a7dae031a2f2a\"" Oct 13 05:52:51.194444 containerd[2012]: time="2025-10-13T05:52:51.194383288Z" level=info msg="connecting to shim 40be793faa6c900340d01b6ed370be4aac89ef307c6ff24fc77a7dae031a2f2a" address="unix:///run/containerd/s/174e7239be406219fb2018d9ad6fe167426e5a595b8b96da5f9e67d62bd77f0a" protocol=ttrpc version=3 Oct 13 05:52:51.265981 systemd[1]: Started cri-containerd-40be793faa6c900340d01b6ed370be4aac89ef307c6ff24fc77a7dae031a2f2a.scope - libcontainer container 40be793faa6c900340d01b6ed370be4aac89ef307c6ff24fc77a7dae031a2f2a. Oct 13 05:52:51.384562 containerd[2012]: time="2025-10-13T05:52:51.384513737Z" level=info msg="StartContainer for \"40be793faa6c900340d01b6ed370be4aac89ef307c6ff24fc77a7dae031a2f2a\" returns successfully" Oct 13 05:52:51.637492 systemd[1]: Started sshd@8-172.31.26.230:22-139.178.68.195:59730.service - OpenSSH per-connection server daemon (139.178.68.195:59730). Oct 13 05:52:51.890567 sshd[5686]: Accepted publickey for core from 139.178.68.195 port 59730 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:52:51.895030 sshd-session[5686]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:52:51.902205 systemd-logind[1980]: New session 9 of user core. Oct 13 05:52:51.907748 systemd[1]: Started session-9.scope - Session 9 of User core. Oct 13 05:52:53.515319 sshd[5689]: Connection closed by 139.178.68.195 port 59730 Oct 13 05:52:53.516202 sshd-session[5686]: pam_unix(sshd:session): session closed for user core Oct 13 05:52:53.527028 systemd[1]: sshd@8-172.31.26.230:22-139.178.68.195:59730.service: Deactivated successfully. Oct 13 05:52:53.532539 systemd[1]: session-9.scope: Deactivated successfully. Oct 13 05:52:53.535677 systemd-logind[1980]: Session 9 logged out. Waiting for processes to exit. Oct 13 05:52:53.539083 systemd-logind[1980]: Removed session 9. Oct 13 05:52:54.655897 containerd[2012]: time="2025-10-13T05:52:54.655830955Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:54.657139 containerd[2012]: time="2025-10-13T05:52:54.657100690Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.3: active requests=0, bytes read=51277746" Oct 13 05:52:54.658440 containerd[2012]: time="2025-10-13T05:52:54.658379240Z" level=info msg="ImageCreate event name:\"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:54.660304 containerd[2012]: time="2025-10-13T05:52:54.660249691Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:54.660972 containerd[2012]: time="2025-10-13T05:52:54.660889114Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" with image id \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:27c4187717f08f0a5727019d8beb7597665eb47e69eaa1d7d091a7e28913e577\", size \"52770417\" in 3.655965431s" Oct 13 05:52:54.660972 containerd[2012]: time="2025-10-13T05:52:54.660920237Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.3\" returns image reference \"sha256:df191a54fb79de3c693f8b1b864a1bd3bd14f63b3fff9d5fa4869c471ce3cd37\"" Oct 13 05:52:54.688766 containerd[2012]: time="2025-10-13T05:52:54.688649058Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\"" Oct 13 05:52:54.859444 containerd[2012]: time="2025-10-13T05:52:54.859076074Z" level=info msg="CreateContainer within sandbox \"7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Oct 13 05:52:54.874040 containerd[2012]: time="2025-10-13T05:52:54.873155565Z" level=info msg="Container 5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:54.880093 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2280159018.mount: Deactivated successfully. Oct 13 05:52:54.889282 containerd[2012]: time="2025-10-13T05:52:54.889244704Z" level=info msg="CreateContainer within sandbox \"7c315cd3d983533627c6768ffd456ff877353b690637d72e9f8959c72694340b\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223\"" Oct 13 05:52:54.895108 containerd[2012]: time="2025-10-13T05:52:54.895074901Z" level=info msg="StartContainer for \"5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223\"" Oct 13 05:52:54.898636 containerd[2012]: time="2025-10-13T05:52:54.898585070Z" level=info msg="connecting to shim 5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223" address="unix:///run/containerd/s/5ca700ea6e3e05f75b256e5e0dca71c5e8f8a521559d21f11ca93931ca7c998b" protocol=ttrpc version=3 Oct 13 05:52:54.939731 systemd[1]: Started cri-containerd-5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223.scope - libcontainer container 5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223. Oct 13 05:52:55.132811 containerd[2012]: time="2025-10-13T05:52:55.132746852Z" level=info msg="StartContainer for \"5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223\" returns successfully" Oct 13 05:52:56.185427 kubelet[3314]: I1013 05:52:56.183684 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-96476c8b9-95bm6" podStartSLOduration=27.686189877 podStartE2EDuration="38.177970636s" podCreationTimestamp="2025-10-13 05:52:18 +0000 UTC" firstStartedPulling="2025-10-13 05:52:44.196691772 +0000 UTC m=+52.073405948" lastFinishedPulling="2025-10-13 05:52:54.688472502 +0000 UTC m=+62.565186707" observedRunningTime="2025-10-13 05:52:56.156829318 +0000 UTC m=+64.033543511" watchObservedRunningTime="2025-10-13 05:52:56.177970636 +0000 UTC m=+64.054684830" Oct 13 05:52:56.186603 kubelet[3314]: I1013 05:52:56.185925 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-794f9d9bfc-8g6v5" podStartSLOduration=7.611261374 podStartE2EDuration="18.185899641s" podCreationTimestamp="2025-10-13 05:52:38 +0000 UTC" firstStartedPulling="2025-10-13 05:52:40.429919158 +0000 UTC m=+48.306633344" lastFinishedPulling="2025-10-13 05:52:51.004557414 +0000 UTC m=+58.881271611" observedRunningTime="2025-10-13 05:52:51.968731212 +0000 UTC m=+59.845445406" watchObservedRunningTime="2025-10-13 05:52:56.185899641 +0000 UTC m=+64.062613838" Oct 13 05:52:56.303323 containerd[2012]: time="2025-10-13T05:52:56.303271687Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:56.311069 containerd[2012]: time="2025-10-13T05:52:56.310583653Z" level=info msg="ImageCreate event name:\"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:56.312525 containerd[2012]: time="2025-10-13T05:52:56.311736472Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3: active requests=0, bytes read=14698542" Oct 13 05:52:56.321111 containerd[2012]: time="2025-10-13T05:52:56.321057262Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:56.321963 containerd[2012]: time="2025-10-13T05:52:56.321928655Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" with image id \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:731ab232ca708102ab332340b1274d5cd656aa896ecc5368ee95850b811df86f\", size \"16191197\" in 1.633242041s" Oct 13 05:52:56.323536 containerd[2012]: time="2025-10-13T05:52:56.321970079Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.3\" returns image reference \"sha256:b8f31c4fdaed3fa08af64de3d37d65a4c2ea0d9f6f522cb60d2e0cb424f8dd8a\"" Oct 13 05:52:56.327296 containerd[2012]: time="2025-10-13T05:52:56.326661908Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\"" Oct 13 05:52:56.331776 containerd[2012]: time="2025-10-13T05:52:56.331740770Z" level=info msg="CreateContainer within sandbox \"e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Oct 13 05:52:56.350856 containerd[2012]: time="2025-10-13T05:52:56.350807774Z" level=info msg="Container 0b090a93fa9dddf6b6b64116c228ab8412d35df0d0cb3c44e804ef3192fa2f8e: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:56.362355 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4128497726.mount: Deactivated successfully. Oct 13 05:52:56.387115 containerd[2012]: time="2025-10-13T05:52:56.387032080Z" level=info msg="CreateContainer within sandbox \"e8ae2704be717da0bcbd0f28025fc4dc8705b446f269dc23eccb7d735d18031b\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"0b090a93fa9dddf6b6b64116c228ab8412d35df0d0cb3c44e804ef3192fa2f8e\"" Oct 13 05:52:56.388687 containerd[2012]: time="2025-10-13T05:52:56.388628420Z" level=info msg="StartContainer for \"0b090a93fa9dddf6b6b64116c228ab8412d35df0d0cb3c44e804ef3192fa2f8e\"" Oct 13 05:52:56.393239 containerd[2012]: time="2025-10-13T05:52:56.391945732Z" level=info msg="connecting to shim 0b090a93fa9dddf6b6b64116c228ab8412d35df0d0cb3c44e804ef3192fa2f8e" address="unix:///run/containerd/s/cf5167d35e739b0137ba751280f7418bd48f70e846ef6e36179157082e408da0" protocol=ttrpc version=3 Oct 13 05:52:56.437812 systemd[1]: Started cri-containerd-0b090a93fa9dddf6b6b64116c228ab8412d35df0d0cb3c44e804ef3192fa2f8e.scope - libcontainer container 0b090a93fa9dddf6b6b64116c228ab8412d35df0d0cb3c44e804ef3192fa2f8e. Oct 13 05:52:56.492049 containerd[2012]: time="2025-10-13T05:52:56.492010533Z" level=info msg="StartContainer for \"0b090a93fa9dddf6b6b64116c228ab8412d35df0d0cb3c44e804ef3192fa2f8e\" returns successfully" Oct 13 05:52:56.750285 kubelet[3314]: I1013 05:52:56.749906 3314 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Oct 13 05:52:56.750872 kubelet[3314]: I1013 05:52:56.750498 3314 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Oct 13 05:52:56.756747 containerd[2012]: time="2025-10-13T05:52:56.756698509Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:52:56.760080 containerd[2012]: time="2025-10-13T05:52:56.759217267Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.3: active requests=0, bytes read=77" Oct 13 05:52:56.760750 containerd[2012]: time="2025-10-13T05:52:56.760698880Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" with image id \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:6a24147f11c1edce9d6ba79bdb0c2beadec53853fb43438a287291e67b41e51b\", size \"48826583\" in 433.999703ms" Oct 13 05:52:56.760750 containerd[2012]: time="2025-10-13T05:52:56.760747626Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.3\" returns image reference \"sha256:879f2443aed0573271114108bfec35d3e76419f98282ef796c646d0986c5ba6a\"" Oct 13 05:52:56.764918 containerd[2012]: time="2025-10-13T05:52:56.764791787Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\"" Oct 13 05:52:56.766445 containerd[2012]: time="2025-10-13T05:52:56.765566534Z" level=info msg="CreateContainer within sandbox \"07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Oct 13 05:52:56.855034 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3606787475.mount: Deactivated successfully. Oct 13 05:52:56.861368 containerd[2012]: time="2025-10-13T05:52:56.861319685Z" level=info msg="Container bb96b44f39f95dfff8b8885d4b9fee918c41c1b1f4b4027368b0355a16ea19a8: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:52:56.936175 containerd[2012]: time="2025-10-13T05:52:56.935142025Z" level=info msg="CreateContainer within sandbox \"07fdd6f8b3cb52514a29386798cd7aaa3ed07c4ebb91587b4d0c15d241dc4b32\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"bb96b44f39f95dfff8b8885d4b9fee918c41c1b1f4b4027368b0355a16ea19a8\"" Oct 13 05:52:56.936175 containerd[2012]: time="2025-10-13T05:52:56.936103518Z" level=info msg="StartContainer for \"bb96b44f39f95dfff8b8885d4b9fee918c41c1b1f4b4027368b0355a16ea19a8\"" Oct 13 05:52:56.938223 containerd[2012]: time="2025-10-13T05:52:56.938151433Z" level=info msg="connecting to shim bb96b44f39f95dfff8b8885d4b9fee918c41c1b1f4b4027368b0355a16ea19a8" address="unix:///run/containerd/s/347a41a53b1d10eb41968fa68bacfb665fca75d93f501ee08c2a17c46c3f3644" protocol=ttrpc version=3 Oct 13 05:52:56.968656 systemd[1]: Started cri-containerd-bb96b44f39f95dfff8b8885d4b9fee918c41c1b1f4b4027368b0355a16ea19a8.scope - libcontainer container bb96b44f39f95dfff8b8885d4b9fee918c41c1b1f4b4027368b0355a16ea19a8. Oct 13 05:52:57.065659 containerd[2012]: time="2025-10-13T05:52:57.065592090Z" level=info msg="StartContainer for \"bb96b44f39f95dfff8b8885d4b9fee918c41c1b1f4b4027368b0355a16ea19a8\" returns successfully" Oct 13 05:52:57.145244 kubelet[3314]: I1013 05:52:57.144648 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6987b98758-wc9wm" podStartSLOduration=34.156747948 podStartE2EDuration="44.108858194s" podCreationTimestamp="2025-10-13 05:52:13 +0000 UTC" firstStartedPulling="2025-10-13 05:52:46.810127794 +0000 UTC m=+54.686841977" lastFinishedPulling="2025-10-13 05:52:56.762238053 +0000 UTC m=+64.638952223" observedRunningTime="2025-10-13 05:52:57.097579039 +0000 UTC m=+64.974293232" watchObservedRunningTime="2025-10-13 05:52:57.108858194 +0000 UTC m=+64.985572383" Oct 13 05:52:57.227751 kubelet[3314]: I1013 05:52:57.227670 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-rf94f" podStartSLOduration=25.778904904 podStartE2EDuration="39.227645104s" podCreationTimestamp="2025-10-13 05:52:18 +0000 UTC" firstStartedPulling="2025-10-13 05:52:42.875663683 +0000 UTC m=+50.752377864" lastFinishedPulling="2025-10-13 05:52:56.324403894 +0000 UTC m=+64.201118064" observedRunningTime="2025-10-13 05:52:57.204970391 +0000 UTC m=+65.081684585" watchObservedRunningTime="2025-10-13 05:52:57.227645104 +0000 UTC m=+65.104359298" Oct 13 05:52:57.302940 containerd[2012]: time="2025-10-13T05:52:57.302763829Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223\" id:\"4871475bedaf80ad65308743139576dc91a22a90876e3abd4701383b7a7ac4b8\" pid:5838 exited_at:{seconds:1760334777 nanos:275961733}" Oct 13 05:52:58.096443 kubelet[3314]: I1013 05:52:58.096404 3314 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 05:52:58.175162 kubelet[3314]: I1013 05:52:58.175127 3314 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 05:52:58.561227 systemd[1]: Started sshd@9-172.31.26.230:22-139.178.68.195:57600.service - OpenSSH per-connection server daemon (139.178.68.195:57600). Oct 13 05:52:58.909965 sshd[5864]: Accepted publickey for core from 139.178.68.195 port 57600 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:52:58.925820 sshd-session[5864]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:52:58.941554 systemd-logind[1980]: New session 10 of user core. Oct 13 05:52:58.946081 systemd[1]: Started session-10.scope - Session 10 of User core. Oct 13 05:53:00.268259 sshd[5867]: Connection closed by 139.178.68.195 port 57600 Oct 13 05:53:00.267935 sshd-session[5864]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:00.279235 systemd[1]: sshd@9-172.31.26.230:22-139.178.68.195:57600.service: Deactivated successfully. Oct 13 05:53:00.288045 systemd[1]: session-10.scope: Deactivated successfully. Oct 13 05:53:00.295925 systemd-logind[1980]: Session 10 logged out. Waiting for processes to exit. Oct 13 05:53:00.331496 systemd[1]: Started sshd@10-172.31.26.230:22-139.178.68.195:57608.service - OpenSSH per-connection server daemon (139.178.68.195:57608). Oct 13 05:53:00.335790 systemd-logind[1980]: Removed session 10. Oct 13 05:53:00.577098 sshd[5884]: Accepted publickey for core from 139.178.68.195 port 57608 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:00.579400 sshd-session[5884]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:00.596622 systemd-logind[1980]: New session 11 of user core. Oct 13 05:53:00.602849 systemd[1]: Started session-11.scope - Session 11 of User core. Oct 13 05:53:01.510607 sshd[5887]: Connection closed by 139.178.68.195 port 57608 Oct 13 05:53:01.511791 sshd-session[5884]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:01.646706 systemd[1]: sshd@10-172.31.26.230:22-139.178.68.195:57608.service: Deactivated successfully. Oct 13 05:53:01.704435 systemd[1]: session-11.scope: Deactivated successfully. Oct 13 05:53:01.731107 systemd-logind[1980]: Session 11 logged out. Waiting for processes to exit. Oct 13 05:53:01.775753 systemd[1]: Started sshd@11-172.31.26.230:22-139.178.68.195:57624.service - OpenSSH per-connection server daemon (139.178.68.195:57624). Oct 13 05:53:01.806945 systemd-logind[1980]: Removed session 11. Oct 13 05:53:02.681601 sshd[5900]: Accepted publickey for core from 139.178.68.195 port 57624 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:02.699824 sshd-session[5900]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:02.812567 systemd-logind[1980]: New session 12 of user core. Oct 13 05:53:02.822069 systemd[1]: Started session-12.scope - Session 12 of User core. Oct 13 05:53:03.298488 sshd[5903]: Connection closed by 139.178.68.195 port 57624 Oct 13 05:53:03.301689 sshd-session[5900]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:03.316774 systemd[1]: sshd@11-172.31.26.230:22-139.178.68.195:57624.service: Deactivated successfully. Oct 13 05:53:03.318149 systemd-logind[1980]: Session 12 logged out. Waiting for processes to exit. Oct 13 05:53:03.325636 systemd[1]: session-12.scope: Deactivated successfully. Oct 13 05:53:03.330193 systemd-logind[1980]: Removed session 12. Oct 13 05:53:04.526049 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3023355503.mount: Deactivated successfully. Oct 13 05:53:05.532376 containerd[2012]: time="2025-10-13T05:53:05.532300733Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:53:05.541873 containerd[2012]: time="2025-10-13T05:53:05.541598001Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.3: active requests=0, bytes read=66357526" Oct 13 05:53:05.680557 containerd[2012]: time="2025-10-13T05:53:05.679430869Z" level=info msg="ImageCreate event name:\"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:53:05.855350 containerd[2012]: time="2025-10-13T05:53:05.855225209Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Oct 13 05:53:05.862630 containerd[2012]: time="2025-10-13T05:53:05.862570942Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" with image id \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.3\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:46297703ab3739331a00a58f0d6a5498c8d3b6523ad947eed68592ee0f3e79f0\", size \"66357372\" in 9.092094073s" Oct 13 05:53:05.862630 containerd[2012]: time="2025-10-13T05:53:05.862629230Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.3\" returns image reference \"sha256:a7d029fd8f6be94c26af980675c1650818e1e6e19dbd2f8c13e6e61963f021e8\"" Oct 13 05:53:06.236874 containerd[2012]: time="2025-10-13T05:53:06.236765789Z" level=info msg="CreateContainer within sandbox \"04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Oct 13 05:53:06.285873 containerd[2012]: time="2025-10-13T05:53:06.284479605Z" level=info msg="Container c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:53:06.380699 containerd[2012]: time="2025-10-13T05:53:06.380360043Z" level=info msg="CreateContainer within sandbox \"04abcbdffdce471036ef660cc8192e0d5a8a425d659f1abb99486e4fa60e0fa8\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\"" Oct 13 05:53:06.403220 containerd[2012]: time="2025-10-13T05:53:06.403155003Z" level=info msg="StartContainer for \"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\"" Oct 13 05:53:06.406471 containerd[2012]: time="2025-10-13T05:53:06.406406086Z" level=info msg="connecting to shim c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa" address="unix:///run/containerd/s/d103e206b8c5f184d6ddf75140e4e5c1067f6d00a5be9195ecdac1b2df72b23d" protocol=ttrpc version=3 Oct 13 05:53:06.580721 systemd[1]: Started cri-containerd-c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa.scope - libcontainer container c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa. Oct 13 05:53:06.711616 containerd[2012]: time="2025-10-13T05:53:06.711563547Z" level=info msg="StartContainer for \"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\" returns successfully" Oct 13 05:53:07.259576 kubelet[3314]: I1013 05:53:07.254800 3314 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-54d579b49d-6f9jm" podStartSLOduration=30.181429577 podStartE2EDuration="49.245257859s" podCreationTimestamp="2025-10-13 05:52:18 +0000 UTC" firstStartedPulling="2025-10-13 05:52:46.871959509 +0000 UTC m=+54.748673693" lastFinishedPulling="2025-10-13 05:53:05.935787738 +0000 UTC m=+73.812501975" observedRunningTime="2025-10-13 05:53:07.217356546 +0000 UTC m=+75.094070739" watchObservedRunningTime="2025-10-13 05:53:07.245257859 +0000 UTC m=+75.121972094" Oct 13 05:53:08.185641 containerd[2012]: time="2025-10-13T05:53:08.185164776Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\" id:\"c8e41903552a4b79e7b98967734c239e289d1cd2bec95d710f37f27d22273b04\" pid:5983 exit_status:1 exited_at:{seconds:1760334788 nanos:162465136}" Oct 13 05:53:08.351914 systemd[1]: Started sshd@12-172.31.26.230:22-139.178.68.195:47658.service - OpenSSH per-connection server daemon (139.178.68.195:47658). Oct 13 05:53:08.641432 sshd[6000]: Accepted publickey for core from 139.178.68.195 port 47658 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:08.647766 sshd-session[6000]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:08.655511 systemd-logind[1980]: New session 13 of user core. Oct 13 05:53:08.660773 systemd[1]: Started session-13.scope - Session 13 of User core. Oct 13 05:53:09.199364 containerd[2012]: time="2025-10-13T05:53:09.199308886Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\" id:\"ae93438c706578d2d8a4372a0d6b53dea9a6454f113529bda6cf08c68683b702\" pid:6019 exit_status:1 exited_at:{seconds:1760334789 nanos:198775539}" Oct 13 05:53:10.098674 sshd[6003]: Connection closed by 139.178.68.195 port 47658 Oct 13 05:53:10.099356 sshd-session[6000]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:10.104065 systemd[1]: sshd@12-172.31.26.230:22-139.178.68.195:47658.service: Deactivated successfully. Oct 13 05:53:10.106737 systemd[1]: session-13.scope: Deactivated successfully. Oct 13 05:53:10.108070 systemd-logind[1980]: Session 13 logged out. Waiting for processes to exit. Oct 13 05:53:10.109887 systemd-logind[1980]: Removed session 13. Oct 13 05:53:11.213848 containerd[2012]: time="2025-10-13T05:53:11.213685319Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\" id:\"16d67760607dcd5aa7e792fec6edd55341f56416cc81d5f739f7d4a00e1034b3\" pid:6048 exited_at:{seconds:1760334791 nanos:213275464}" Oct 13 05:53:15.136658 systemd[1]: Started sshd@13-172.31.26.230:22-139.178.68.195:47672.service - OpenSSH per-connection server daemon (139.178.68.195:47672). Oct 13 05:53:15.461521 sshd[6061]: Accepted publickey for core from 139.178.68.195 port 47672 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:15.466074 sshd-session[6061]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:15.472260 systemd-logind[1980]: New session 14 of user core. Oct 13 05:53:15.478691 systemd[1]: Started session-14.scope - Session 14 of User core. Oct 13 05:53:16.361004 sshd[6064]: Connection closed by 139.178.68.195 port 47672 Oct 13 05:53:16.362019 sshd-session[6061]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:16.366126 systemd[1]: sshd@13-172.31.26.230:22-139.178.68.195:47672.service: Deactivated successfully. Oct 13 05:53:16.369992 systemd[1]: session-14.scope: Deactivated successfully. Oct 13 05:53:16.372525 systemd-logind[1980]: Session 14 logged out. Waiting for processes to exit. Oct 13 05:53:16.374744 systemd-logind[1980]: Removed session 14. Oct 13 05:53:19.224775 kubelet[3314]: I1013 05:53:19.224714 3314 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 05:53:21.394423 systemd[1]: Started sshd@14-172.31.26.230:22-139.178.68.195:43090.service - OpenSSH per-connection server daemon (139.178.68.195:43090). Oct 13 05:53:21.689178 sshd[6079]: Accepted publickey for core from 139.178.68.195 port 43090 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:21.691092 sshd-session[6079]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:21.700264 systemd-logind[1980]: New session 15 of user core. Oct 13 05:53:21.708769 systemd[1]: Started session-15.scope - Session 15 of User core. Oct 13 05:53:22.469310 sshd[6093]: Connection closed by 139.178.68.195 port 43090 Oct 13 05:53:22.470279 sshd-session[6079]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:22.476569 systemd-logind[1980]: Session 15 logged out. Waiting for processes to exit. Oct 13 05:53:22.476854 systemd[1]: sshd@14-172.31.26.230:22-139.178.68.195:43090.service: Deactivated successfully. Oct 13 05:53:22.479609 systemd[1]: session-15.scope: Deactivated successfully. Oct 13 05:53:22.483709 systemd-logind[1980]: Removed session 15. Oct 13 05:53:22.506316 systemd[1]: Started sshd@15-172.31.26.230:22-139.178.68.195:43100.service - OpenSSH per-connection server daemon (139.178.68.195:43100). Oct 13 05:53:22.701105 sshd[6105]: Accepted publickey for core from 139.178.68.195 port 43100 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:22.703783 sshd-session[6105]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:22.711712 systemd-logind[1980]: New session 16 of user core. Oct 13 05:53:22.719716 systemd[1]: Started session-16.scope - Session 16 of User core. Oct 13 05:53:23.446081 sshd[6108]: Connection closed by 139.178.68.195 port 43100 Oct 13 05:53:23.451204 sshd-session[6105]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:23.460479 systemd[1]: sshd@15-172.31.26.230:22-139.178.68.195:43100.service: Deactivated successfully. Oct 13 05:53:23.463609 systemd[1]: session-16.scope: Deactivated successfully. Oct 13 05:53:23.465153 systemd-logind[1980]: Session 16 logged out. Waiting for processes to exit. Oct 13 05:53:23.468530 systemd-logind[1980]: Removed session 16. Oct 13 05:53:23.479743 systemd[1]: Started sshd@16-172.31.26.230:22-139.178.68.195:43108.service - OpenSSH per-connection server daemon (139.178.68.195:43108). Oct 13 05:53:23.708483 sshd[6117]: Accepted publickey for core from 139.178.68.195 port 43108 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:23.712237 sshd-session[6117]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:23.719968 systemd-logind[1980]: New session 17 of user core. Oct 13 05:53:23.725801 systemd[1]: Started session-17.scope - Session 17 of User core. Oct 13 05:53:24.859100 sshd[6120]: Connection closed by 139.178.68.195 port 43108 Oct 13 05:53:24.859889 sshd-session[6117]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:24.898708 containerd[2012]: time="2025-10-13T05:53:24.898651323Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223\" id:\"2014adfbb4a2d1a962fa405722d378fd7522eb38af5ef016210c84a2ff2f429e\" pid:6144 exited_at:{seconds:1760334804 nanos:843515350}" Oct 13 05:53:24.901623 systemd[1]: sshd@16-172.31.26.230:22-139.178.68.195:43108.service: Deactivated successfully. Oct 13 05:53:24.908327 systemd[1]: session-17.scope: Deactivated successfully. Oct 13 05:53:24.911642 systemd-logind[1980]: Session 17 logged out. Waiting for processes to exit. Oct 13 05:53:24.923109 systemd[1]: Started sshd@17-172.31.26.230:22-139.178.68.195:43116.service - OpenSSH per-connection server daemon (139.178.68.195:43116). Oct 13 05:53:24.949797 systemd-logind[1980]: Removed session 17. Oct 13 05:53:25.157395 sshd[6160]: Accepted publickey for core from 139.178.68.195 port 43116 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:25.159835 sshd-session[6160]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:25.165687 systemd-logind[1980]: New session 18 of user core. Oct 13 05:53:25.171659 systemd[1]: Started session-18.scope - Session 18 of User core. Oct 13 05:53:26.611905 sshd[6164]: Connection closed by 139.178.68.195 port 43116 Oct 13 05:53:26.616633 sshd-session[6160]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:26.630444 systemd[1]: sshd@17-172.31.26.230:22-139.178.68.195:43116.service: Deactivated successfully. Oct 13 05:53:26.630956 systemd-logind[1980]: Session 18 logged out. Waiting for processes to exit. Oct 13 05:53:26.637167 systemd[1]: session-18.scope: Deactivated successfully. Oct 13 05:53:26.653370 systemd-logind[1980]: Removed session 18. Oct 13 05:53:26.655735 systemd[1]: Started sshd@18-172.31.26.230:22-139.178.68.195:43120.service - OpenSSH per-connection server daemon (139.178.68.195:43120). Oct 13 05:53:26.923186 sshd[6176]: Accepted publickey for core from 139.178.68.195 port 43120 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:26.927938 sshd-session[6176]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:26.935109 systemd-logind[1980]: New session 19 of user core. Oct 13 05:53:26.940704 systemd[1]: Started session-19.scope - Session 19 of User core. Oct 13 05:53:27.147017 containerd[2012]: time="2025-10-13T05:53:27.146969353Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223\" id:\"e7c152aaac1089acdbacfd365cf178fcecbe70e2081f9823a41260dcdd28bf19\" pid:6199 exited_at:{seconds:1760334807 nanos:146318695}" Oct 13 05:53:27.263987 sshd[6179]: Connection closed by 139.178.68.195 port 43120 Oct 13 05:53:27.264356 sshd-session[6176]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:27.271039 systemd[1]: sshd@18-172.31.26.230:22-139.178.68.195:43120.service: Deactivated successfully. Oct 13 05:53:27.273824 systemd[1]: session-19.scope: Deactivated successfully. Oct 13 05:53:27.275746 systemd-logind[1980]: Session 19 logged out. Waiting for processes to exit. Oct 13 05:53:27.278101 systemd-logind[1980]: Removed session 19. Oct 13 05:53:31.971121 containerd[2012]: time="2025-10-13T05:53:31.971078835Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\" id:\"71fd0e26f2e1ca3d1b3c49b5e13af053270595307e52c7ae52f0c9406f45fd8f\" pid:6236 exited_at:{seconds:1760334811 nanos:970608311}" Oct 13 05:53:32.299931 systemd[1]: Started sshd@19-172.31.26.230:22-139.178.68.195:45742.service - OpenSSH per-connection server daemon (139.178.68.195:45742). Oct 13 05:53:32.565380 sshd[6249]: Accepted publickey for core from 139.178.68.195 port 45742 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:32.567768 sshd-session[6249]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:32.573871 systemd-logind[1980]: New session 20 of user core. Oct 13 05:53:32.576639 systemd[1]: Started session-20.scope - Session 20 of User core. Oct 13 05:53:33.314153 sshd[6252]: Connection closed by 139.178.68.195 port 45742 Oct 13 05:53:33.314667 sshd-session[6249]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:33.319010 systemd[1]: sshd@19-172.31.26.230:22-139.178.68.195:45742.service: Deactivated successfully. Oct 13 05:53:33.321200 systemd[1]: session-20.scope: Deactivated successfully. Oct 13 05:53:33.322286 systemd-logind[1980]: Session 20 logged out. Waiting for processes to exit. Oct 13 05:53:33.324427 systemd-logind[1980]: Removed session 20. Oct 13 05:53:38.352532 systemd[1]: Started sshd@20-172.31.26.230:22-139.178.68.195:49126.service - OpenSSH per-connection server daemon (139.178.68.195:49126). Oct 13 05:53:38.656942 sshd[6263]: Accepted publickey for core from 139.178.68.195 port 49126 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:38.665217 sshd-session[6263]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:38.674191 systemd-logind[1980]: New session 21 of user core. Oct 13 05:53:38.682873 systemd[1]: Started session-21.scope - Session 21 of User core. Oct 13 05:53:40.254866 sshd[6266]: Connection closed by 139.178.68.195 port 49126 Oct 13 05:53:40.257321 sshd-session[6263]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:40.267198 systemd[1]: sshd@20-172.31.26.230:22-139.178.68.195:49126.service: Deactivated successfully. Oct 13 05:53:40.269796 systemd-logind[1980]: Session 21 logged out. Waiting for processes to exit. Oct 13 05:53:40.278193 systemd[1]: session-21.scope: Deactivated successfully. Oct 13 05:53:40.286139 systemd-logind[1980]: Removed session 21. Oct 13 05:53:40.438279 containerd[2012]: time="2025-10-13T05:53:40.436636583Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\" id:\"b5737973757fc9a0564e90e319bc2256328d5c62dc11caf59104fb7f968e7b1b\" pid:6282 exited_at:{seconds:1760334820 nanos:349655451}" Oct 13 05:53:41.577209 containerd[2012]: time="2025-10-13T05:53:41.577154354Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\" id:\"67236ab4287d2c23dafdb9ee914345c3fa4e7895cef2f9f78842cd276c5cd357\" pid:6314 exited_at:{seconds:1760334821 nanos:576734204}" Oct 13 05:53:45.298111 systemd[1]: Started sshd@21-172.31.26.230:22-139.178.68.195:49134.service - OpenSSH per-connection server daemon (139.178.68.195:49134). Oct 13 05:53:45.601051 sshd[6329]: Accepted publickey for core from 139.178.68.195 port 49134 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:45.603407 sshd-session[6329]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:45.613186 systemd-logind[1980]: New session 22 of user core. Oct 13 05:53:45.621775 systemd[1]: Started session-22.scope - Session 22 of User core. Oct 13 05:53:46.570710 sshd[6332]: Connection closed by 139.178.68.195 port 49134 Oct 13 05:53:46.572724 sshd-session[6329]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:46.580689 systemd-logind[1980]: Session 22 logged out. Waiting for processes to exit. Oct 13 05:53:46.582416 systemd[1]: sshd@21-172.31.26.230:22-139.178.68.195:49134.service: Deactivated successfully. Oct 13 05:53:46.586776 systemd[1]: session-22.scope: Deactivated successfully. Oct 13 05:53:46.592015 systemd-logind[1980]: Removed session 22. Oct 13 05:53:51.611386 systemd[1]: Started sshd@22-172.31.26.230:22-139.178.68.195:37730.service - OpenSSH per-connection server daemon (139.178.68.195:37730). Oct 13 05:53:51.910740 sshd[6344]: Accepted publickey for core from 139.178.68.195 port 37730 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:51.915665 sshd-session[6344]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:51.928029 systemd-logind[1980]: New session 23 of user core. Oct 13 05:53:51.933807 systemd[1]: Started session-23.scope - Session 23 of User core. Oct 13 05:53:52.933486 sshd[6347]: Connection closed by 139.178.68.195 port 37730 Oct 13 05:53:52.931778 sshd-session[6344]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:52.945309 systemd[1]: sshd@22-172.31.26.230:22-139.178.68.195:37730.service: Deactivated successfully. Oct 13 05:53:52.952955 systemd[1]: session-23.scope: Deactivated successfully. Oct 13 05:53:52.960408 systemd-logind[1980]: Session 23 logged out. Waiting for processes to exit. Oct 13 05:53:52.968087 systemd-logind[1980]: Removed session 23. Oct 13 05:53:57.276002 containerd[2012]: time="2025-10-13T05:53:57.275942077Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5cd44722a8334d10cf471867a370d1b087c00b52b723d5658d22c1135d060223\" id:\"ed38a2ca494b59bacc999ae1a40cd6010135319cd7189601e7bb8897e3d32dcf\" pid:6371 exited_at:{seconds:1760334837 nanos:275597839}" Oct 13 05:53:57.968620 systemd[1]: Started sshd@23-172.31.26.230:22-139.178.68.195:39760.service - OpenSSH per-connection server daemon (139.178.68.195:39760). Oct 13 05:53:58.216867 sshd[6382]: Accepted publickey for core from 139.178.68.195 port 39760 ssh2: RSA SHA256:s1xe7fF9ictGJLsZ+yMvgojxUb4KGTBdg2jCJy6Gu0o Oct 13 05:53:58.220197 sshd-session[6382]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Oct 13 05:53:58.229807 systemd-logind[1980]: New session 24 of user core. Oct 13 05:53:58.233866 systemd[1]: Started session-24.scope - Session 24 of User core. Oct 13 05:53:59.146865 sshd[6385]: Connection closed by 139.178.68.195 port 39760 Oct 13 05:53:59.151863 sshd-session[6382]: pam_unix(sshd:session): session closed for user core Oct 13 05:53:59.168473 systemd[1]: sshd@23-172.31.26.230:22-139.178.68.195:39760.service: Deactivated successfully. Oct 13 05:53:59.174526 systemd[1]: session-24.scope: Deactivated successfully. Oct 13 05:53:59.176722 systemd-logind[1980]: Session 24 logged out. Waiting for processes to exit. Oct 13 05:53:59.182560 systemd-logind[1980]: Removed session 24. Oct 13 05:54:09.349087 containerd[2012]: time="2025-10-13T05:54:09.349033699Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c4051edf2b22ecd0cfbe63aca1dd65bf148898816c244e675921bf4cbd1c3faa\" id:\"6bf2fcbf86b6bb73740dec8ce7749cddea23f9a47ac1f6cdba4598418d6f7900\" pid:6418 exited_at:{seconds:1760334849 nanos:348681996}" Oct 13 05:54:11.263999 containerd[2012]: time="2025-10-13T05:54:11.263946484Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4b419cdb632b79bbd28e1e35ff1f87645f550b5f87a785df3e38218694c12ebb\" id:\"945a2698c9378d5c27887a385bc835ea1f40524f281d99bd82fcbd2f118c84ba\" pid:6441 exited_at:{seconds:1760334851 nanos:263576071}" Oct 13 05:54:12.926848 systemd[1]: cri-containerd-7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e.scope: Deactivated successfully. Oct 13 05:54:12.927875 systemd[1]: cri-containerd-7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e.scope: Consumed 13.474s CPU time, 101.5M memory peak, 84.6M read from disk. Oct 13 05:54:13.101090 containerd[2012]: time="2025-10-13T05:54:13.100860426Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e\" id:\"7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e\" pid:3637 exit_status:1 exited_at:{seconds:1760334853 nanos:59856733}" Oct 13 05:54:13.117254 containerd[2012]: time="2025-10-13T05:54:13.117178790Z" level=info msg="received exit event container_id:\"7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e\" id:\"7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e\" pid:3637 exit_status:1 exited_at:{seconds:1760334853 nanos:59856733}" Oct 13 05:54:13.298131 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e-rootfs.mount: Deactivated successfully. Oct 13 05:54:13.698321 systemd[1]: cri-containerd-2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd.scope: Deactivated successfully. Oct 13 05:54:13.699421 systemd[1]: cri-containerd-2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd.scope: Consumed 4.271s CPU time, 85.6M memory peak, 118.4M read from disk. Oct 13 05:54:13.704762 containerd[2012]: time="2025-10-13T05:54:13.704727729Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd\" id:\"2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd\" pid:3124 exit_status:1 exited_at:{seconds:1760334853 nanos:703677261}" Oct 13 05:54:13.705216 containerd[2012]: time="2025-10-13T05:54:13.705190413Z" level=info msg="received exit event container_id:\"2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd\" id:\"2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd\" pid:3124 exit_status:1 exited_at:{seconds:1760334853 nanos:703677261}" Oct 13 05:54:13.762254 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd-rootfs.mount: Deactivated successfully. Oct 13 05:54:13.924878 kubelet[3314]: I1013 05:54:13.924821 3314 scope.go:117] "RemoveContainer" containerID="7a0367d8506397bd46691ae96abfe7507c394dbeb91b6e66982e9a458a93782e" Oct 13 05:54:14.051843 containerd[2012]: time="2025-10-13T05:54:14.051702939Z" level=info msg="CreateContainer within sandbox \"884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Oct 13 05:54:14.206738 containerd[2012]: time="2025-10-13T05:54:14.206276426Z" level=info msg="Container f320396c1dd26bb5923d5907928cf243ce2c908434beb854e3107b6316709326: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:54:14.244241 containerd[2012]: time="2025-10-13T05:54:14.244176355Z" level=info msg="CreateContainer within sandbox \"884e5b12995c1a52e3c81b8cfc4a634be13a7e1bda8ed9a06a12db98231e9dfa\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"f320396c1dd26bb5923d5907928cf243ce2c908434beb854e3107b6316709326\"" Oct 13 05:54:14.250440 containerd[2012]: time="2025-10-13T05:54:14.250372779Z" level=info msg="StartContainer for \"f320396c1dd26bb5923d5907928cf243ce2c908434beb854e3107b6316709326\"" Oct 13 05:54:14.257681 containerd[2012]: time="2025-10-13T05:54:14.257596520Z" level=info msg="connecting to shim f320396c1dd26bb5923d5907928cf243ce2c908434beb854e3107b6316709326" address="unix:///run/containerd/s/4bf00479a77d1b0129eb2cdd185e3d2f5f530a71abce3f40d4f2ccf71dc003a3" protocol=ttrpc version=3 Oct 13 05:54:14.357879 systemd[1]: Started cri-containerd-f320396c1dd26bb5923d5907928cf243ce2c908434beb854e3107b6316709326.scope - libcontainer container f320396c1dd26bb5923d5907928cf243ce2c908434beb854e3107b6316709326. Oct 13 05:54:14.434515 containerd[2012]: time="2025-10-13T05:54:14.431848887Z" level=info msg="StartContainer for \"f320396c1dd26bb5923d5907928cf243ce2c908434beb854e3107b6316709326\" returns successfully" Oct 13 05:54:14.888550 kubelet[3314]: I1013 05:54:14.888254 3314 scope.go:117] "RemoveContainer" containerID="2b170f204c30eca19d1bbfa4e1461f7634a3444a4e89b00f25873b5ef3fac5bd" Oct 13 05:54:14.892065 containerd[2012]: time="2025-10-13T05:54:14.892016215Z" level=info msg="CreateContainer within sandbox \"4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Oct 13 05:54:14.930081 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1452240732.mount: Deactivated successfully. Oct 13 05:54:14.931654 containerd[2012]: time="2025-10-13T05:54:14.930071930Z" level=info msg="Container b0f8d5efb3ad156283ed1911342a52eae683c8db551a2eb1aa6390a3f437e23a: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:54:14.955759 containerd[2012]: time="2025-10-13T05:54:14.955713375Z" level=info msg="CreateContainer within sandbox \"4232c0794c8d8bc405c8ff3a55bc5acb370876a46545bb298542d227bc11a1ff\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"b0f8d5efb3ad156283ed1911342a52eae683c8db551a2eb1aa6390a3f437e23a\"" Oct 13 05:54:14.972165 kubelet[3314]: E1013 05:54:14.972105 3314 controller.go:195] "Failed to update lease" err="the server was unable to return a response in the time allotted, but may still be processing the request (put leases.coordination.k8s.io ip-172-31-26-230)" Oct 13 05:54:14.973963 containerd[2012]: time="2025-10-13T05:54:14.973922664Z" level=info msg="StartContainer for \"b0f8d5efb3ad156283ed1911342a52eae683c8db551a2eb1aa6390a3f437e23a\"" Oct 13 05:54:14.976954 containerd[2012]: time="2025-10-13T05:54:14.976913388Z" level=info msg="connecting to shim b0f8d5efb3ad156283ed1911342a52eae683c8db551a2eb1aa6390a3f437e23a" address="unix:///run/containerd/s/07b7bad936943a8db40cd17d2d6ebe98b9d13be2944471daffd130d1d6635b8c" protocol=ttrpc version=3 Oct 13 05:54:15.009807 systemd[1]: Started cri-containerd-b0f8d5efb3ad156283ed1911342a52eae683c8db551a2eb1aa6390a3f437e23a.scope - libcontainer container b0f8d5efb3ad156283ed1911342a52eae683c8db551a2eb1aa6390a3f437e23a. Oct 13 05:54:15.152613 containerd[2012]: time="2025-10-13T05:54:15.152025766Z" level=info msg="StartContainer for \"b0f8d5efb3ad156283ed1911342a52eae683c8db551a2eb1aa6390a3f437e23a\" returns successfully" Oct 13 05:54:18.861051 systemd[1]: cri-containerd-5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d.scope: Deactivated successfully. Oct 13 05:54:18.862692 systemd[1]: cri-containerd-5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d.scope: Consumed 2.667s CPU time, 41.3M memory peak, 73.5M read from disk. Oct 13 05:54:18.867751 containerd[2012]: time="2025-10-13T05:54:18.867707467Z" level=info msg="TaskExit event in podsandbox handler container_id:\"5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d\" id:\"5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d\" pid:3162 exit_status:1 exited_at:{seconds:1760334858 nanos:867242379}" Oct 13 05:54:18.868544 containerd[2012]: time="2025-10-13T05:54:18.867707564Z" level=info msg="received exit event container_id:\"5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d\" id:\"5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d\" pid:3162 exit_status:1 exited_at:{seconds:1760334858 nanos:867242379}" Oct 13 05:54:18.897871 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d-rootfs.mount: Deactivated successfully. Oct 13 05:54:19.912316 kubelet[3314]: I1013 05:54:19.912282 3314 scope.go:117] "RemoveContainer" containerID="5b4c96b501955e24e0ba972b2d3d77fa14ec5b70b27eeb268d7bdde09542197d" Oct 13 05:54:19.915477 containerd[2012]: time="2025-10-13T05:54:19.915103437Z" level=info msg="CreateContainer within sandbox \"9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Oct 13 05:54:19.932598 containerd[2012]: time="2025-10-13T05:54:19.932558908Z" level=info msg="Container 3d8650b7d711fb70dd661cb917a9b6e0e8510519046d8ac29682fbd70a635869: CDI devices from CRI Config.CDIDevices: []" Oct 13 05:54:19.950447 containerd[2012]: time="2025-10-13T05:54:19.950297795Z" level=info msg="CreateContainer within sandbox \"9d4425765210c2d5a8472aefcf054073abe9b87c3e32279408cf7519cf6dc937\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"3d8650b7d711fb70dd661cb917a9b6e0e8510519046d8ac29682fbd70a635869\"" Oct 13 05:54:19.951884 containerd[2012]: time="2025-10-13T05:54:19.951848191Z" level=info msg="StartContainer for \"3d8650b7d711fb70dd661cb917a9b6e0e8510519046d8ac29682fbd70a635869\"" Oct 13 05:54:19.953574 containerd[2012]: time="2025-10-13T05:54:19.953240708Z" level=info msg="connecting to shim 3d8650b7d711fb70dd661cb917a9b6e0e8510519046d8ac29682fbd70a635869" address="unix:///run/containerd/s/03b2d2f144bfd4fc661ba908dae25b5f7804d7746f11078cfcd2afd401f76d72" protocol=ttrpc version=3 Oct 13 05:54:19.986714 systemd[1]: Started cri-containerd-3d8650b7d711fb70dd661cb917a9b6e0e8510519046d8ac29682fbd70a635869.scope - libcontainer container 3d8650b7d711fb70dd661cb917a9b6e0e8510519046d8ac29682fbd70a635869. Oct 13 05:54:20.057900 containerd[2012]: time="2025-10-13T05:54:20.057858685Z" level=info msg="StartContainer for \"3d8650b7d711fb70dd661cb917a9b6e0e8510519046d8ac29682fbd70a635869\" returns successfully"