Aug 13 07:12:36.986343 kernel: Linux version 6.6.100-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Aug 12 22:14:58 -00 2025 Aug 13 07:12:36.986383 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:12:36.986404 kernel: BIOS-provided physical RAM map: Aug 13 07:12:36.986416 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Aug 13 07:12:36.986428 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable Aug 13 07:12:36.986440 kernel: BIOS-e820: [mem 0x00000000786ce000-0x00000000787cdfff] type 20 Aug 13 07:12:36.986455 kernel: BIOS-e820: [mem 0x00000000787ce000-0x000000007894dfff] reserved Aug 13 07:12:36.986468 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Aug 13 07:12:36.986481 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Aug 13 07:12:36.986497 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable Aug 13 07:12:36.986510 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Aug 13 07:12:36.986523 kernel: NX (Execute Disable) protection: active Aug 13 07:12:36.986536 kernel: APIC: Static calls initialized Aug 13 07:12:36.986549 kernel: efi: EFI v2.7 by EDK II Aug 13 07:12:36.986566 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77003518 Aug 13 07:12:36.986584 kernel: SMBIOS 2.7 present. Aug 13 07:12:36.986599 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Aug 13 07:12:36.986613 kernel: Hypervisor detected: KVM Aug 13 07:12:36.986627 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Aug 13 07:12:36.986641 kernel: kvm-clock: using sched offset of 3967099494 cycles Aug 13 07:12:36.986656 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Aug 13 07:12:36.986671 kernel: tsc: Detected 2499.998 MHz processor Aug 13 07:12:36.986686 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Aug 13 07:12:36.986701 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Aug 13 07:12:36.986716 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 Aug 13 07:12:36.986733 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Aug 13 07:12:36.986748 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Aug 13 07:12:36.986762 kernel: Using GB pages for direct mapping Aug 13 07:12:36.986775 kernel: Secure boot disabled Aug 13 07:12:36.986790 kernel: ACPI: Early table checksum verification disabled Aug 13 07:12:36.986804 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) Aug 13 07:12:36.986820 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) Aug 13 07:12:36.986835 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Aug 13 07:12:36.986850 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Aug 13 07:12:36.986868 kernel: ACPI: FACS 0x00000000789D0000 000040 Aug 13 07:12:36.986882 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Aug 13 07:12:36.986897 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Aug 13 07:12:36.986912 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Aug 13 07:12:36.986926 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Aug 13 07:12:36.986941 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Aug 13 07:12:36.986962 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Aug 13 07:12:36.986981 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Aug 13 07:12:36.986996 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) Aug 13 07:12:36.987012 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] Aug 13 07:12:36.987027 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] Aug 13 07:12:36.987043 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] Aug 13 07:12:36.987059 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] Aug 13 07:12:36.987074 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] Aug 13 07:12:36.987093 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] Aug 13 07:12:36.987129 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] Aug 13 07:12:36.987142 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] Aug 13 07:12:36.987153 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] Aug 13 07:12:36.987166 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] Aug 13 07:12:36.987178 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] Aug 13 07:12:36.987189 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Aug 13 07:12:36.987200 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Aug 13 07:12:36.987212 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Aug 13 07:12:36.987228 kernel: NUMA: Initialized distance table, cnt=1 Aug 13 07:12:36.987241 kernel: NODE_DATA(0) allocated [mem 0x7a8ef000-0x7a8f4fff] Aug 13 07:12:36.987256 kernel: Zone ranges: Aug 13 07:12:36.987271 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Aug 13 07:12:36.987285 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] Aug 13 07:12:36.987300 kernel: Normal empty Aug 13 07:12:36.987314 kernel: Movable zone start for each node Aug 13 07:12:36.987329 kernel: Early memory node ranges Aug 13 07:12:36.987343 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Aug 13 07:12:36.987361 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] Aug 13 07:12:36.987375 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] Aug 13 07:12:36.987390 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] Aug 13 07:12:36.987404 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Aug 13 07:12:36.987418 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Aug 13 07:12:36.987433 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Aug 13 07:12:36.987448 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges Aug 13 07:12:36.987462 kernel: ACPI: PM-Timer IO Port: 0xb008 Aug 13 07:12:36.987477 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Aug 13 07:12:36.987495 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Aug 13 07:12:36.987509 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Aug 13 07:12:36.987524 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Aug 13 07:12:36.987539 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Aug 13 07:12:36.987553 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Aug 13 07:12:36.987568 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Aug 13 07:12:36.987582 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Aug 13 07:12:36.987596 kernel: TSC deadline timer available Aug 13 07:12:36.987610 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Aug 13 07:12:36.987624 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Aug 13 07:12:36.987642 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices Aug 13 07:12:36.987656 kernel: Booting paravirtualized kernel on KVM Aug 13 07:12:36.987671 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Aug 13 07:12:36.987685 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Aug 13 07:12:36.987700 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u1048576 Aug 13 07:12:36.987714 kernel: pcpu-alloc: s197096 r8192 d32280 u1048576 alloc=1*2097152 Aug 13 07:12:36.987728 kernel: pcpu-alloc: [0] 0 1 Aug 13 07:12:36.987743 kernel: kvm-guest: PV spinlocks enabled Aug 13 07:12:36.987757 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Aug 13 07:12:36.987778 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:12:36.987793 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 07:12:36.987807 kernel: random: crng init done Aug 13 07:12:36.987820 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 07:12:36.987831 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Aug 13 07:12:36.987842 kernel: Fallback order for Node 0: 0 Aug 13 07:12:36.987854 kernel: Built 1 zonelists, mobility grouping on. Total pages: 501318 Aug 13 07:12:36.987870 kernel: Policy zone: DMA32 Aug 13 07:12:36.987882 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 07:12:36.987895 kernel: Memory: 1874604K/2037804K available (12288K kernel code, 2295K rwdata, 22748K rodata, 42876K init, 2316K bss, 162940K reserved, 0K cma-reserved) Aug 13 07:12:36.987907 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Aug 13 07:12:36.987919 kernel: Kernel/User page tables isolation: enabled Aug 13 07:12:36.987930 kernel: ftrace: allocating 37968 entries in 149 pages Aug 13 07:12:36.987943 kernel: ftrace: allocated 149 pages with 4 groups Aug 13 07:12:36.987954 kernel: Dynamic Preempt: voluntary Aug 13 07:12:36.987966 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 07:12:36.987989 kernel: rcu: RCU event tracing is enabled. Aug 13 07:12:36.988003 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Aug 13 07:12:36.988017 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 07:12:36.988030 kernel: Rude variant of Tasks RCU enabled. Aug 13 07:12:36.988043 kernel: Tracing variant of Tasks RCU enabled. Aug 13 07:12:36.988056 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 07:12:36.988067 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Aug 13 07:12:36.988081 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Aug 13 07:12:36.991333 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 07:12:36.991363 kernel: Console: colour dummy device 80x25 Aug 13 07:12:36.991384 kernel: printk: console [tty0] enabled Aug 13 07:12:36.991402 kernel: printk: console [ttyS0] enabled Aug 13 07:12:36.991426 kernel: ACPI: Core revision 20230628 Aug 13 07:12:36.991445 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Aug 13 07:12:36.991464 kernel: APIC: Switch to symmetric I/O mode setup Aug 13 07:12:36.991484 kernel: x2apic enabled Aug 13 07:12:36.991503 kernel: APIC: Switched APIC routing to: physical x2apic Aug 13 07:12:36.991523 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Aug 13 07:12:36.991542 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499998) Aug 13 07:12:36.991561 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Aug 13 07:12:36.991576 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Aug 13 07:12:36.991590 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Aug 13 07:12:36.991602 kernel: Spectre V2 : Mitigation: Retpolines Aug 13 07:12:36.991615 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Aug 13 07:12:36.991630 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Aug 13 07:12:36.991645 kernel: RETBleed: Vulnerable Aug 13 07:12:36.991659 kernel: Speculative Store Bypass: Vulnerable Aug 13 07:12:36.991677 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Aug 13 07:12:36.991692 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Aug 13 07:12:36.991705 kernel: GDS: Unknown: Dependent on hypervisor status Aug 13 07:12:36.991718 kernel: ITS: Mitigation: Aligned branch/return thunks Aug 13 07:12:36.991732 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Aug 13 07:12:36.991746 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Aug 13 07:12:36.991761 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Aug 13 07:12:36.991777 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Aug 13 07:12:36.991793 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Aug 13 07:12:36.991809 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Aug 13 07:12:36.991825 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Aug 13 07:12:36.991847 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Aug 13 07:12:36.991863 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Aug 13 07:12:36.991880 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Aug 13 07:12:36.991896 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Aug 13 07:12:36.991912 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Aug 13 07:12:36.991927 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Aug 13 07:12:36.991944 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Aug 13 07:12:36.991959 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Aug 13 07:12:36.991975 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Aug 13 07:12:36.991992 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Aug 13 07:12:36.992008 kernel: Freeing SMP alternatives memory: 32K Aug 13 07:12:36.992024 kernel: pid_max: default: 32768 minimum: 301 Aug 13 07:12:36.992048 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Aug 13 07:12:36.992064 kernel: landlock: Up and running. Aug 13 07:12:36.992078 kernel: SELinux: Initializing. Aug 13 07:12:36.997335 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Aug 13 07:12:36.997386 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Aug 13 07:12:36.997404 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8175M CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x4) Aug 13 07:12:36.997421 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 07:12:36.997439 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 07:12:36.997456 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 07:12:36.997474 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Aug 13 07:12:36.997500 kernel: signal: max sigframe size: 3632 Aug 13 07:12:36.997517 kernel: rcu: Hierarchical SRCU implementation. Aug 13 07:12:36.997535 kernel: rcu: Max phase no-delay instances is 400. Aug 13 07:12:36.997553 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Aug 13 07:12:36.997570 kernel: smp: Bringing up secondary CPUs ... Aug 13 07:12:36.997587 kernel: smpboot: x86: Booting SMP configuration: Aug 13 07:12:36.997605 kernel: .... node #0, CPUs: #1 Aug 13 07:12:36.997623 kernel: Transient Scheduler Attacks: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Aug 13 07:12:36.997642 kernel: Transient Scheduler Attacks: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Aug 13 07:12:36.997664 kernel: smp: Brought up 1 node, 2 CPUs Aug 13 07:12:36.997681 kernel: smpboot: Max logical packages: 1 Aug 13 07:12:36.997698 kernel: smpboot: Total of 2 processors activated (9999.99 BogoMIPS) Aug 13 07:12:36.997715 kernel: devtmpfs: initialized Aug 13 07:12:36.997732 kernel: x86/mm: Memory block size: 128MB Aug 13 07:12:36.997749 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) Aug 13 07:12:36.997856 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 07:12:36.997872 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Aug 13 07:12:36.997893 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 07:12:36.997909 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 07:12:36.997924 kernel: audit: initializing netlink subsys (disabled) Aug 13 07:12:36.997939 kernel: audit: type=2000 audit(1755069156.848:1): state=initialized audit_enabled=0 res=1 Aug 13 07:12:36.997955 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 07:12:36.997972 kernel: thermal_sys: Registered thermal governor 'user_space' Aug 13 07:12:36.997988 kernel: cpuidle: using governor menu Aug 13 07:12:36.998003 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 07:12:36.998018 kernel: dca service started, version 1.12.1 Aug 13 07:12:36.998037 kernel: PCI: Using configuration type 1 for base access Aug 13 07:12:36.998052 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Aug 13 07:12:36.998067 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 07:12:36.998082 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 07:12:36.998116 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 07:12:36.998182 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 07:12:36.998196 kernel: ACPI: Added _OSI(Module Device) Aug 13 07:12:36.998307 kernel: ACPI: Added _OSI(Processor Device) Aug 13 07:12:36.998409 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 07:12:36.998487 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Aug 13 07:12:36.998504 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Aug 13 07:12:36.998521 kernel: ACPI: Interpreter enabled Aug 13 07:12:36.998537 kernel: ACPI: PM: (supports S0 S5) Aug 13 07:12:36.998553 kernel: ACPI: Using IOAPIC for interrupt routing Aug 13 07:12:36.998567 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Aug 13 07:12:36.998582 kernel: PCI: Using E820 reservations for host bridge windows Aug 13 07:12:36.998596 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Aug 13 07:12:36.998671 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Aug 13 07:12:36.999007 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Aug 13 07:12:36.999207 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Aug 13 07:12:36.999353 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Aug 13 07:12:36.999374 kernel: acpiphp: Slot [3] registered Aug 13 07:12:36.999390 kernel: acpiphp: Slot [4] registered Aug 13 07:12:36.999406 kernel: acpiphp: Slot [5] registered Aug 13 07:12:36.999421 kernel: acpiphp: Slot [6] registered Aug 13 07:12:36.999443 kernel: acpiphp: Slot [7] registered Aug 13 07:12:36.999462 kernel: acpiphp: Slot [8] registered Aug 13 07:12:36.999475 kernel: acpiphp: Slot [9] registered Aug 13 07:12:36.999614 kernel: acpiphp: Slot [10] registered Aug 13 07:12:37.000282 kernel: acpiphp: Slot [11] registered Aug 13 07:12:37.000778 kernel: acpiphp: Slot [12] registered Aug 13 07:12:37.001457 kernel: acpiphp: Slot [13] registered Aug 13 07:12:37.001793 kernel: acpiphp: Slot [14] registered Aug 13 07:12:37.001819 kernel: acpiphp: Slot [15] registered Aug 13 07:12:37.001833 kernel: acpiphp: Slot [16] registered Aug 13 07:12:37.001852 kernel: acpiphp: Slot [17] registered Aug 13 07:12:37.001865 kernel: acpiphp: Slot [18] registered Aug 13 07:12:37.001878 kernel: acpiphp: Slot [19] registered Aug 13 07:12:37.001891 kernel: acpiphp: Slot [20] registered Aug 13 07:12:37.001904 kernel: acpiphp: Slot [21] registered Aug 13 07:12:37.001918 kernel: acpiphp: Slot [22] registered Aug 13 07:12:37.001930 kernel: acpiphp: Slot [23] registered Aug 13 07:12:37.001942 kernel: acpiphp: Slot [24] registered Aug 13 07:12:37.001955 kernel: acpiphp: Slot [25] registered Aug 13 07:12:37.001971 kernel: acpiphp: Slot [26] registered Aug 13 07:12:37.001985 kernel: acpiphp: Slot [27] registered Aug 13 07:12:37.001999 kernel: acpiphp: Slot [28] registered Aug 13 07:12:37.002015 kernel: acpiphp: Slot [29] registered Aug 13 07:12:37.002029 kernel: acpiphp: Slot [30] registered Aug 13 07:12:37.002043 kernel: acpiphp: Slot [31] registered Aug 13 07:12:37.002055 kernel: PCI host bridge to bus 0000:00 Aug 13 07:12:37.006386 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Aug 13 07:12:37.006663 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Aug 13 07:12:37.006789 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Aug 13 07:12:37.006902 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Aug 13 07:12:37.007017 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] Aug 13 07:12:37.007160 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Aug 13 07:12:37.007315 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Aug 13 07:12:37.007461 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Aug 13 07:12:37.007609 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 Aug 13 07:12:37.007744 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Aug 13 07:12:37.007887 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Aug 13 07:12:37.008043 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Aug 13 07:12:37.010226 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Aug 13 07:12:37.010392 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Aug 13 07:12:37.010535 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Aug 13 07:12:37.010680 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Aug 13 07:12:37.010826 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 Aug 13 07:12:37.010961 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x80000000-0x803fffff pref] Aug 13 07:12:37.011195 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Aug 13 07:12:37.011334 kernel: pci 0000:00:03.0: BAR 0: assigned to efifb Aug 13 07:12:37.011467 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Aug 13 07:12:37.011606 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Aug 13 07:12:37.011748 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80404000-0x80407fff] Aug 13 07:12:37.011888 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Aug 13 07:12:37.012020 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80400000-0x80403fff] Aug 13 07:12:37.012041 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Aug 13 07:12:37.012058 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Aug 13 07:12:37.012074 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Aug 13 07:12:37.012090 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Aug 13 07:12:37.012157 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Aug 13 07:12:37.012173 kernel: iommu: Default domain type: Translated Aug 13 07:12:37.012190 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Aug 13 07:12:37.012206 kernel: efivars: Registered efivars operations Aug 13 07:12:37.012222 kernel: PCI: Using ACPI for IRQ routing Aug 13 07:12:37.012238 kernel: PCI: pci_cache_line_size set to 64 bytes Aug 13 07:12:37.012255 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] Aug 13 07:12:37.012270 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] Aug 13 07:12:37.012404 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Aug 13 07:12:37.012541 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Aug 13 07:12:37.012671 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Aug 13 07:12:37.012691 kernel: vgaarb: loaded Aug 13 07:12:37.012707 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Aug 13 07:12:37.012724 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Aug 13 07:12:37.012740 kernel: clocksource: Switched to clocksource kvm-clock Aug 13 07:12:37.012756 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 07:12:37.012773 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 07:12:37.012789 kernel: pnp: PnP ACPI init Aug 13 07:12:37.012808 kernel: pnp: PnP ACPI: found 5 devices Aug 13 07:12:37.012825 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Aug 13 07:12:37.012841 kernel: NET: Registered PF_INET protocol family Aug 13 07:12:37.012857 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Aug 13 07:12:37.012873 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Aug 13 07:12:37.012890 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 07:12:37.012906 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Aug 13 07:12:37.012922 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Aug 13 07:12:37.012942 kernel: TCP: Hash tables configured (established 16384 bind 16384) Aug 13 07:12:37.012958 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Aug 13 07:12:37.012973 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Aug 13 07:12:37.012988 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 07:12:37.013044 kernel: NET: Registered PF_XDP protocol family Aug 13 07:12:37.013283 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Aug 13 07:12:37.013446 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Aug 13 07:12:37.013582 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Aug 13 07:12:37.013706 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Aug 13 07:12:37.013846 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] Aug 13 07:12:37.013995 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Aug 13 07:12:37.014017 kernel: PCI: CLS 0 bytes, default 64 Aug 13 07:12:37.014034 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Aug 13 07:12:37.014052 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x240937b9988, max_idle_ns: 440795218083 ns Aug 13 07:12:37.014069 kernel: clocksource: Switched to clocksource tsc Aug 13 07:12:37.014084 kernel: Initialise system trusted keyrings Aug 13 07:12:37.014128 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Aug 13 07:12:37.014147 kernel: Key type asymmetric registered Aug 13 07:12:37.014163 kernel: Asymmetric key parser 'x509' registered Aug 13 07:12:37.014179 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Aug 13 07:12:37.014195 kernel: io scheduler mq-deadline registered Aug 13 07:12:37.014212 kernel: io scheduler kyber registered Aug 13 07:12:37.014228 kernel: io scheduler bfq registered Aug 13 07:12:37.014244 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Aug 13 07:12:37.014260 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 07:12:37.014276 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Aug 13 07:12:37.014296 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Aug 13 07:12:37.014312 kernel: i8042: Warning: Keylock active Aug 13 07:12:37.014328 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Aug 13 07:12:37.014342 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Aug 13 07:12:37.014499 kernel: rtc_cmos 00:00: RTC can wake from S4 Aug 13 07:12:37.014625 kernel: rtc_cmos 00:00: registered as rtc0 Aug 13 07:12:37.014747 kernel: rtc_cmos 00:00: setting system clock to 2025-08-13T07:12:36 UTC (1755069156) Aug 13 07:12:37.014869 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Aug 13 07:12:37.014892 kernel: intel_pstate: CPU model not supported Aug 13 07:12:37.014908 kernel: efifb: probing for efifb Aug 13 07:12:37.014923 kernel: efifb: framebuffer at 0x80000000, using 1920k, total 1920k Aug 13 07:12:37.014939 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Aug 13 07:12:37.014954 kernel: efifb: scrolling: redraw Aug 13 07:12:37.014970 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Aug 13 07:12:37.014985 kernel: Console: switching to colour frame buffer device 100x37 Aug 13 07:12:37.015000 kernel: fb0: EFI VGA frame buffer device Aug 13 07:12:37.015016 kernel: pstore: Using crash dump compression: deflate Aug 13 07:12:37.015035 kernel: pstore: Registered efi_pstore as persistent store backend Aug 13 07:12:37.015049 kernel: NET: Registered PF_INET6 protocol family Aug 13 07:12:37.015065 kernel: Segment Routing with IPv6 Aug 13 07:12:37.015081 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 07:12:37.015143 kernel: NET: Registered PF_PACKET protocol family Aug 13 07:12:37.015158 kernel: Key type dns_resolver registered Aug 13 07:12:37.015171 kernel: IPI shorthand broadcast: enabled Aug 13 07:12:37.015208 kernel: sched_clock: Marking stable (521002066, 188358637)->(834373923, -125013220) Aug 13 07:12:37.015227 kernel: registered taskstats version 1 Aug 13 07:12:37.015245 kernel: Loading compiled-in X.509 certificates Aug 13 07:12:37.015262 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.100-flatcar: 264e720147fa8df9744bb9dc1c08171c0cb20041' Aug 13 07:12:37.015278 kernel: Key type .fscrypt registered Aug 13 07:12:37.015294 kernel: Key type fscrypt-provisioning registered Aug 13 07:12:37.015312 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 07:12:37.015327 kernel: ima: Allocated hash algorithm: sha1 Aug 13 07:12:37.015346 kernel: ima: No architecture policies found Aug 13 07:12:37.015361 kernel: clk: Disabling unused clocks Aug 13 07:12:37.015378 kernel: Freeing unused kernel image (initmem) memory: 42876K Aug 13 07:12:37.015398 kernel: Write protecting the kernel read-only data: 36864k Aug 13 07:12:37.015413 kernel: Freeing unused kernel image (rodata/data gap) memory: 1828K Aug 13 07:12:37.015428 kernel: Run /init as init process Aug 13 07:12:37.015450 kernel: with arguments: Aug 13 07:12:37.015466 kernel: /init Aug 13 07:12:37.015481 kernel: with environment: Aug 13 07:12:37.015495 kernel: HOME=/ Aug 13 07:12:37.015510 kernel: TERM=linux Aug 13 07:12:37.015523 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 07:12:37.015544 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 07:12:37.015563 systemd[1]: Detected virtualization amazon. Aug 13 07:12:37.015577 systemd[1]: Detected architecture x86-64. Aug 13 07:12:37.015592 systemd[1]: Running in initrd. Aug 13 07:12:37.015609 systemd[1]: No hostname configured, using default hostname. Aug 13 07:12:37.015624 systemd[1]: Hostname set to . Aug 13 07:12:37.015640 systemd[1]: Initializing machine ID from VM UUID. Aug 13 07:12:37.015659 systemd[1]: Queued start job for default target initrd.target. Aug 13 07:12:37.015674 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:12:37.016205 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:12:37.016225 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 07:12:37.016242 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:12:37.016262 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 07:12:37.017022 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 07:12:37.018485 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 07:12:37.018503 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 07:12:37.018519 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:12:37.018535 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:12:37.018552 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:12:37.018573 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:12:37.018588 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:12:37.018621 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:12:37.018638 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:12:37.018653 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:12:37.018667 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 07:12:37.018684 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Aug 13 07:12:37.018701 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:12:37.018726 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:12:37.018741 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:12:37.018756 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:12:37.018772 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 07:12:37.018787 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:12:37.018806 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 07:12:37.018824 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 07:12:37.018842 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:12:37.018860 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:12:37.018883 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:12:37.018938 systemd-journald[178]: Collecting audit messages is disabled. Aug 13 07:12:37.018978 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 07:12:37.018994 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:12:37.019015 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 07:12:37.019034 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Aug 13 07:12:37.019053 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 07:12:37.019071 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:12:37.019093 systemd-journald[178]: Journal started Aug 13 07:12:37.021178 systemd-journald[178]: Runtime Journal (/run/log/journal/ec2c78c10ae2aab33d5047134f04b0aa) is 4.7M, max 38.2M, 33.4M free. Aug 13 07:12:37.013501 systemd-modules-load[179]: Inserted module 'overlay' Aug 13 07:12:37.027136 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:12:37.030378 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 07:12:37.045502 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:12:37.050310 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:12:37.068158 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 07:12:37.064655 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:12:37.077959 systemd-modules-load[179]: Inserted module 'br_netfilter' Aug 13 07:12:37.078755 kernel: Bridge firewalling registered Aug 13 07:12:37.082431 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:12:37.084895 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:12:37.086020 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:12:37.094353 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 07:12:37.102317 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:12:37.105520 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:12:37.117431 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:12:37.119939 dracut-cmdline[207]: dracut-dracut-053 Aug 13 07:12:37.124912 dracut-cmdline[207]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=8b1c4c6202e70eaa8c6477427259ab5e403c8f1de8515605304942a21d23450a Aug 13 07:12:37.127358 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:12:37.173307 systemd-resolved[220]: Positive Trust Anchors: Aug 13 07:12:37.173326 systemd-resolved[220]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:12:37.173392 systemd-resolved[220]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:12:37.181900 systemd-resolved[220]: Defaulting to hostname 'linux'. Aug 13 07:12:37.185924 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:12:37.186663 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:12:37.220150 kernel: SCSI subsystem initialized Aug 13 07:12:37.231136 kernel: Loading iSCSI transport class v2.0-870. Aug 13 07:12:37.243141 kernel: iscsi: registered transport (tcp) Aug 13 07:12:37.267069 kernel: iscsi: registered transport (qla4xxx) Aug 13 07:12:37.267335 kernel: QLogic iSCSI HBA Driver Aug 13 07:12:37.308711 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 07:12:37.315388 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 07:12:37.344043 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 07:12:37.344144 kernel: device-mapper: uevent: version 1.0.3 Aug 13 07:12:37.344168 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Aug 13 07:12:37.389143 kernel: raid6: avx512x4 gen() 17399 MB/s Aug 13 07:12:37.407144 kernel: raid6: avx512x2 gen() 17163 MB/s Aug 13 07:12:37.425135 kernel: raid6: avx512x1 gen() 17143 MB/s Aug 13 07:12:37.443130 kernel: raid6: avx2x4 gen() 17321 MB/s Aug 13 07:12:37.461138 kernel: raid6: avx2x2 gen() 17195 MB/s Aug 13 07:12:37.480160 kernel: raid6: avx2x1 gen() 13064 MB/s Aug 13 07:12:37.480217 kernel: raid6: using algorithm avx512x4 gen() 17399 MB/s Aug 13 07:12:37.500245 kernel: raid6: .... xor() 7160 MB/s, rmw enabled Aug 13 07:12:37.500307 kernel: raid6: using avx512x2 recovery algorithm Aug 13 07:12:37.524151 kernel: xor: automatically using best checksumming function avx Aug 13 07:12:37.688142 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 07:12:37.699362 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:12:37.705352 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:12:37.727843 systemd-udevd[396]: Using default interface naming scheme 'v255'. Aug 13 07:12:37.733200 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:12:37.743698 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 07:12:37.762328 dracut-pre-trigger[401]: rd.md=0: removing MD RAID activation Aug 13 07:12:37.795375 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:12:37.800301 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:12:37.853596 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:12:37.862407 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 07:12:37.888646 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 07:12:37.891659 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:12:37.893286 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:12:37.893784 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:12:37.901349 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 07:12:37.935058 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:12:37.960144 kernel: ena 0000:00:05.0: ENA device version: 0.10 Aug 13 07:12:37.960441 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Aug 13 07:12:37.966138 kernel: cryptd: max_cpu_qlen set to 1000 Aug 13 07:12:37.979467 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:12:37.979636 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:12:37.983879 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:12:37.988583 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Aug 13 07:12:37.984490 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:12:37.989284 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:12:37.990263 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:12:38.004417 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:12:38.008604 kernel: nvme nvme0: pci function 0000:00:04.0 Aug 13 07:12:38.008853 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Aug 13 07:12:38.019779 kernel: AVX2 version of gcm_enc/dec engaged. Aug 13 07:12:38.019859 kernel: AES CTR mode by8 optimization enabled Aug 13 07:12:38.029208 kernel: nvme nvme0: 2/0/0 default/read/poll queues Aug 13 07:12:38.048834 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 07:12:38.048902 kernel: GPT:9289727 != 16777215 Aug 13 07:12:38.048922 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 07:12:38.051470 kernel: GPT:9289727 != 16777215 Aug 13 07:12:38.052467 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 07:12:38.061132 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 07:12:38.061200 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:3e:37:5b:65:f1 Aug 13 07:12:38.070857 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:12:38.072348 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:12:38.074687 (udev-worker)[456]: Network interface NamePolicy= disabled on kernel command line. Aug 13 07:12:38.082410 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:12:38.103954 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:12:38.109279 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:12:38.128394 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:12:38.139137 kernel: BTRFS: device label OEM devid 1 transid 9 /dev/nvme0n1p6 scanned by (udev-worker) (442) Aug 13 07:12:38.173925 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Aug 13 07:12:38.181129 kernel: BTRFS: device fsid 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad devid 1 transid 37 /dev/nvme0n1p3 scanned by (udev-worker) (449) Aug 13 07:12:38.211836 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Aug 13 07:12:38.223845 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Aug 13 07:12:38.230316 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Aug 13 07:12:38.231200 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Aug 13 07:12:38.243390 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 07:12:38.251844 disk-uuid[631]: Primary Header is updated. Aug 13 07:12:38.251844 disk-uuid[631]: Secondary Entries is updated. Aug 13 07:12:38.251844 disk-uuid[631]: Secondary Header is updated. Aug 13 07:12:38.258130 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 07:12:38.264143 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 07:12:38.268177 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 07:12:39.275765 disk-uuid[632]: The operation has completed successfully. Aug 13 07:12:39.276752 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 07:12:39.420091 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 07:12:39.420247 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 07:12:39.447363 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 07:12:39.451495 sh[975]: Success Aug 13 07:12:39.473219 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Aug 13 07:12:39.568695 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 07:12:39.577262 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 07:12:39.578393 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 07:12:39.605131 kernel: BTRFS info (device dm-0): first mount of filesystem 6f4baebc-7e60-4ee7-93a9-8bedb08a33ad Aug 13 07:12:39.605203 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:12:39.605223 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Aug 13 07:12:39.607635 kernel: BTRFS info (device dm-0): disabling log replay at mount time Aug 13 07:12:39.611197 kernel: BTRFS info (device dm-0): using free space tree Aug 13 07:12:39.714125 kernel: BTRFS info (device dm-0): enabling ssd optimizations Aug 13 07:12:39.728047 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 07:12:39.729401 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 07:12:39.740405 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 07:12:39.745502 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 07:12:39.771022 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:12:39.773781 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:12:39.774460 kernel: BTRFS info (device nvme0n1p6): using free space tree Aug 13 07:12:39.782210 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Aug 13 07:12:39.794418 systemd[1]: mnt-oem.mount: Deactivated successfully. Aug 13 07:12:39.798414 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:12:39.806324 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 07:12:39.814566 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 07:12:39.865790 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:12:39.872326 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:12:39.902230 systemd-networkd[1167]: lo: Link UP Aug 13 07:12:39.902248 systemd-networkd[1167]: lo: Gained carrier Aug 13 07:12:39.903988 systemd-networkd[1167]: Enumeration completed Aug 13 07:12:39.904522 systemd-networkd[1167]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:12:39.904527 systemd-networkd[1167]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:12:39.905469 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:12:39.907749 systemd-networkd[1167]: eth0: Link UP Aug 13 07:12:39.907754 systemd-networkd[1167]: eth0: Gained carrier Aug 13 07:12:39.907769 systemd-networkd[1167]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:12:39.908355 systemd[1]: Reached target network.target - Network. Aug 13 07:12:39.915221 systemd-networkd[1167]: eth0: DHCPv4 address 172.31.16.138/20, gateway 172.31.16.1 acquired from 172.31.16.1 Aug 13 07:12:40.190644 ignition[1097]: Ignition 2.19.0 Aug 13 07:12:40.190656 ignition[1097]: Stage: fetch-offline Aug 13 07:12:40.190852 ignition[1097]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:12:40.192288 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:12:40.190861 ignition[1097]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 07:12:40.191217 ignition[1097]: Ignition finished successfully Aug 13 07:12:40.197282 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Aug 13 07:12:40.212624 ignition[1176]: Ignition 2.19.0 Aug 13 07:12:40.213375 ignition[1176]: Stage: fetch Aug 13 07:12:40.213891 ignition[1176]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:12:40.213902 ignition[1176]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 07:12:40.213998 ignition[1176]: PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 07:12:40.223554 ignition[1176]: PUT result: OK Aug 13 07:12:40.225150 ignition[1176]: parsed url from cmdline: "" Aug 13 07:12:40.225161 ignition[1176]: no config URL provided Aug 13 07:12:40.225168 ignition[1176]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 07:12:40.225180 ignition[1176]: no config at "/usr/lib/ignition/user.ign" Aug 13 07:12:40.225198 ignition[1176]: PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 07:12:40.225865 ignition[1176]: PUT result: OK Aug 13 07:12:40.225924 ignition[1176]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Aug 13 07:12:40.226486 ignition[1176]: GET result: OK Aug 13 07:12:40.226933 ignition[1176]: parsing config with SHA512: b89a8c431d82e87c652a533de9c48bcfc0c37b21c21e99b745770d0623e0368ae28720a7e1a881b4938cf4be15344ad1fe23eafa39acf48801bef72630f14c7e Aug 13 07:12:40.230711 unknown[1176]: fetched base config from "system" Aug 13 07:12:40.230725 unknown[1176]: fetched base config from "system" Aug 13 07:12:40.230731 unknown[1176]: fetched user config from "aws" Aug 13 07:12:40.232710 ignition[1176]: fetch: fetch complete Aug 13 07:12:40.232725 ignition[1176]: fetch: fetch passed Aug 13 07:12:40.232779 ignition[1176]: Ignition finished successfully Aug 13 07:12:40.234662 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Aug 13 07:12:40.239400 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 07:12:40.260076 ignition[1182]: Ignition 2.19.0 Aug 13 07:12:40.260093 ignition[1182]: Stage: kargs Aug 13 07:12:40.260618 ignition[1182]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:12:40.260632 ignition[1182]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 07:12:40.260751 ignition[1182]: PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 07:12:40.261825 ignition[1182]: PUT result: OK Aug 13 07:12:40.267984 ignition[1182]: kargs: kargs passed Aug 13 07:12:40.268137 ignition[1182]: Ignition finished successfully Aug 13 07:12:40.270252 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 07:12:40.275353 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 07:12:40.293320 ignition[1188]: Ignition 2.19.0 Aug 13 07:12:40.293341 ignition[1188]: Stage: disks Aug 13 07:12:40.293954 ignition[1188]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:12:40.293969 ignition[1188]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 07:12:40.294126 ignition[1188]: PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 07:12:40.295029 ignition[1188]: PUT result: OK Aug 13 07:12:40.298066 ignition[1188]: disks: disks passed Aug 13 07:12:40.298176 ignition[1188]: Ignition finished successfully Aug 13 07:12:40.299788 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 07:12:40.300855 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 07:12:40.301551 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 07:12:40.302063 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:12:40.302675 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:12:40.303306 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:12:40.308346 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 07:12:40.339908 systemd-fsck[1196]: ROOT: clean, 14/553520 files, 52654/553472 blocks Aug 13 07:12:40.343899 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 07:12:40.351598 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 07:12:40.470145 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 98cc0201-e9ec-4d2c-8a62-5b521bf9317d r/w with ordered data mode. Quota mode: none. Aug 13 07:12:40.470335 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 07:12:40.471326 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 07:12:40.477263 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:12:40.481239 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 07:12:40.483079 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Aug 13 07:12:40.484244 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 07:12:40.484282 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:12:40.490984 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 07:12:40.497336 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 07:12:40.509130 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/nvme0n1p6 scanned by mount (1215) Aug 13 07:12:40.513253 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:12:40.513314 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:12:40.515419 kernel: BTRFS info (device nvme0n1p6): using free space tree Aug 13 07:12:40.524415 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Aug 13 07:12:40.526705 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:12:40.843563 initrd-setup-root[1239]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 07:12:40.870866 initrd-setup-root[1246]: cut: /sysroot/etc/group: No such file or directory Aug 13 07:12:40.875402 initrd-setup-root[1253]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 07:12:40.902185 initrd-setup-root[1260]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 07:12:41.193225 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 07:12:41.199250 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 07:12:41.201640 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 07:12:41.215301 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 07:12:41.216169 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:12:41.246885 ignition[1328]: INFO : Ignition 2.19.0 Aug 13 07:12:41.246885 ignition[1328]: INFO : Stage: mount Aug 13 07:12:41.250011 ignition[1328]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:12:41.250011 ignition[1328]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 07:12:41.250011 ignition[1328]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 07:12:41.253806 ignition[1328]: INFO : PUT result: OK Aug 13 07:12:41.255173 ignition[1328]: INFO : mount: mount passed Aug 13 07:12:41.255173 ignition[1328]: INFO : Ignition finished successfully Aug 13 07:12:41.256459 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 07:12:41.263237 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 07:12:41.267466 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 07:12:41.283418 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:12:41.305147 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme0n1p6 scanned by mount (1340) Aug 13 07:12:41.305211 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 7cc37ed4-8461-447f-bee4-dfe5b4695079 Aug 13 07:12:41.309832 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Aug 13 07:12:41.309898 kernel: BTRFS info (device nvme0n1p6): using free space tree Aug 13 07:12:41.317119 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Aug 13 07:12:41.319138 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:12:41.354456 ignition[1356]: INFO : Ignition 2.19.0 Aug 13 07:12:41.354456 ignition[1356]: INFO : Stage: files Aug 13 07:12:41.355936 ignition[1356]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:12:41.355936 ignition[1356]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 07:12:41.355936 ignition[1356]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 07:12:41.357325 ignition[1356]: INFO : PUT result: OK Aug 13 07:12:41.359168 ignition[1356]: DEBUG : files: compiled without relabeling support, skipping Aug 13 07:12:41.372714 ignition[1356]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 07:12:41.372714 ignition[1356]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 07:12:41.399906 ignition[1356]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 07:12:41.400865 ignition[1356]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 07:12:41.400865 ignition[1356]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 07:12:41.400450 unknown[1356]: wrote ssh authorized keys file for user: core Aug 13 07:12:41.403745 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Aug 13 07:12:41.403745 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Aug 13 07:12:41.448523 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Aug 13 07:12:41.510265 systemd-networkd[1167]: eth0: Gained IPv6LL Aug 13 07:12:41.631687 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Aug 13 07:12:41.631687 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Aug 13 07:12:41.633569 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 07:12:41.633569 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:12:41.633569 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:12:41.633569 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:12:41.633569 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:12:41.633569 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:12:41.633569 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:12:41.638613 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:12:41.638613 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:12:41.638613 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:12:41.638613 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:12:41.638613 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:12:41.638613 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Aug 13 07:12:42.038998 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Aug 13 07:12:42.464988 ignition[1356]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Aug 13 07:12:42.464988 ignition[1356]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Aug 13 07:12:42.467826 ignition[1356]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:12:42.469305 ignition[1356]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:12:42.469305 ignition[1356]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Aug 13 07:12:42.469305 ignition[1356]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Aug 13 07:12:42.469305 ignition[1356]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 07:12:42.469305 ignition[1356]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:12:42.469305 ignition[1356]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:12:42.469305 ignition[1356]: INFO : files: files passed Aug 13 07:12:42.469305 ignition[1356]: INFO : Ignition finished successfully Aug 13 07:12:42.469971 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 07:12:42.475388 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 07:12:42.476679 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 07:12:42.481408 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 07:12:42.481515 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 07:12:42.505332 initrd-setup-root-after-ignition[1386]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:12:42.507435 initrd-setup-root-after-ignition[1386]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:12:42.509601 initrd-setup-root-after-ignition[1390]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:12:42.510812 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:12:42.511921 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 07:12:42.516387 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 07:12:42.548959 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 07:12:42.549165 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 07:12:42.550504 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 07:12:42.551664 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 07:12:42.552491 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 07:12:42.557442 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 07:12:42.572464 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:12:42.579326 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 07:12:42.593325 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:12:42.594228 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:12:42.595278 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 07:12:42.596247 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 07:12:42.596439 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:12:42.597842 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 07:12:42.598754 systemd[1]: Stopped target basic.target - Basic System. Aug 13 07:12:42.599591 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 07:12:42.600430 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:12:42.601245 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 07:12:42.602154 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 07:12:42.602938 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:12:42.603792 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 07:12:42.605021 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 07:12:42.605950 systemd[1]: Stopped target swap.target - Swaps. Aug 13 07:12:42.606721 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 07:12:42.606903 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:12:42.608035 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:12:42.608881 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:12:42.609581 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 07:12:42.610477 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:12:42.611071 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 07:12:42.611270 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 07:12:42.612798 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 07:12:42.612984 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:12:42.613804 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 07:12:42.613962 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 07:12:42.621442 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 07:12:42.623212 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 07:12:42.624261 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:12:42.629610 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 07:12:42.632228 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 07:12:42.632902 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:12:42.636800 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 07:12:42.638798 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:12:42.643886 ignition[1410]: INFO : Ignition 2.19.0 Aug 13 07:12:42.643886 ignition[1410]: INFO : Stage: umount Aug 13 07:12:42.645386 ignition[1410]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:12:42.645386 ignition[1410]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Aug 13 07:12:42.645386 ignition[1410]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Aug 13 07:12:42.648186 ignition[1410]: INFO : PUT result: OK Aug 13 07:12:42.649621 ignition[1410]: INFO : umount: umount passed Aug 13 07:12:42.650970 ignition[1410]: INFO : Ignition finished successfully Aug 13 07:12:42.651635 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 07:12:42.651758 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 07:12:42.654456 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 07:12:42.654597 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 07:12:42.656947 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 07:12:42.657026 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 07:12:42.659735 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 07:12:42.659812 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 07:12:42.660219 systemd[1]: ignition-fetch.service: Deactivated successfully. Aug 13 07:12:42.660275 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Aug 13 07:12:42.661211 systemd[1]: Stopped target network.target - Network. Aug 13 07:12:42.663094 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 07:12:42.663234 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:12:42.665534 systemd[1]: Stopped target paths.target - Path Units. Aug 13 07:12:42.666073 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 07:12:42.666187 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:12:42.666667 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 07:12:42.667123 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 07:12:42.667607 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 07:12:42.667665 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:12:42.670209 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 07:12:42.670270 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:12:42.671120 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 07:12:42.671194 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 07:12:42.671742 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 07:12:42.671803 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 07:12:42.673429 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 07:12:42.674147 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 07:12:42.676451 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 07:12:42.679174 systemd-networkd[1167]: eth0: DHCPv6 lease lost Aug 13 07:12:42.681631 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 07:12:42.681819 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 07:12:42.683823 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 07:12:42.683875 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:12:42.694300 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 07:12:42.694937 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 07:12:42.695027 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:12:42.696046 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:12:42.698741 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 07:12:42.698879 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 07:12:42.703540 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 07:12:42.703693 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 07:12:42.710833 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 07:12:42.711283 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:12:42.716713 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 07:12:42.716882 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 07:12:42.718836 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 07:12:42.718910 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 07:12:42.719915 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 07:12:42.719962 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:12:42.720752 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 07:12:42.720818 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:12:42.722084 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 07:12:42.722171 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 07:12:42.723324 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:12:42.723387 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:12:42.724533 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 07:12:42.724592 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 07:12:42.731312 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 07:12:42.732894 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 07:12:42.733807 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:12:42.734495 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 07:12:42.734560 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 07:12:42.736349 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 07:12:42.736424 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:12:42.737620 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 07:12:42.737697 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:12:42.738318 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:12:42.738379 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:12:42.740912 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 07:12:42.741050 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 07:12:42.742428 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 07:12:42.751428 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 07:12:42.763711 systemd[1]: Switching root. Aug 13 07:12:42.796080 systemd-journald[178]: Journal stopped Aug 13 07:12:44.800131 systemd-journald[178]: Received SIGTERM from PID 1 (systemd). Aug 13 07:12:44.800230 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 07:12:44.800254 kernel: SELinux: policy capability open_perms=1 Aug 13 07:12:44.800274 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 07:12:44.800294 kernel: SELinux: policy capability always_check_network=0 Aug 13 07:12:44.800329 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 07:12:44.800356 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 07:12:44.800376 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 07:12:44.800396 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 07:12:44.800417 kernel: audit: type=1403 audit(1755069163.304:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Aug 13 07:12:44.800439 systemd[1]: Successfully loaded SELinux policy in 80.592ms. Aug 13 07:12:44.800467 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.684ms. Aug 13 07:12:44.800494 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 07:12:44.800519 systemd[1]: Detected virtualization amazon. Aug 13 07:12:44.800540 systemd[1]: Detected architecture x86-64. Aug 13 07:12:44.800563 systemd[1]: Detected first boot. Aug 13 07:12:44.800585 systemd[1]: Initializing machine ID from VM UUID. Aug 13 07:12:44.800607 zram_generator::config[1453]: No configuration found. Aug 13 07:12:44.800639 systemd[1]: Populated /etc with preset unit settings. Aug 13 07:12:44.800660 systemd[1]: initrd-switch-root.service: Deactivated successfully. Aug 13 07:12:44.800680 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Aug 13 07:12:44.800701 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Aug 13 07:12:44.800727 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Aug 13 07:12:44.800750 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Aug 13 07:12:44.800772 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Aug 13 07:12:44.800794 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Aug 13 07:12:44.800815 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Aug 13 07:12:44.800837 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Aug 13 07:12:44.800858 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Aug 13 07:12:44.800879 systemd[1]: Created slice user.slice - User and Session Slice. Aug 13 07:12:44.800908 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:12:44.800931 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:12:44.800952 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Aug 13 07:12:44.800973 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Aug 13 07:12:44.800995 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Aug 13 07:12:44.801017 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:12:44.801043 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Aug 13 07:12:44.801064 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:12:44.801086 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Aug 13 07:12:44.805193 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Aug 13 07:12:44.805228 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Aug 13 07:12:44.805250 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Aug 13 07:12:44.805272 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:12:44.805293 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:12:44.805314 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:12:44.805336 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:12:44.805357 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Aug 13 07:12:44.805383 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Aug 13 07:12:44.805404 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:12:44.805425 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:12:44.805447 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:12:44.805468 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Aug 13 07:12:44.805491 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Aug 13 07:12:44.805511 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Aug 13 07:12:44.805533 systemd[1]: Mounting media.mount - External Media Directory... Aug 13 07:12:44.805555 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:44.805581 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Aug 13 07:12:44.805613 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Aug 13 07:12:44.806881 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Aug 13 07:12:44.807221 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 07:12:44.807251 systemd[1]: Reached target machines.target - Containers. Aug 13 07:12:44.807275 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Aug 13 07:12:44.807296 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:12:44.807318 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:12:44.807345 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Aug 13 07:12:44.807367 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:12:44.807389 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:12:44.807411 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:12:44.807433 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Aug 13 07:12:44.807455 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:12:44.807477 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 07:12:44.807499 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Aug 13 07:12:44.807521 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Aug 13 07:12:44.807547 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Aug 13 07:12:44.807569 systemd[1]: Stopped systemd-fsck-usr.service. Aug 13 07:12:44.807590 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:12:44.807612 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:12:44.807634 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Aug 13 07:12:44.807656 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Aug 13 07:12:44.807677 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:12:44.807699 systemd[1]: verity-setup.service: Deactivated successfully. Aug 13 07:12:44.807722 systemd[1]: Stopped verity-setup.service. Aug 13 07:12:44.807748 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:44.807769 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Aug 13 07:12:44.807791 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Aug 13 07:12:44.807813 systemd[1]: Mounted media.mount - External Media Directory. Aug 13 07:12:44.807835 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Aug 13 07:12:44.807861 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Aug 13 07:12:44.807883 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Aug 13 07:12:44.807905 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:12:44.807927 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 07:12:44.807949 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Aug 13 07:12:44.807970 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:12:44.807992 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:12:44.808015 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:12:44.808040 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:12:44.808062 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:12:44.808089 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 07:12:44.810167 systemd[1]: Reached target network-pre.target - Preparation for Network. Aug 13 07:12:44.810198 kernel: loop: module loaded Aug 13 07:12:44.811668 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Aug 13 07:12:44.817217 kernel: fuse: init (API version 7.39) Aug 13 07:12:44.817291 systemd-journald[1538]: Collecting audit messages is disabled. Aug 13 07:12:44.817340 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:12:44.817363 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:12:44.817383 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:12:44.817405 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Aug 13 07:12:44.817427 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Aug 13 07:12:44.817458 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 07:12:44.817479 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:12:44.817501 systemd-journald[1538]: Journal started Aug 13 07:12:44.817544 systemd-journald[1538]: Runtime Journal (/run/log/journal/ec2c78c10ae2aab33d5047134f04b0aa) is 4.7M, max 38.2M, 33.4M free. Aug 13 07:12:44.313505 systemd[1]: Queued start job for default target multi-user.target. Aug 13 07:12:44.378691 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Aug 13 07:12:44.379190 systemd[1]: systemd-journald.service: Deactivated successfully. Aug 13 07:12:44.824175 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Aug 13 07:12:44.837998 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Aug 13 07:12:44.857239 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Aug 13 07:12:44.864124 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:12:44.894805 kernel: ACPI: bus type drm_connector registered Aug 13 07:12:44.894891 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Aug 13 07:12:44.899134 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:12:44.913161 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Aug 13 07:12:44.918131 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:12:44.946083 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Aug 13 07:12:44.946201 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:12:44.948528 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Aug 13 07:12:44.949966 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:12:44.952233 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:12:44.953508 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 07:12:44.953719 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Aug 13 07:12:44.955202 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Aug 13 07:12:44.957182 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:12:44.965720 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Aug 13 07:12:44.979066 kernel: loop0: detected capacity change from 0 to 142488 Aug 13 07:12:44.988018 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Aug 13 07:12:45.004297 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Aug 13 07:12:45.014288 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Aug 13 07:12:45.025358 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Aug 13 07:12:45.031891 systemd[1]: Starting systemd-sysusers.service - Create System Users... Aug 13 07:12:45.035727 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:12:45.037466 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Aug 13 07:12:45.046817 systemd-journald[1538]: Time spent on flushing to /var/log/journal/ec2c78c10ae2aab33d5047134f04b0aa is 79.436ms for 989 entries. Aug 13 07:12:45.046817 systemd-journald[1538]: System Journal (/var/log/journal/ec2c78c10ae2aab33d5047134f04b0aa) is 8.0M, max 195.6M, 187.6M free. Aug 13 07:12:45.137185 systemd-journald[1538]: Received client request to flush runtime journal. Aug 13 07:12:45.137304 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 07:12:45.054363 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Aug 13 07:12:45.098565 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 07:12:45.102710 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Aug 13 07:12:45.125778 udevadm[1593]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Aug 13 07:12:45.142619 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Aug 13 07:12:45.161762 kernel: loop1: detected capacity change from 0 to 140768 Aug 13 07:12:45.190433 systemd[1]: Finished systemd-sysusers.service - Create System Users. Aug 13 07:12:45.200690 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:12:45.235694 systemd-tmpfiles[1601]: ACLs are not supported, ignoring. Aug 13 07:12:45.236136 systemd-tmpfiles[1601]: ACLs are not supported, ignoring. Aug 13 07:12:45.246089 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:12:45.286154 kernel: loop2: detected capacity change from 0 to 229808 Aug 13 07:12:45.339672 kernel: loop3: detected capacity change from 0 to 61336 Aug 13 07:12:45.385128 kernel: loop4: detected capacity change from 0 to 142488 Aug 13 07:12:45.425145 kernel: loop5: detected capacity change from 0 to 140768 Aug 13 07:12:45.452139 kernel: loop6: detected capacity change from 0 to 229808 Aug 13 07:12:45.492148 kernel: loop7: detected capacity change from 0 to 61336 Aug 13 07:12:45.523736 (sd-merge)[1608]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Aug 13 07:12:45.525529 (sd-merge)[1608]: Merged extensions into '/usr'. Aug 13 07:12:45.536653 systemd[1]: Reloading requested from client PID 1565 ('systemd-sysext') (unit systemd-sysext.service)... Aug 13 07:12:45.536798 systemd[1]: Reloading... Aug 13 07:12:45.677139 zram_generator::config[1634]: No configuration found. Aug 13 07:12:45.894215 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:12:46.000909 systemd[1]: Reloading finished in 463 ms. Aug 13 07:12:46.034532 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Aug 13 07:12:46.044199 systemd[1]: Starting ensure-sysext.service... Aug 13 07:12:46.047322 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:12:46.070269 systemd[1]: Reloading requested from client PID 1685 ('systemctl') (unit ensure-sysext.service)... Aug 13 07:12:46.070286 systemd[1]: Reloading... Aug 13 07:12:46.124477 systemd-tmpfiles[1686]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 07:12:46.124997 systemd-tmpfiles[1686]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Aug 13 07:12:46.129036 systemd-tmpfiles[1686]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 07:12:46.129666 systemd-tmpfiles[1686]: ACLs are not supported, ignoring. Aug 13 07:12:46.132214 systemd-tmpfiles[1686]: ACLs are not supported, ignoring. Aug 13 07:12:46.159741 systemd-tmpfiles[1686]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:12:46.159767 systemd-tmpfiles[1686]: Skipping /boot Aug 13 07:12:46.175168 zram_generator::config[1709]: No configuration found. Aug 13 07:12:46.214575 systemd-tmpfiles[1686]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:12:46.214598 systemd-tmpfiles[1686]: Skipping /boot Aug 13 07:12:46.374135 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:12:46.433299 systemd[1]: Reloading finished in 362 ms. Aug 13 07:12:46.447183 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Aug 13 07:12:46.452904 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:12:46.463374 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 07:12:46.474395 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Aug 13 07:12:46.477986 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Aug 13 07:12:46.483396 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:12:46.494457 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:12:46.499776 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Aug 13 07:12:46.514501 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:46.514812 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:12:46.522633 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:12:46.533178 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:12:46.540514 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:12:46.541500 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:12:46.541802 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:46.549238 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:46.549892 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:12:46.550308 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:12:46.570348 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Aug 13 07:12:46.571541 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:46.572705 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:12:46.572892 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:12:46.576904 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:12:46.578190 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:12:46.589743 systemd-udevd[1772]: Using default interface naming scheme 'v255'. Aug 13 07:12:46.592419 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:12:46.599041 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:46.600160 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:12:46.610289 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:12:46.616359 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:12:46.620331 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:12:46.621709 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:12:46.622845 systemd[1]: Reached target time-set.target - System Time Set. Aug 13 07:12:46.624319 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Aug 13 07:12:46.629893 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Aug 13 07:12:46.636866 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:12:46.637065 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:12:46.647199 systemd[1]: Finished ensure-sysext.service. Aug 13 07:12:46.649929 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:12:46.650185 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:12:46.652488 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:12:46.652682 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:12:46.664667 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:12:46.664816 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:12:46.668041 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:12:46.668658 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:12:46.698185 ldconfig[1561]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 07:12:46.703930 augenrules[1803]: No rules Aug 13 07:12:46.701822 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Aug 13 07:12:46.704800 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 07:12:46.710608 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Aug 13 07:12:46.721381 systemd[1]: Starting systemd-update-done.service - Update is Completed... Aug 13 07:12:46.728268 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:12:46.729947 systemd[1]: Started systemd-userdbd.service - User Database Manager. Aug 13 07:12:46.739234 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:12:46.782226 systemd[1]: Finished systemd-update-done.service - Update is Completed. Aug 13 07:12:46.799581 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Aug 13 07:12:46.801061 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 07:12:46.895639 systemd-resolved[1771]: Positive Trust Anchors: Aug 13 07:12:46.896029 systemd-resolved[1771]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:12:46.896190 systemd-resolved[1771]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:12:46.898826 systemd-networkd[1818]: lo: Link UP Aug 13 07:12:46.898836 systemd-networkd[1818]: lo: Gained carrier Aug 13 07:12:46.899707 systemd-networkd[1818]: Enumeration completed Aug 13 07:12:46.900285 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:12:46.908366 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Aug 13 07:12:46.909200 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Aug 13 07:12:46.920919 systemd-resolved[1771]: Defaulting to hostname 'linux'. Aug 13 07:12:46.927008 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:12:46.927759 systemd[1]: Reached target network.target - Network. Aug 13 07:12:46.929413 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:12:46.948775 (udev-worker)[1820]: Network interface NamePolicy= disabled on kernel command line. Aug 13 07:12:47.002183 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Aug 13 07:12:47.015123 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Aug 13 07:12:47.030053 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input3 Aug 13 07:12:47.037222 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (1830) Aug 13 07:12:47.043597 systemd-networkd[1818]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:12:47.043614 systemd-networkd[1818]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:12:47.046841 systemd-networkd[1818]: eth0: Link UP Aug 13 07:12:47.047407 systemd-networkd[1818]: eth0: Gained carrier Aug 13 07:12:47.048214 systemd-networkd[1818]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:12:47.053155 kernel: ACPI: button: Power Button [PWRF] Aug 13 07:12:47.053226 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input4 Aug 13 07:12:47.058130 kernel: ACPI: button: Sleep Button [SLPF] Aug 13 07:12:47.062539 systemd-networkd[1818]: eth0: DHCPv4 address 172.31.16.138/20, gateway 172.31.16.1 acquired from 172.31.16.1 Aug 13 07:12:47.171581 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:12:47.190940 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:12:47.191537 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:12:47.220354 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:12:47.264180 kernel: mousedev: PS/2 mouse device common for all mice Aug 13 07:12:47.289965 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Aug 13 07:12:47.294362 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Aug 13 07:12:47.295383 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Aug 13 07:12:47.299070 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Aug 13 07:12:47.316485 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Aug 13 07:12:47.340455 lvm[1937]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:12:47.372051 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Aug 13 07:12:47.372710 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:12:47.377445 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Aug 13 07:12:47.378537 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:12:47.379769 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:12:47.380801 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Aug 13 07:12:47.381630 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Aug 13 07:12:47.382352 systemd[1]: Started logrotate.timer - Daily rotation of log files. Aug 13 07:12:47.382922 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Aug 13 07:12:47.383573 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Aug 13 07:12:47.384022 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 07:12:47.384119 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:12:47.384504 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:12:47.385024 lvm[1943]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:12:47.386962 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Aug 13 07:12:47.390564 systemd[1]: Starting docker.socket - Docker Socket for the API... Aug 13 07:12:47.397460 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Aug 13 07:12:47.399329 systemd[1]: Listening on docker.socket - Docker Socket for the API. Aug 13 07:12:47.400275 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:12:47.400918 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:12:47.401634 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:12:47.401679 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:12:47.413324 systemd[1]: Starting containerd.service - containerd container runtime... Aug 13 07:12:47.417765 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Aug 13 07:12:47.420348 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Aug 13 07:12:47.423617 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Aug 13 07:12:47.433831 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Aug 13 07:12:47.435068 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Aug 13 07:12:47.439549 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Aug 13 07:12:47.451361 systemd[1]: Started ntpd.service - Network Time Service. Aug 13 07:12:47.454347 jq[1950]: false Aug 13 07:12:47.460334 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Aug 13 07:12:47.465936 systemd[1]: Starting setup-oem.service - Setup OEM... Aug 13 07:12:47.482901 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Aug 13 07:12:47.506323 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Aug 13 07:12:47.516604 systemd[1]: Starting systemd-logind.service - User Login Management... Aug 13 07:12:47.521523 extend-filesystems[1951]: Found loop4 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found loop5 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found loop6 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found loop7 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1p1 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1p2 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1p3 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found usr Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1p4 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1p6 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1p7 Aug 13 07:12:47.521523 extend-filesystems[1951]: Found nvme0n1p9 Aug 13 07:12:47.521523 extend-filesystems[1951]: Checking size of /dev/nvme0n1p9 Aug 13 07:12:47.611227 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Aug 13 07:12:47.519724 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Aug 13 07:12:47.611608 extend-filesystems[1951]: Resized partition /dev/nvme0n1p9 Aug 13 07:12:47.520467 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Aug 13 07:12:47.616706 extend-filesystems[1972]: resize2fs 1.47.1 (20-May-2024) Aug 13 07:12:47.537780 systemd[1]: Starting update-engine.service - Update Engine... Aug 13 07:12:47.553020 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Aug 13 07:12:47.556969 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Aug 13 07:12:47.618682 jq[1968]: true Aug 13 07:12:47.569446 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 07:12:47.570184 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Aug 13 07:12:47.584637 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 07:12:47.585202 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Aug 13 07:12:47.628893 dbus-daemon[1949]: [system] SELinux support is enabled Aug 13 07:12:47.630001 systemd[1]: Started dbus.service - D-Bus System Message Bus. Aug 13 07:12:47.639682 ntpd[1953]: ntpd 4.2.8p17@1.4004-o Tue Aug 12 21:30:10 UTC 2025 (1): Starting Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: ntpd 4.2.8p17@1.4004-o Tue Aug 12 21:30:10 UTC 2025 (1): Starting Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: ---------------------------------------------------- Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: ntp-4 is maintained by Network Time Foundation, Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: corporation. Support and training for ntp-4 are Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: available at https://www.nwtime.org/support Aug 13 07:12:47.642157 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: ---------------------------------------------------- Aug 13 07:12:47.641038 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 07:12:47.639717 ntpd[1953]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Aug 13 07:12:47.641193 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Aug 13 07:12:47.639728 ntpd[1953]: ---------------------------------------------------- Aug 13 07:12:47.642366 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 07:12:47.639739 ntpd[1953]: ntp-4 is maintained by Network Time Foundation, Aug 13 07:12:47.642394 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Aug 13 07:12:47.639749 ntpd[1953]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Aug 13 07:12:47.639759 ntpd[1953]: corporation. Support and training for ntp-4 are Aug 13 07:12:47.639769 ntpd[1953]: available at https://www.nwtime.org/support Aug 13 07:12:47.639779 ntpd[1953]: ---------------------------------------------------- Aug 13 07:12:47.650701 ntpd[1953]: proto: precision = 0.082 usec (-23) Aug 13 07:12:47.655284 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: proto: precision = 0.082 usec (-23) Aug 13 07:12:47.655284 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: basedate set to 2025-07-31 Aug 13 07:12:47.655284 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: gps base set to 2025-08-03 (week 2378) Aug 13 07:12:47.651077 ntpd[1953]: basedate set to 2025-07-31 Aug 13 07:12:47.651114 ntpd[1953]: gps base set to 2025-08-03 (week 2378) Aug 13 07:12:47.657339 ntpd[1953]: Listen and drop on 0 v6wildcard [::]:123 Aug 13 07:12:47.659389 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Listen and drop on 0 v6wildcard [::]:123 Aug 13 07:12:47.659389 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Aug 13 07:12:47.659295 ntpd[1953]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Aug 13 07:12:47.659735 ntpd[1953]: Listen normally on 2 lo 127.0.0.1:123 Aug 13 07:12:47.662188 ntpd[1953]: Listen normally on 3 eth0 172.31.16.138:123 Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Listen normally on 2 lo 127.0.0.1:123 Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Listen normally on 3 eth0 172.31.16.138:123 Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Listen normally on 4 lo [::1]:123 Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: bind(21) AF_INET6 fe80::43e:37ff:fe5b:65f1%2#123 flags 0x11 failed: Cannot assign requested address Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: unable to create socket on eth0 (5) for fe80::43e:37ff:fe5b:65f1%2#123 Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: failed to init interface for address fe80::43e:37ff:fe5b:65f1%2 Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: Listening on routing socket on fd #21 for interface updates Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Aug 13 07:12:47.664400 ntpd[1953]: 13 Aug 07:12:47 ntpd[1953]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Aug 13 07:12:47.662237 ntpd[1953]: Listen normally on 4 lo [::1]:123 Aug 13 07:12:47.662291 ntpd[1953]: bind(21) AF_INET6 fe80::43e:37ff:fe5b:65f1%2#123 flags 0x11 failed: Cannot assign requested address Aug 13 07:12:47.662312 ntpd[1953]: unable to create socket on eth0 (5) for fe80::43e:37ff:fe5b:65f1%2#123 Aug 13 07:12:47.662327 ntpd[1953]: failed to init interface for address fe80::43e:37ff:fe5b:65f1%2 Aug 13 07:12:47.662364 ntpd[1953]: Listening on routing socket on fd #21 for interface updates Aug 13 07:12:47.663809 ntpd[1953]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Aug 13 07:12:47.663838 ntpd[1953]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Aug 13 07:12:47.668483 dbus-daemon[1949]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1818 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Aug 13 07:12:47.681957 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 07:12:47.682222 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Aug 13 07:12:47.689406 jq[1984]: true Aug 13 07:12:47.699029 (ntainerd)[1991]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Aug 13 07:12:47.699392 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Aug 13 07:12:47.717126 tar[1971]: linux-amd64/LICENSE Aug 13 07:12:47.722599 tar[1971]: linux-amd64/helm Aug 13 07:12:47.724127 update_engine[1963]: I20250813 07:12:47.723986 1963 main.cc:92] Flatcar Update Engine starting Aug 13 07:12:47.746146 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Aug 13 07:12:47.754648 systemd[1]: Started update-engine.service - Update Engine. Aug 13 07:12:47.766185 update_engine[1963]: I20250813 07:12:47.756376 1963 update_check_scheduler.cc:74] Next update check in 11m18s Aug 13 07:12:47.764355 systemd[1]: Started locksmithd.service - Cluster reboot manager. Aug 13 07:12:47.772129 extend-filesystems[1972]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Aug 13 07:12:47.772129 extend-filesystems[1972]: old_desc_blocks = 1, new_desc_blocks = 1 Aug 13 07:12:47.772129 extend-filesystems[1972]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Aug 13 07:12:47.777602 extend-filesystems[1951]: Resized filesystem in /dev/nvme0n1p9 Aug 13 07:12:47.773341 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 07:12:47.773800 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Aug 13 07:12:47.785046 coreos-metadata[1948]: Aug 13 07:12:47.784 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Aug 13 07:12:47.802150 coreos-metadata[1948]: Aug 13 07:12:47.800 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Aug 13 07:12:47.802150 coreos-metadata[1948]: Aug 13 07:12:47.801 INFO Fetch successful Aug 13 07:12:47.802150 coreos-metadata[1948]: Aug 13 07:12:47.801 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Aug 13 07:12:47.804166 systemd[1]: Finished setup-oem.service - Setup OEM. Aug 13 07:12:47.805835 coreos-metadata[1948]: Aug 13 07:12:47.805 INFO Fetch successful Aug 13 07:12:47.805835 coreos-metadata[1948]: Aug 13 07:12:47.805 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Aug 13 07:12:47.810360 coreos-metadata[1948]: Aug 13 07:12:47.810 INFO Fetch successful Aug 13 07:12:47.810547 coreos-metadata[1948]: Aug 13 07:12:47.810 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Aug 13 07:12:47.813714 coreos-metadata[1948]: Aug 13 07:12:47.813 INFO Fetch successful Aug 13 07:12:47.813808 coreos-metadata[1948]: Aug 13 07:12:47.813 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Aug 13 07:12:47.814627 coreos-metadata[1948]: Aug 13 07:12:47.814 INFO Fetch failed with 404: resource not found Aug 13 07:12:47.814627 coreos-metadata[1948]: Aug 13 07:12:47.814 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Aug 13 07:12:47.816978 systemd-logind[1961]: Watching system buttons on /dev/input/event2 (Power Button) Aug 13 07:12:47.817012 systemd-logind[1961]: Watching system buttons on /dev/input/event3 (Sleep Button) Aug 13 07:12:47.817037 systemd-logind[1961]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Aug 13 07:12:47.818014 coreos-metadata[1948]: Aug 13 07:12:47.817 INFO Fetch successful Aug 13 07:12:47.818014 coreos-metadata[1948]: Aug 13 07:12:47.817 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Aug 13 07:12:47.821550 coreos-metadata[1948]: Aug 13 07:12:47.819 INFO Fetch successful Aug 13 07:12:47.821550 coreos-metadata[1948]: Aug 13 07:12:47.819 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Aug 13 07:12:47.821550 coreos-metadata[1948]: Aug 13 07:12:47.820 INFO Fetch successful Aug 13 07:12:47.821550 coreos-metadata[1948]: Aug 13 07:12:47.820 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Aug 13 07:12:47.821354 systemd-logind[1961]: New seat seat0. Aug 13 07:12:47.822605 coreos-metadata[1948]: Aug 13 07:12:47.822 INFO Fetch successful Aug 13 07:12:47.822605 coreos-metadata[1948]: Aug 13 07:12:47.822 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Aug 13 07:12:47.823878 coreos-metadata[1948]: Aug 13 07:12:47.823 INFO Fetch successful Aug 13 07:12:47.827624 systemd[1]: Started systemd-logind.service - User Login Management. Aug 13 07:12:47.913715 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Aug 13 07:12:47.931985 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Aug 13 07:12:47.944615 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (1820) Aug 13 07:12:47.955435 bash[2032]: Updated "/home/core/.ssh/authorized_keys" Aug 13 07:12:47.957523 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Aug 13 07:12:47.969513 systemd[1]: Starting sshkeys.service... Aug 13 07:12:47.999980 locksmithd[2010]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 07:12:48.036156 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Aug 13 07:12:48.046276 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Aug 13 07:12:48.166331 systemd-networkd[1818]: eth0: Gained IPv6LL Aug 13 07:12:48.171511 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Aug 13 07:12:48.208338 systemd[1]: Reached target network-online.target - Network is Online. Aug 13 07:12:48.216595 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Aug 13 07:12:48.228575 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:12:48.237624 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Aug 13 07:12:48.247497 coreos-metadata[2090]: Aug 13 07:12:48.247 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Aug 13 07:12:48.250589 coreos-metadata[2090]: Aug 13 07:12:48.248 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Aug 13 07:12:48.250589 coreos-metadata[2090]: Aug 13 07:12:48.249 INFO Fetch successful Aug 13 07:12:48.250589 coreos-metadata[2090]: Aug 13 07:12:48.249 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Aug 13 07:12:48.259586 coreos-metadata[2090]: Aug 13 07:12:48.251 INFO Fetch successful Aug 13 07:12:48.260854 unknown[2090]: wrote ssh authorized keys file for user: core Aug 13 07:12:48.378585 dbus-daemon[1949]: [system] Successfully activated service 'org.freedesktop.hostname1' Aug 13 07:12:48.378769 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Aug 13 07:12:48.388697 dbus-daemon[1949]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.7' (uid=0 pid=1999 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Aug 13 07:12:48.400552 systemd[1]: Starting polkit.service - Authorization Manager... Aug 13 07:12:48.405588 update-ssh-keys[2140]: Updated "/home/core/.ssh/authorized_keys" Aug 13 07:12:48.413080 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Aug 13 07:12:48.416531 systemd[1]: Finished sshkeys.service. Aug 13 07:12:48.439863 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Aug 13 07:12:48.442185 polkitd[2146]: Started polkitd version 121 Aug 13 07:12:48.468464 polkitd[2146]: Loading rules from directory /etc/polkit-1/rules.d Aug 13 07:12:48.468561 polkitd[2146]: Loading rules from directory /usr/share/polkit-1/rules.d Aug 13 07:12:48.474575 polkitd[2146]: Finished loading, compiling and executing 2 rules Aug 13 07:12:48.475239 dbus-daemon[1949]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Aug 13 07:12:48.475724 polkitd[2146]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Aug 13 07:12:48.476276 systemd[1]: Started polkit.service - Authorization Manager. Aug 13 07:12:48.536039 systemd-hostnamed[1999]: Hostname set to (transient) Aug 13 07:12:48.536403 systemd-resolved[1771]: System hostname changed to 'ip-172-31-16-138'. Aug 13 07:12:48.543147 amazon-ssm-agent[2135]: Initializing new seelog logger Aug 13 07:12:48.543147 amazon-ssm-agent[2135]: New Seelog Logger Creation Complete Aug 13 07:12:48.543147 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.543147 amazon-ssm-agent[2135]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.543818 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 processing appconfig overrides Aug 13 07:12:48.544335 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.544418 amazon-ssm-agent[2135]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.545669 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 processing appconfig overrides Aug 13 07:12:48.545669 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.545669 amazon-ssm-agent[2135]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.545669 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 processing appconfig overrides Aug 13 07:12:48.545669 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO Proxy environment variables: Aug 13 07:12:48.552387 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.552387 amazon-ssm-agent[2135]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Aug 13 07:12:48.552556 amazon-ssm-agent[2135]: 2025/08/13 07:12:48 processing appconfig overrides Aug 13 07:12:48.648627 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO https_proxy: Aug 13 07:12:48.676165 sshd_keygen[1994]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 07:12:48.746566 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO http_proxy: Aug 13 07:12:48.753629 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Aug 13 07:12:48.762867 systemd[1]: Starting issuegen.service - Generate /run/issue... Aug 13 07:12:48.797608 containerd[1991]: time="2025-08-13T07:12:48.795908846Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Aug 13 07:12:48.813858 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 07:12:48.814161 systemd[1]: Finished issuegen.service - Generate /run/issue. Aug 13 07:12:48.827473 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Aug 13 07:12:48.848625 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO no_proxy: Aug 13 07:12:48.880823 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Aug 13 07:12:48.891753 systemd[1]: Started getty@tty1.service - Getty on tty1. Aug 13 07:12:48.895511 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Aug 13 07:12:48.897200 systemd[1]: Reached target getty.target - Login Prompts. Aug 13 07:12:48.947252 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO Checking if agent identity type OnPrem can be assumed Aug 13 07:12:48.951662 containerd[1991]: time="2025-08-13T07:12:48.951567597Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:12:48.956026 containerd[1991]: time="2025-08-13T07:12:48.955525518Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.100-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:12:48.956026 containerd[1991]: time="2025-08-13T07:12:48.955579817Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Aug 13 07:12:48.956026 containerd[1991]: time="2025-08-13T07:12:48.955607696Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Aug 13 07:12:48.956026 containerd[1991]: time="2025-08-13T07:12:48.955795156Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Aug 13 07:12:48.956026 containerd[1991]: time="2025-08-13T07:12:48.955817929Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Aug 13 07:12:48.956026 containerd[1991]: time="2025-08-13T07:12:48.955886341Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:12:48.956026 containerd[1991]: time="2025-08-13T07:12:48.955905343Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959037 containerd[1991]: time="2025-08-13T07:12:48.958339208Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959037 containerd[1991]: time="2025-08-13T07:12:48.958387533Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959037 containerd[1991]: time="2025-08-13T07:12:48.958412404Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959037 containerd[1991]: time="2025-08-13T07:12:48.958428469Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959037 containerd[1991]: time="2025-08-13T07:12:48.958580280Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959037 containerd[1991]: time="2025-08-13T07:12:48.958854694Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959346 containerd[1991]: time="2025-08-13T07:12:48.959045825Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:12:48.959346 containerd[1991]: time="2025-08-13T07:12:48.959065589Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Aug 13 07:12:48.959346 containerd[1991]: time="2025-08-13T07:12:48.959255680Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Aug 13 07:12:48.959346 containerd[1991]: time="2025-08-13T07:12:48.959319137Z" level=info msg="metadata content store policy set" policy=shared Aug 13 07:12:48.965865 containerd[1991]: time="2025-08-13T07:12:48.965535725Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Aug 13 07:12:48.965865 containerd[1991]: time="2025-08-13T07:12:48.965618472Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Aug 13 07:12:48.965865 containerd[1991]: time="2025-08-13T07:12:48.965642518Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Aug 13 07:12:48.965865 containerd[1991]: time="2025-08-13T07:12:48.965665282Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Aug 13 07:12:48.965865 containerd[1991]: time="2025-08-13T07:12:48.965687678Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Aug 13 07:12:48.965865 containerd[1991]: time="2025-08-13T07:12:48.965875104Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966268626Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966410595Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966432859Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966454429Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966477099Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966496687Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966515112Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966537248Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966558150Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966580163Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966600560Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966618476Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966656467Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.966727 containerd[1991]: time="2025-08-13T07:12:48.966674471Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966691223Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966712113Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966731963Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966750569Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966776298Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966795810Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966815431Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966837425Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966858776Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966879230Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966905181Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966929280Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966959213Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966977808Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.967243 containerd[1991]: time="2025-08-13T07:12:48.966994290Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Aug 13 07:12:48.967791 containerd[1991]: time="2025-08-13T07:12:48.967045597Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Aug 13 07:12:48.967791 containerd[1991]: time="2025-08-13T07:12:48.967073384Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Aug 13 07:12:48.971186 containerd[1991]: time="2025-08-13T07:12:48.967091278Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Aug 13 07:12:48.971186 containerd[1991]: time="2025-08-13T07:12:48.970346434Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Aug 13 07:12:48.971186 containerd[1991]: time="2025-08-13T07:12:48.970372617Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.971186 containerd[1991]: time="2025-08-13T07:12:48.970398183Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Aug 13 07:12:48.971186 containerd[1991]: time="2025-08-13T07:12:48.970414223Z" level=info msg="NRI interface is disabled by configuration." Aug 13 07:12:48.971186 containerd[1991]: time="2025-08-13T07:12:48.970430987Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Aug 13 07:12:48.971473 containerd[1991]: time="2025-08-13T07:12:48.970838273Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Aug 13 07:12:48.971473 containerd[1991]: time="2025-08-13T07:12:48.970938399Z" level=info msg="Connect containerd service" Aug 13 07:12:48.971473 containerd[1991]: time="2025-08-13T07:12:48.970991124Z" level=info msg="using legacy CRI server" Aug 13 07:12:48.971473 containerd[1991]: time="2025-08-13T07:12:48.971002772Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Aug 13 07:12:48.971473 containerd[1991]: time="2025-08-13T07:12:48.971154101Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Aug 13 07:12:48.975597 containerd[1991]: time="2025-08-13T07:12:48.974611172Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 07:12:48.977435 containerd[1991]: time="2025-08-13T07:12:48.977398748Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 07:12:48.977526 containerd[1991]: time="2025-08-13T07:12:48.977472382Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 07:12:48.979291 containerd[1991]: time="2025-08-13T07:12:48.977542903Z" level=info msg="Start subscribing containerd event" Aug 13 07:12:48.979291 containerd[1991]: time="2025-08-13T07:12:48.977609394Z" level=info msg="Start recovering state" Aug 13 07:12:48.979291 containerd[1991]: time="2025-08-13T07:12:48.977696566Z" level=info msg="Start event monitor" Aug 13 07:12:48.979291 containerd[1991]: time="2025-08-13T07:12:48.977721515Z" level=info msg="Start snapshots syncer" Aug 13 07:12:48.979291 containerd[1991]: time="2025-08-13T07:12:48.977735145Z" level=info msg="Start cni network conf syncer for default" Aug 13 07:12:48.979291 containerd[1991]: time="2025-08-13T07:12:48.977745173Z" level=info msg="Start streaming server" Aug 13 07:12:48.979291 containerd[1991]: time="2025-08-13T07:12:48.977821913Z" level=info msg="containerd successfully booted in 0.184340s" Aug 13 07:12:48.977927 systemd[1]: Started containerd.service - containerd container runtime. Aug 13 07:12:49.047805 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO Checking if agent identity type EC2 can be assumed Aug 13 07:12:49.145797 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO Agent will take identity from EC2 Aug 13 07:12:49.244692 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [amazon-ssm-agent] using named pipe channel for IPC Aug 13 07:12:49.330595 tar[1971]: linux-amd64/README.md Aug 13 07:12:49.344259 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [amazon-ssm-agent] using named pipe channel for IPC Aug 13 07:12:49.348367 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Aug 13 07:12:49.443743 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [amazon-ssm-agent] using named pipe channel for IPC Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [amazon-ssm-agent] Starting Core Agent Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [amazon-ssm-agent] registrar detected. Attempting registration Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [Registrar] Starting registrar module Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:48 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:49 INFO [EC2Identity] EC2 registration was successful. Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:49 INFO [CredentialRefresher] credentialRefresher has started Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:49 INFO [CredentialRefresher] Starting credentials refresher loop Aug 13 07:12:49.514534 amazon-ssm-agent[2135]: 2025-08-13 07:12:49 INFO EC2RoleProvider Successfully connected with instance profile role credentials Aug 13 07:12:49.543163 amazon-ssm-agent[2135]: 2025-08-13 07:12:49 INFO [CredentialRefresher] Next credential rotation will be in 31.7416614184 minutes Aug 13 07:12:50.338376 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:12:50.339530 systemd[1]: Reached target multi-user.target - Multi-User System. Aug 13 07:12:50.340611 systemd[1]: Startup finished in 655ms (kernel) + 6.559s (initrd) + 7.113s (userspace) = 14.328s. Aug 13 07:12:50.346828 (kubelet)[2197]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:12:50.529424 amazon-ssm-agent[2135]: 2025-08-13 07:12:50 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Aug 13 07:12:50.604985 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Aug 13 07:12:50.614484 systemd[1]: Started sshd@0-172.31.16.138:22-147.75.109.163:40144.service - OpenSSH per-connection server daemon (147.75.109.163:40144). Aug 13 07:12:50.630205 amazon-ssm-agent[2135]: 2025-08-13 07:12:50 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2207) started Aug 13 07:12:50.642429 ntpd[1953]: Listen normally on 6 eth0 [fe80::43e:37ff:fe5b:65f1%2]:123 Aug 13 07:12:50.644488 ntpd[1953]: 13 Aug 07:12:50 ntpd[1953]: Listen normally on 6 eth0 [fe80::43e:37ff:fe5b:65f1%2]:123 Aug 13 07:12:50.730419 amazon-ssm-agent[2135]: 2025-08-13 07:12:50 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Aug 13 07:12:50.800708 sshd[2213]: Accepted publickey for core from 147.75.109.163 port 40144 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:12:50.803414 sshd[2213]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:50.817770 systemd-logind[1961]: New session 1 of user core. Aug 13 07:12:50.819316 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Aug 13 07:12:50.826512 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Aug 13 07:12:50.844572 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Aug 13 07:12:50.854549 systemd[1]: Starting user@500.service - User Manager for UID 500... Aug 13 07:12:50.860512 (systemd)[2223]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 07:12:51.023852 systemd[2223]: Queued start job for default target default.target. Aug 13 07:12:51.030917 systemd[2223]: Created slice app.slice - User Application Slice. Aug 13 07:12:51.031240 systemd[2223]: Reached target paths.target - Paths. Aug 13 07:12:51.031290 systemd[2223]: Reached target timers.target - Timers. Aug 13 07:12:51.033938 systemd[2223]: Starting dbus.socket - D-Bus User Message Bus Socket... Aug 13 07:12:51.049866 systemd[2223]: Listening on dbus.socket - D-Bus User Message Bus Socket. Aug 13 07:12:51.050837 systemd[2223]: Reached target sockets.target - Sockets. Aug 13 07:12:51.050866 systemd[2223]: Reached target basic.target - Basic System. Aug 13 07:12:51.051147 systemd[1]: Started user@500.service - User Manager for UID 500. Aug 13 07:12:51.052718 systemd[2223]: Reached target default.target - Main User Target. Aug 13 07:12:51.052782 systemd[2223]: Startup finished in 183ms. Aug 13 07:12:51.058378 systemd[1]: Started session-1.scope - Session 1 of User core. Aug 13 07:12:51.201223 systemd[1]: Started sshd@1-172.31.16.138:22-147.75.109.163:40148.service - OpenSSH per-connection server daemon (147.75.109.163:40148). Aug 13 07:12:51.303074 kubelet[2197]: E0813 07:12:51.302988 2197 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:12:51.304729 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:12:51.304887 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:12:51.305294 systemd[1]: kubelet.service: Consumed 1.159s CPU time. Aug 13 07:12:51.356807 sshd[2234]: Accepted publickey for core from 147.75.109.163 port 40148 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:12:51.358356 sshd[2234]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:51.364477 systemd-logind[1961]: New session 2 of user core. Aug 13 07:12:51.372469 systemd[1]: Started session-2.scope - Session 2 of User core. Aug 13 07:12:51.487972 sshd[2234]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:51.498327 systemd[1]: sshd@1-172.31.16.138:22-147.75.109.163:40148.service: Deactivated successfully. Aug 13 07:12:51.500292 systemd[1]: session-2.scope: Deactivated successfully. Aug 13 07:12:51.501092 systemd-logind[1961]: Session 2 logged out. Waiting for processes to exit. Aug 13 07:12:51.502401 systemd-logind[1961]: Removed session 2. Aug 13 07:12:51.517919 systemd[1]: Started sshd@2-172.31.16.138:22-147.75.109.163:40152.service - OpenSSH per-connection server daemon (147.75.109.163:40152). Aug 13 07:12:51.680416 sshd[2243]: Accepted publickey for core from 147.75.109.163 port 40152 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:12:51.682008 sshd[2243]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:51.686960 systemd-logind[1961]: New session 3 of user core. Aug 13 07:12:51.693358 systemd[1]: Started session-3.scope - Session 3 of User core. Aug 13 07:12:51.810209 sshd[2243]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:51.813737 systemd[1]: sshd@2-172.31.16.138:22-147.75.109.163:40152.service: Deactivated successfully. Aug 13 07:12:51.815454 systemd[1]: session-3.scope: Deactivated successfully. Aug 13 07:12:51.816027 systemd-logind[1961]: Session 3 logged out. Waiting for processes to exit. Aug 13 07:12:51.817014 systemd-logind[1961]: Removed session 3. Aug 13 07:12:51.848503 systemd[1]: Started sshd@3-172.31.16.138:22-147.75.109.163:40162.service - OpenSSH per-connection server daemon (147.75.109.163:40162). Aug 13 07:12:52.008285 sshd[2250]: Accepted publickey for core from 147.75.109.163 port 40162 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:12:52.009841 sshd[2250]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:52.015299 systemd-logind[1961]: New session 4 of user core. Aug 13 07:12:52.022413 systemd[1]: Started session-4.scope - Session 4 of User core. Aug 13 07:12:52.141959 sshd[2250]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:52.145811 systemd[1]: sshd@3-172.31.16.138:22-147.75.109.163:40162.service: Deactivated successfully. Aug 13 07:12:52.147852 systemd[1]: session-4.scope: Deactivated successfully. Aug 13 07:12:52.149694 systemd-logind[1961]: Session 4 logged out. Waiting for processes to exit. Aug 13 07:12:52.150973 systemd-logind[1961]: Removed session 4. Aug 13 07:12:52.172675 systemd[1]: Started sshd@4-172.31.16.138:22-147.75.109.163:40168.service - OpenSSH per-connection server daemon (147.75.109.163:40168). Aug 13 07:12:52.339692 sshd[2257]: Accepted publickey for core from 147.75.109.163 port 40168 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:12:52.341117 sshd[2257]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:52.346222 systemd-logind[1961]: New session 5 of user core. Aug 13 07:12:52.356464 systemd[1]: Started session-5.scope - Session 5 of User core. Aug 13 07:12:52.466684 sudo[2260]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Aug 13 07:12:52.467017 sudo[2260]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:12:52.483005 sudo[2260]: pam_unix(sudo:session): session closed for user root Aug 13 07:12:52.506089 sshd[2257]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:52.509741 systemd[1]: sshd@4-172.31.16.138:22-147.75.109.163:40168.service: Deactivated successfully. Aug 13 07:12:52.511399 systemd[1]: session-5.scope: Deactivated successfully. Aug 13 07:12:52.512083 systemd-logind[1961]: Session 5 logged out. Waiting for processes to exit. Aug 13 07:12:52.512982 systemd-logind[1961]: Removed session 5. Aug 13 07:12:52.542541 systemd[1]: Started sshd@5-172.31.16.138:22-147.75.109.163:40180.service - OpenSSH per-connection server daemon (147.75.109.163:40180). Aug 13 07:12:52.695573 sshd[2265]: Accepted publickey for core from 147.75.109.163 port 40180 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:12:52.697262 sshd[2265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:52.702195 systemd-logind[1961]: New session 6 of user core. Aug 13 07:12:52.709379 systemd[1]: Started session-6.scope - Session 6 of User core. Aug 13 07:12:52.806213 sudo[2269]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Aug 13 07:12:52.806506 sudo[2269]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:12:52.810604 sudo[2269]: pam_unix(sudo:session): session closed for user root Aug 13 07:12:52.816457 sudo[2268]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Aug 13 07:12:52.816749 sudo[2268]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:12:52.830499 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Aug 13 07:12:52.834745 auditctl[2272]: No rules Aug 13 07:12:52.835234 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 07:12:52.835485 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Aug 13 07:12:52.838628 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 07:12:52.872326 augenrules[2290]: No rules Aug 13 07:12:52.873963 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 07:12:52.875346 sudo[2268]: pam_unix(sudo:session): session closed for user root Aug 13 07:12:52.897976 sshd[2265]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:52.901232 systemd[1]: sshd@5-172.31.16.138:22-147.75.109.163:40180.service: Deactivated successfully. Aug 13 07:12:52.903409 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 07:12:52.905699 systemd-logind[1961]: Session 6 logged out. Waiting for processes to exit. Aug 13 07:12:52.906873 systemd-logind[1961]: Removed session 6. Aug 13 07:12:52.930577 systemd[1]: Started sshd@6-172.31.16.138:22-147.75.109.163:40190.service - OpenSSH per-connection server daemon (147.75.109.163:40190). Aug 13 07:12:53.100209 sshd[2298]: Accepted publickey for core from 147.75.109.163 port 40190 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:12:53.101720 sshd[2298]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:53.106784 systemd-logind[1961]: New session 7 of user core. Aug 13 07:12:53.118360 systemd[1]: Started session-7.scope - Session 7 of User core. Aug 13 07:12:53.215468 sudo[2301]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 07:12:53.215952 sudo[2301]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:12:53.608795 systemd[1]: Starting docker.service - Docker Application Container Engine... Aug 13 07:12:53.610992 (dockerd)[2318]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Aug 13 07:12:53.992063 dockerd[2318]: time="2025-08-13T07:12:53.990331383Z" level=info msg="Starting up" Aug 13 07:12:54.167723 dockerd[2318]: time="2025-08-13T07:12:54.167671163Z" level=info msg="Loading containers: start." Aug 13 07:12:54.316144 kernel: Initializing XFRM netlink socket Aug 13 07:12:54.352694 (udev-worker)[2342]: Network interface NamePolicy= disabled on kernel command line. Aug 13 07:12:54.422991 systemd-networkd[1818]: docker0: Link UP Aug 13 07:12:54.446046 dockerd[2318]: time="2025-08-13T07:12:54.445986761Z" level=info msg="Loading containers: done." Aug 13 07:12:54.466955 dockerd[2318]: time="2025-08-13T07:12:54.466889084Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 07:12:54.467156 dockerd[2318]: time="2025-08-13T07:12:54.467012296Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Aug 13 07:12:54.467156 dockerd[2318]: time="2025-08-13T07:12:54.467140590Z" level=info msg="Daemon has completed initialization" Aug 13 07:12:54.523472 dockerd[2318]: time="2025-08-13T07:12:54.523324817Z" level=info msg="API listen on /run/docker.sock" Aug 13 07:12:54.523568 systemd[1]: Started docker.service - Docker Application Container Engine. Aug 13 07:12:55.739079 systemd-resolved[1771]: Clock change detected. Flushing caches. Aug 13 07:12:56.543309 containerd[1991]: time="2025-08-13T07:12:56.543253264Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\"" Aug 13 07:12:57.170769 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount932889786.mount: Deactivated successfully. Aug 13 07:12:58.796748 containerd[1991]: time="2025-08-13T07:12:58.795174103Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:12:58.798406 containerd[1991]: time="2025-08-13T07:12:58.798334515Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.3: active requests=0, bytes read=30078237" Aug 13 07:12:58.798803 containerd[1991]: time="2025-08-13T07:12:58.798738719Z" level=info msg="ImageCreate event name:\"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:12:58.803273 containerd[1991]: time="2025-08-13T07:12:58.803208927Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:12:58.804756 containerd[1991]: time="2025-08-13T07:12:58.804557453Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.3\" with image id \"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\", size \"30075037\" in 2.261252556s" Aug 13 07:12:58.804756 containerd[1991]: time="2025-08-13T07:12:58.804609190Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\" returns image reference \"sha256:a92b4b92a991677d355596cc4aa9b0b12cbc38e8cbdc1e476548518ae045bc4a\"" Aug 13 07:12:58.805658 containerd[1991]: time="2025-08-13T07:12:58.805622707Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\"" Aug 13 07:13:00.581345 containerd[1991]: time="2025-08-13T07:13:00.581273954Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:00.582525 containerd[1991]: time="2025-08-13T07:13:00.582474506Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.3: active requests=0, bytes read=26019361" Aug 13 07:13:00.583485 containerd[1991]: time="2025-08-13T07:13:00.583432872Z" level=info msg="ImageCreate event name:\"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:00.586688 containerd[1991]: time="2025-08-13T07:13:00.586633369Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:00.587888 containerd[1991]: time="2025-08-13T07:13:00.587765805Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.3\" with image id \"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\", size \"27646922\" in 1.782104735s" Aug 13 07:13:00.587888 containerd[1991]: time="2025-08-13T07:13:00.587800597Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\" returns image reference \"sha256:bf97fadcef43049604abcf0caf4f35229fbee25bd0cdb6fdc1d2bbb4f03d9660\"" Aug 13 07:13:00.588741 containerd[1991]: time="2025-08-13T07:13:00.588661556Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\"" Aug 13 07:13:02.657889 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 07:13:02.683817 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:13:03.387031 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:13:03.418490 (kubelet)[2529]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:13:03.610424 kubelet[2529]: E0813 07:13:03.610199 2529 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:13:03.626563 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:13:03.626791 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:13:03.866501 containerd[1991]: time="2025-08-13T07:13:03.865834687Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:03.877973 containerd[1991]: time="2025-08-13T07:13:03.877333139Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.3: active requests=0, bytes read=20155013" Aug 13 07:13:03.886093 containerd[1991]: time="2025-08-13T07:13:03.886020226Z" level=info msg="ImageCreate event name:\"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:03.907402 containerd[1991]: time="2025-08-13T07:13:03.907324224Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:03.909624 containerd[1991]: time="2025-08-13T07:13:03.909202804Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.3\" with image id \"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\", size \"21782592\" in 3.320441191s" Aug 13 07:13:03.909624 containerd[1991]: time="2025-08-13T07:13:03.909257355Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\" returns image reference \"sha256:41376797d5122e388663ab6d0ad583e58cff63e1a0f1eebfb31d615d8f1c1c87\"" Aug 13 07:13:03.912093 containerd[1991]: time="2025-08-13T07:13:03.910484399Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\"" Aug 13 07:13:05.258604 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2290806462.mount: Deactivated successfully. Aug 13 07:13:05.957514 containerd[1991]: time="2025-08-13T07:13:05.957447864Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:05.959627 containerd[1991]: time="2025-08-13T07:13:05.959348701Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.3: active requests=0, bytes read=31892666" Aug 13 07:13:05.962770 containerd[1991]: time="2025-08-13T07:13:05.961666949Z" level=info msg="ImageCreate event name:\"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:05.966665 containerd[1991]: time="2025-08-13T07:13:05.964986666Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:05.966665 containerd[1991]: time="2025-08-13T07:13:05.965774529Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.3\" with image id \"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\", repo tag \"registry.k8s.io/kube-proxy:v1.33.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\", size \"31891685\" in 2.055242296s" Aug 13 07:13:05.966665 containerd[1991]: time="2025-08-13T07:13:05.965809389Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\" returns image reference \"sha256:af855adae796077ff822e22c0102f686b2ca7b7c51948889b1825388eaac9234\"" Aug 13 07:13:05.968944 containerd[1991]: time="2025-08-13T07:13:05.968912145Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Aug 13 07:13:06.555227 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2020551434.mount: Deactivated successfully. Aug 13 07:13:07.782501 containerd[1991]: time="2025-08-13T07:13:07.782436874Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:07.786285 containerd[1991]: time="2025-08-13T07:13:07.785516437Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Aug 13 07:13:07.789028 containerd[1991]: time="2025-08-13T07:13:07.788090403Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:07.793954 containerd[1991]: time="2025-08-13T07:13:07.793905206Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:07.795161 containerd[1991]: time="2025-08-13T07:13:07.795125478Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.826174334s" Aug 13 07:13:07.795289 containerd[1991]: time="2025-08-13T07:13:07.795275119Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Aug 13 07:13:07.796713 containerd[1991]: time="2025-08-13T07:13:07.796651710Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 07:13:08.280319 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2620402827.mount: Deactivated successfully. Aug 13 07:13:08.293742 containerd[1991]: time="2025-08-13T07:13:08.293684684Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:08.295947 containerd[1991]: time="2025-08-13T07:13:08.295578882Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Aug 13 07:13:08.299176 containerd[1991]: time="2025-08-13T07:13:08.297972374Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:08.302544 containerd[1991]: time="2025-08-13T07:13:08.302490829Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:08.304734 containerd[1991]: time="2025-08-13T07:13:08.304670198Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 507.975666ms" Aug 13 07:13:08.304838 containerd[1991]: time="2025-08-13T07:13:08.304743820Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Aug 13 07:13:08.306429 containerd[1991]: time="2025-08-13T07:13:08.306394076Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Aug 13 07:13:08.817947 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4125374044.mount: Deactivated successfully. Aug 13 07:13:10.978803 containerd[1991]: time="2025-08-13T07:13:10.978709500Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:10.980843 containerd[1991]: time="2025-08-13T07:13:10.980769387Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58247175" Aug 13 07:13:10.983316 containerd[1991]: time="2025-08-13T07:13:10.983222131Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:10.987580 containerd[1991]: time="2025-08-13T07:13:10.987542980Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:10.989011 containerd[1991]: time="2025-08-13T07:13:10.988847319Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 2.682419074s" Aug 13 07:13:10.989011 containerd[1991]: time="2025-08-13T07:13:10.988891232Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Aug 13 07:13:13.733207 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 07:13:13.740123 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:13:14.054025 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:13:14.055893 (kubelet)[2684]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:13:14.135744 kubelet[2684]: E0813 07:13:14.134787 2684 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:13:14.138090 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:13:14.138284 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:13:15.312920 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:13:15.320128 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:13:15.356107 systemd[1]: Reloading requested from client PID 2698 ('systemctl') (unit session-7.scope)... Aug 13 07:13:15.356131 systemd[1]: Reloading... Aug 13 07:13:15.497776 zram_generator::config[2739]: No configuration found. Aug 13 07:13:15.647022 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:13:15.734598 systemd[1]: Reloading finished in 377 ms. Aug 13 07:13:15.791452 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Aug 13 07:13:15.791575 systemd[1]: kubelet.service: Failed with result 'signal'. Aug 13 07:13:15.792034 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:13:15.799137 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:13:16.010961 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:13:16.022272 (kubelet)[2802]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:13:16.082765 kubelet[2802]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:13:16.082765 kubelet[2802]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 07:13:16.082765 kubelet[2802]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:13:16.086762 kubelet[2802]: I0813 07:13:16.086278 2802 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:13:16.361089 kubelet[2802]: I0813 07:13:16.360594 2802 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 07:13:16.361089 kubelet[2802]: I0813 07:13:16.360630 2802 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:13:16.361089 kubelet[2802]: I0813 07:13:16.360958 2802 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 07:13:16.421748 kubelet[2802]: E0813 07:13:16.420831 2802 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.31.16.138:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Aug 13 07:13:16.421748 kubelet[2802]: I0813 07:13:16.421152 2802 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:13:16.455740 kubelet[2802]: E0813 07:13:16.455542 2802 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:13:16.455740 kubelet[2802]: I0813 07:13:16.455742 2802 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:13:16.469619 kubelet[2802]: I0813 07:13:16.469582 2802 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:13:16.476479 kubelet[2802]: I0813 07:13:16.476394 2802 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:13:16.481110 kubelet[2802]: I0813 07:13:16.476457 2802 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-16-138","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 07:13:16.481110 kubelet[2802]: I0813 07:13:16.481115 2802 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:13:16.481334 kubelet[2802]: I0813 07:13:16.481132 2802 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 07:13:16.481334 kubelet[2802]: I0813 07:13:16.481276 2802 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:13:16.486949 kubelet[2802]: I0813 07:13:16.486459 2802 kubelet.go:480] "Attempting to sync node with API server" Aug 13 07:13:16.486949 kubelet[2802]: I0813 07:13:16.486516 2802 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:13:16.486949 kubelet[2802]: I0813 07:13:16.486561 2802 kubelet.go:386] "Adding apiserver pod source" Aug 13 07:13:16.489805 kubelet[2802]: I0813 07:13:16.489448 2802 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:13:16.496342 kubelet[2802]: E0813 07:13:16.496122 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.16.138:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-16-138&limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Aug 13 07:13:16.499397 kubelet[2802]: E0813 07:13:16.499199 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.16.138:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Aug 13 07:13:16.499397 kubelet[2802]: I0813 07:13:16.499300 2802 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 07:13:16.499991 kubelet[2802]: I0813 07:13:16.499862 2802 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 07:13:16.501667 kubelet[2802]: W0813 07:13:16.500861 2802 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 07:13:16.507219 kubelet[2802]: I0813 07:13:16.507182 2802 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 07:13:16.507331 kubelet[2802]: I0813 07:13:16.507249 2802 server.go:1289] "Started kubelet" Aug 13 07:13:16.512825 kubelet[2802]: I0813 07:13:16.512768 2802 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:13:16.518505 kubelet[2802]: I0813 07:13:16.517357 2802 server.go:317] "Adding debug handlers to kubelet server" Aug 13 07:13:16.518505 kubelet[2802]: I0813 07:13:16.517939 2802 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:13:16.518505 kubelet[2802]: I0813 07:13:16.518286 2802 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:13:16.521755 kubelet[2802]: E0813 07:13:16.518417 2802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.16.138:6443/api/v1/namespaces/default/events\": dial tcp 172.31.16.138:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-16-138.185b4222ae0af2f3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-16-138,UID:ip-172-31-16-138,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-16-138,},FirstTimestamp:2025-08-13 07:13:16.507214579 +0000 UTC m=+0.479013891,LastTimestamp:2025-08-13 07:13:16.507214579 +0000 UTC m=+0.479013891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-16-138,}" Aug 13 07:13:16.523751 kubelet[2802]: I0813 07:13:16.523571 2802 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:13:16.530515 kubelet[2802]: I0813 07:13:16.529667 2802 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:13:16.532657 kubelet[2802]: I0813 07:13:16.532626 2802 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 07:13:16.532985 kubelet[2802]: E0813 07:13:16.532964 2802 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-16-138\" not found" Aug 13 07:13:16.536676 kubelet[2802]: I0813 07:13:16.536635 2802 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 07:13:16.536845 kubelet[2802]: I0813 07:13:16.536696 2802 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:13:16.540789 kubelet[2802]: I0813 07:13:16.540765 2802 factory.go:223] Registration of the systemd container factory successfully Aug 13 07:13:16.540949 kubelet[2802]: I0813 07:13:16.540873 2802 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:13:16.541253 kubelet[2802]: E0813 07:13:16.541229 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.16.138:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Aug 13 07:13:16.541402 kubelet[2802]: E0813 07:13:16.541326 2802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.16.138:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-16-138?timeout=10s\": dial tcp 172.31.16.138:6443: connect: connection refused" interval="200ms" Aug 13 07:13:16.546809 kubelet[2802]: I0813 07:13:16.545956 2802 factory.go:223] Registration of the containerd container factory successfully Aug 13 07:13:16.565305 kubelet[2802]: I0813 07:13:16.565250 2802 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 07:13:16.567898 kubelet[2802]: I0813 07:13:16.566855 2802 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 07:13:16.567898 kubelet[2802]: I0813 07:13:16.566881 2802 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 07:13:16.567898 kubelet[2802]: I0813 07:13:16.566909 2802 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 07:13:16.567898 kubelet[2802]: I0813 07:13:16.566920 2802 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 07:13:16.567898 kubelet[2802]: E0813 07:13:16.566968 2802 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:13:16.576419 kubelet[2802]: E0813 07:13:16.576385 2802 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:13:16.577409 kubelet[2802]: E0813 07:13:16.577363 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.16.138:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Aug 13 07:13:16.594160 kubelet[2802]: I0813 07:13:16.593936 2802 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 07:13:16.594160 kubelet[2802]: I0813 07:13:16.593951 2802 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 07:13:16.594160 kubelet[2802]: I0813 07:13:16.593968 2802 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:13:16.598237 kubelet[2802]: I0813 07:13:16.597960 2802 policy_none.go:49] "None policy: Start" Aug 13 07:13:16.598237 kubelet[2802]: I0813 07:13:16.597999 2802 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 07:13:16.598237 kubelet[2802]: I0813 07:13:16.598014 2802 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:13:16.609916 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Aug 13 07:13:16.623571 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Aug 13 07:13:16.633752 kubelet[2802]: E0813 07:13:16.633677 2802 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ip-172-31-16-138\" not found" Aug 13 07:13:16.636170 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Aug 13 07:13:16.638993 kubelet[2802]: E0813 07:13:16.638392 2802 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 07:13:16.638993 kubelet[2802]: I0813 07:13:16.638633 2802 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:13:16.638993 kubelet[2802]: I0813 07:13:16.638646 2802 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:13:16.638993 kubelet[2802]: I0813 07:13:16.638897 2802 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:13:16.641849 kubelet[2802]: E0813 07:13:16.641819 2802 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 07:13:16.641981 kubelet[2802]: E0813 07:13:16.641875 2802 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-16-138\" not found" Aug 13 07:13:16.680889 systemd[1]: Created slice kubepods-burstable-pod9e680469fcafab1b016eee57455e5fba.slice - libcontainer container kubepods-burstable-pod9e680469fcafab1b016eee57455e5fba.slice. Aug 13 07:13:16.687862 kubelet[2802]: E0813 07:13:16.687822 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:16.694198 systemd[1]: Created slice kubepods-burstable-pod85fcb035caa1c4a4ffa6d58b654622c7.slice - libcontainer container kubepods-burstable-pod85fcb035caa1c4a4ffa6d58b654622c7.slice. Aug 13 07:13:16.697532 kubelet[2802]: E0813 07:13:16.697277 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:16.700127 systemd[1]: Created slice kubepods-burstable-pod285c9ec3cdd51491cd921923dc3172ca.slice - libcontainer container kubepods-burstable-pod285c9ec3cdd51491cd921923dc3172ca.slice. Aug 13 07:13:16.702063 kubelet[2802]: E0813 07:13:16.702026 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:16.738765 kubelet[2802]: I0813 07:13:16.738522 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-kubeconfig\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:16.738765 kubelet[2802]: I0813 07:13:16.738562 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:16.738765 kubelet[2802]: I0813 07:13:16.738597 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/285c9ec3cdd51491cd921923dc3172ca-kubeconfig\") pod \"kube-scheduler-ip-172-31-16-138\" (UID: \"285c9ec3cdd51491cd921923dc3172ca\") " pod="kube-system/kube-scheduler-ip-172-31-16-138" Aug 13 07:13:16.738765 kubelet[2802]: I0813 07:13:16.738615 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9e680469fcafab1b016eee57455e5fba-ca-certs\") pod \"kube-apiserver-ip-172-31-16-138\" (UID: \"9e680469fcafab1b016eee57455e5fba\") " pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:16.738765 kubelet[2802]: I0813 07:13:16.738632 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9e680469fcafab1b016eee57455e5fba-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-16-138\" (UID: \"9e680469fcafab1b016eee57455e5fba\") " pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:16.739003 kubelet[2802]: I0813 07:13:16.738648 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-ca-certs\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:16.739003 kubelet[2802]: I0813 07:13:16.738663 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:16.739003 kubelet[2802]: I0813 07:13:16.738678 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9e680469fcafab1b016eee57455e5fba-k8s-certs\") pod \"kube-apiserver-ip-172-31-16-138\" (UID: \"9e680469fcafab1b016eee57455e5fba\") " pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:16.739003 kubelet[2802]: I0813 07:13:16.738694 2802 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-k8s-certs\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:16.740417 kubelet[2802]: I0813 07:13:16.740367 2802 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-16-138" Aug 13 07:13:16.740704 kubelet[2802]: E0813 07:13:16.740680 2802 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.16.138:6443/api/v1/nodes\": dial tcp 172.31.16.138:6443: connect: connection refused" node="ip-172-31-16-138" Aug 13 07:13:16.742194 kubelet[2802]: E0813 07:13:16.742154 2802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.16.138:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-16-138?timeout=10s\": dial tcp 172.31.16.138:6443: connect: connection refused" interval="400ms" Aug 13 07:13:16.943078 kubelet[2802]: I0813 07:13:16.942940 2802 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-16-138" Aug 13 07:13:16.943958 kubelet[2802]: E0813 07:13:16.943392 2802 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.16.138:6443/api/v1/nodes\": dial tcp 172.31.16.138:6443: connect: connection refused" node="ip-172-31-16-138" Aug 13 07:13:16.989792 containerd[1991]: time="2025-08-13T07:13:16.989696317Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-16-138,Uid:9e680469fcafab1b016eee57455e5fba,Namespace:kube-system,Attempt:0,}" Aug 13 07:13:16.998927 containerd[1991]: time="2025-08-13T07:13:16.998584177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-16-138,Uid:85fcb035caa1c4a4ffa6d58b654622c7,Namespace:kube-system,Attempt:0,}" Aug 13 07:13:17.003321 containerd[1991]: time="2025-08-13T07:13:17.003264369Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-16-138,Uid:285c9ec3cdd51491cd921923dc3172ca,Namespace:kube-system,Attempt:0,}" Aug 13 07:13:17.143364 kubelet[2802]: E0813 07:13:17.143301 2802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.16.138:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-16-138?timeout=10s\": dial tcp 172.31.16.138:6443: connect: connection refused" interval="800ms" Aug 13 07:13:17.214224 kubelet[2802]: E0813 07:13:17.213967 2802 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.16.138:6443/api/v1/namespaces/default/events\": dial tcp 172.31.16.138:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-16-138.185b4222ae0af2f3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-16-138,UID:ip-172-31-16-138,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-16-138,},FirstTimestamp:2025-08-13 07:13:16.507214579 +0000 UTC m=+0.479013891,LastTimestamp:2025-08-13 07:13:16.507214579 +0000 UTC m=+0.479013891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-16-138,}" Aug 13 07:13:17.346835 kubelet[2802]: I0813 07:13:17.346795 2802 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-16-138" Aug 13 07:13:17.347194 kubelet[2802]: E0813 07:13:17.347161 2802 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.16.138:6443/api/v1/nodes\": dial tcp 172.31.16.138:6443: connect: connection refused" node="ip-172-31-16-138" Aug 13 07:13:17.396199 kubelet[2802]: E0813 07:13:17.396118 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://172.31.16.138:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Aug 13 07:13:17.462571 kubelet[2802]: E0813 07:13:17.461455 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://172.31.16.138:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Aug 13 07:13:17.462980 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount469152918.mount: Deactivated successfully. Aug 13 07:13:17.479846 containerd[1991]: time="2025-08-13T07:13:17.479112731Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:13:17.481307 containerd[1991]: time="2025-08-13T07:13:17.481247984Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:13:17.484743 containerd[1991]: time="2025-08-13T07:13:17.483710680Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Aug 13 07:13:17.488830 containerd[1991]: time="2025-08-13T07:13:17.488754433Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:13:17.491046 containerd[1991]: time="2025-08-13T07:13:17.490988808Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:13:17.494221 containerd[1991]: time="2025-08-13T07:13:17.494177536Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:13:17.494951 containerd[1991]: time="2025-08-13T07:13:17.494803699Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:13:17.497619 containerd[1991]: time="2025-08-13T07:13:17.497567939Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:13:17.498519 containerd[1991]: time="2025-08-13T07:13:17.498472636Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 499.816781ms" Aug 13 07:13:17.501116 containerd[1991]: time="2025-08-13T07:13:17.501075283Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 511.237244ms" Aug 13 07:13:17.504751 containerd[1991]: time="2025-08-13T07:13:17.504691079Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 501.343696ms" Aug 13 07:13:17.718403 containerd[1991]: time="2025-08-13T07:13:17.718045199Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:17.718403 containerd[1991]: time="2025-08-13T07:13:17.718136114Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:17.718403 containerd[1991]: time="2025-08-13T07:13:17.718162572Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:17.718403 containerd[1991]: time="2025-08-13T07:13:17.718278720Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:17.733671 containerd[1991]: time="2025-08-13T07:13:17.732391270Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:17.733671 containerd[1991]: time="2025-08-13T07:13:17.732458767Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:17.733671 containerd[1991]: time="2025-08-13T07:13:17.732482178Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:17.737853 containerd[1991]: time="2025-08-13T07:13:17.737047340Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:17.737853 containerd[1991]: time="2025-08-13T07:13:17.735687162Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:17.738565 containerd[1991]: time="2025-08-13T07:13:17.738172495Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:17.739658 containerd[1991]: time="2025-08-13T07:13:17.738993947Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:17.740511 containerd[1991]: time="2025-08-13T07:13:17.740059802Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:17.750191 systemd[1]: Started cri-containerd-486c2f31a6fc50342c4b46a19fa7f2ccdaf93af30fb4785f28591d65622e60e3.scope - libcontainer container 486c2f31a6fc50342c4b46a19fa7f2ccdaf93af30fb4785f28591d65622e60e3. Aug 13 07:13:17.793977 systemd[1]: Started cri-containerd-2a000a6a33e265605c7f98fc371406ed77d3f61d7f92f04c69dd36045adc2daf.scope - libcontainer container 2a000a6a33e265605c7f98fc371406ed77d3f61d7f92f04c69dd36045adc2daf. Aug 13 07:13:17.806179 systemd[1]: Started cri-containerd-317ab3aed66b7818d543696747191e7c454b1b4d41ef0f880f4d39e55aacb815.scope - libcontainer container 317ab3aed66b7818d543696747191e7c454b1b4d41ef0f880f4d39e55aacb815. Aug 13 07:13:17.850518 containerd[1991]: time="2025-08-13T07:13:17.850454675Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-16-138,Uid:285c9ec3cdd51491cd921923dc3172ca,Namespace:kube-system,Attempt:0,} returns sandbox id \"486c2f31a6fc50342c4b46a19fa7f2ccdaf93af30fb4785f28591d65622e60e3\"" Aug 13 07:13:17.857481 kubelet[2802]: E0813 07:13:17.857415 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://172.31.16.138:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-16-138&limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Aug 13 07:13:17.862987 containerd[1991]: time="2025-08-13T07:13:17.862931301Z" level=info msg="CreateContainer within sandbox \"486c2f31a6fc50342c4b46a19fa7f2ccdaf93af30fb4785f28591d65622e60e3\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 07:13:17.899493 containerd[1991]: time="2025-08-13T07:13:17.899360329Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-16-138,Uid:9e680469fcafab1b016eee57455e5fba,Namespace:kube-system,Attempt:0,} returns sandbox id \"2a000a6a33e265605c7f98fc371406ed77d3f61d7f92f04c69dd36045adc2daf\"" Aug 13 07:13:17.902100 containerd[1991]: time="2025-08-13T07:13:17.901957312Z" level=info msg="CreateContainer within sandbox \"486c2f31a6fc50342c4b46a19fa7f2ccdaf93af30fb4785f28591d65622e60e3\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"3b3e714bb4ab69b7ddbbfc5f25d87099e76151e73d6d0817987ebaf5cbcaa2e1\"" Aug 13 07:13:17.904751 containerd[1991]: time="2025-08-13T07:13:17.903931115Z" level=info msg="StartContainer for \"3b3e714bb4ab69b7ddbbfc5f25d87099e76151e73d6d0817987ebaf5cbcaa2e1\"" Aug 13 07:13:17.909702 containerd[1991]: time="2025-08-13T07:13:17.909664906Z" level=info msg="CreateContainer within sandbox \"2a000a6a33e265605c7f98fc371406ed77d3f61d7f92f04c69dd36045adc2daf\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 07:13:17.918834 containerd[1991]: time="2025-08-13T07:13:17.918782897Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-16-138,Uid:85fcb035caa1c4a4ffa6d58b654622c7,Namespace:kube-system,Attempt:0,} returns sandbox id \"317ab3aed66b7818d543696747191e7c454b1b4d41ef0f880f4d39e55aacb815\"" Aug 13 07:13:17.928104 containerd[1991]: time="2025-08-13T07:13:17.928053102Z" level=info msg="CreateContainer within sandbox \"317ab3aed66b7818d543696747191e7c454b1b4d41ef0f880f4d39e55aacb815\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 07:13:17.945586 kubelet[2802]: E0813 07:13:17.945377 2802 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.16.138:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-16-138?timeout=10s\": dial tcp 172.31.16.138:6443: connect: connection refused" interval="1.6s" Aug 13 07:13:17.946026 containerd[1991]: time="2025-08-13T07:13:17.945988598Z" level=info msg="CreateContainer within sandbox \"2a000a6a33e265605c7f98fc371406ed77d3f61d7f92f04c69dd36045adc2daf\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"27db264ec26ef49c693e8c940c8cac50f55807177c11f68dc924f53f02b593ac\"" Aug 13 07:13:17.947050 containerd[1991]: time="2025-08-13T07:13:17.947001480Z" level=info msg="StartContainer for \"27db264ec26ef49c693e8c940c8cac50f55807177c11f68dc924f53f02b593ac\"" Aug 13 07:13:17.954026 systemd[1]: Started cri-containerd-3b3e714bb4ab69b7ddbbfc5f25d87099e76151e73d6d0817987ebaf5cbcaa2e1.scope - libcontainer container 3b3e714bb4ab69b7ddbbfc5f25d87099e76151e73d6d0817987ebaf5cbcaa2e1. Aug 13 07:13:17.998960 systemd[1]: Started cri-containerd-27db264ec26ef49c693e8c940c8cac50f55807177c11f68dc924f53f02b593ac.scope - libcontainer container 27db264ec26ef49c693e8c940c8cac50f55807177c11f68dc924f53f02b593ac. Aug 13 07:13:18.021449 containerd[1991]: time="2025-08-13T07:13:18.021216444Z" level=info msg="CreateContainer within sandbox \"317ab3aed66b7818d543696747191e7c454b1b4d41ef0f880f4d39e55aacb815\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"dc6be2948eab8185f4901a6c14226a3272f2f0037f9a2eebb8f9ecc64c87c13e\"" Aug 13 07:13:18.022771 containerd[1991]: time="2025-08-13T07:13:18.022490149Z" level=info msg="StartContainer for \"dc6be2948eab8185f4901a6c14226a3272f2f0037f9a2eebb8f9ecc64c87c13e\"" Aug 13 07:13:18.055640 containerd[1991]: time="2025-08-13T07:13:18.055578780Z" level=info msg="StartContainer for \"3b3e714bb4ab69b7ddbbfc5f25d87099e76151e73d6d0817987ebaf5cbcaa2e1\" returns successfully" Aug 13 07:13:18.074476 kubelet[2802]: E0813 07:13:18.074424 2802 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://172.31.16.138:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Aug 13 07:13:18.095263 systemd[1]: Started cri-containerd-dc6be2948eab8185f4901a6c14226a3272f2f0037f9a2eebb8f9ecc64c87c13e.scope - libcontainer container dc6be2948eab8185f4901a6c14226a3272f2f0037f9a2eebb8f9ecc64c87c13e. Aug 13 07:13:18.114891 containerd[1991]: time="2025-08-13T07:13:18.114853897Z" level=info msg="StartContainer for \"27db264ec26ef49c693e8c940c8cac50f55807177c11f68dc924f53f02b593ac\" returns successfully" Aug 13 07:13:18.151572 kubelet[2802]: I0813 07:13:18.151507 2802 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-16-138" Aug 13 07:13:18.152276 kubelet[2802]: E0813 07:13:18.152134 2802 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://172.31.16.138:6443/api/v1/nodes\": dial tcp 172.31.16.138:6443: connect: connection refused" node="ip-172-31-16-138" Aug 13 07:13:18.179341 containerd[1991]: time="2025-08-13T07:13:18.179271044Z" level=info msg="StartContainer for \"dc6be2948eab8185f4901a6c14226a3272f2f0037f9a2eebb8f9ecc64c87c13e\" returns successfully" Aug 13 07:13:18.520342 kubelet[2802]: E0813 07:13:18.520296 2802 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://172.31.16.138:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.16.138:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Aug 13 07:13:18.596404 kubelet[2802]: E0813 07:13:18.595904 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:18.600064 kubelet[2802]: E0813 07:13:18.600032 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:18.601758 kubelet[2802]: E0813 07:13:18.601703 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:19.606500 kubelet[2802]: E0813 07:13:19.606460 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:19.610074 kubelet[2802]: E0813 07:13:19.610038 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:19.670364 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Aug 13 07:13:19.757397 kubelet[2802]: I0813 07:13:19.756558 2802 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-16-138" Aug 13 07:13:20.607772 kubelet[2802]: E0813 07:13:20.606842 2802 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:21.434537 kubelet[2802]: E0813 07:13:21.434489 2802 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-16-138\" not found" node="ip-172-31-16-138" Aug 13 07:13:21.502356 kubelet[2802]: I0813 07:13:21.502314 2802 apiserver.go:52] "Watching apiserver" Aug 13 07:13:21.522652 kubelet[2802]: I0813 07:13:21.522536 2802 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-16-138" Aug 13 07:13:21.522652 kubelet[2802]: E0813 07:13:21.522576 2802 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"ip-172-31-16-138\": node \"ip-172-31-16-138\" not found" Aug 13 07:13:21.537147 kubelet[2802]: I0813 07:13:21.537113 2802 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 07:13:21.537282 kubelet[2802]: I0813 07:13:21.536984 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:21.549643 kubelet[2802]: E0813 07:13:21.549608 2802 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-16-138\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:21.549643 kubelet[2802]: I0813 07:13:21.549639 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-16-138" Aug 13 07:13:21.553247 kubelet[2802]: E0813 07:13:21.553216 2802 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-16-138\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-16-138" Aug 13 07:13:21.553247 kubelet[2802]: I0813 07:13:21.553245 2802 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:21.555181 kubelet[2802]: E0813 07:13:21.555145 2802 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-16-138\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:23.525218 systemd[1]: Reloading requested from client PID 3084 ('systemctl') (unit session-7.scope)... Aug 13 07:13:23.525242 systemd[1]: Reloading... Aug 13 07:13:23.636756 zram_generator::config[3120]: No configuration found. Aug 13 07:13:23.769138 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:13:23.874867 systemd[1]: Reloading finished in 349 ms. Aug 13 07:13:23.917881 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:13:23.936149 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 07:13:23.936371 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:13:23.944175 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:13:24.185295 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:13:24.192447 (kubelet)[3184]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:13:24.272796 kubelet[3184]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:13:24.272796 kubelet[3184]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 07:13:24.272796 kubelet[3184]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:13:24.272796 kubelet[3184]: I0813 07:13:24.271867 3184 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:13:24.278542 kubelet[3184]: I0813 07:13:24.278491 3184 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 07:13:24.278542 kubelet[3184]: I0813 07:13:24.278531 3184 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:13:24.278875 kubelet[3184]: I0813 07:13:24.278849 3184 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 07:13:24.280444 kubelet[3184]: I0813 07:13:24.280411 3184 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Aug 13 07:13:24.290933 kubelet[3184]: I0813 07:13:24.290895 3184 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:13:24.301802 kubelet[3184]: E0813 07:13:24.300127 3184 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:13:24.301802 kubelet[3184]: I0813 07:13:24.300170 3184 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:13:24.304654 kubelet[3184]: I0813 07:13:24.304538 3184 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:13:24.305716 kubelet[3184]: I0813 07:13:24.305180 3184 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:13:24.305716 kubelet[3184]: I0813 07:13:24.305225 3184 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-16-138","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 07:13:24.305716 kubelet[3184]: I0813 07:13:24.305465 3184 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:13:24.305716 kubelet[3184]: I0813 07:13:24.305480 3184 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 07:13:24.305716 kubelet[3184]: I0813 07:13:24.305536 3184 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:13:24.311409 kubelet[3184]: I0813 07:13:24.311366 3184 kubelet.go:480] "Attempting to sync node with API server" Aug 13 07:13:24.311825 kubelet[3184]: I0813 07:13:24.311691 3184 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:13:24.312810 kubelet[3184]: I0813 07:13:24.312643 3184 kubelet.go:386] "Adding apiserver pod source" Aug 13 07:13:24.312810 kubelet[3184]: I0813 07:13:24.312672 3184 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:13:24.323766 kubelet[3184]: I0813 07:13:24.323056 3184 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 07:13:24.323936 kubelet[3184]: I0813 07:13:24.323850 3184 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 07:13:24.333224 kubelet[3184]: I0813 07:13:24.333159 3184 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 07:13:24.333224 kubelet[3184]: I0813 07:13:24.333213 3184 server.go:1289] "Started kubelet" Aug 13 07:13:24.335084 kubelet[3184]: I0813 07:13:24.334907 3184 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:13:24.335084 kubelet[3184]: I0813 07:13:24.334950 3184 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:13:24.337435 kubelet[3184]: I0813 07:13:24.337406 3184 server.go:317] "Adding debug handlers to kubelet server" Aug 13 07:13:24.350112 kubelet[3184]: I0813 07:13:24.350076 3184 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:13:24.360746 kubelet[3184]: I0813 07:13:24.358992 3184 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:13:24.360746 kubelet[3184]: I0813 07:13:24.359339 3184 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:13:24.361997 kubelet[3184]: I0813 07:13:24.361974 3184 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 07:13:24.371085 kubelet[3184]: I0813 07:13:24.371058 3184 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 07:13:24.371878 kubelet[3184]: I0813 07:13:24.371638 3184 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:13:24.372509 kubelet[3184]: I0813 07:13:24.372484 3184 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:13:24.381905 kubelet[3184]: I0813 07:13:24.381871 3184 factory.go:223] Registration of the containerd container factory successfully Aug 13 07:13:24.382200 kubelet[3184]: I0813 07:13:24.382178 3184 factory.go:223] Registration of the systemd container factory successfully Aug 13 07:13:24.388617 kubelet[3184]: E0813 07:13:24.388112 3184 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:13:24.407401 kubelet[3184]: I0813 07:13:24.407328 3184 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 07:13:24.410931 kubelet[3184]: I0813 07:13:24.410901 3184 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 07:13:24.411534 kubelet[3184]: I0813 07:13:24.411508 3184 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 07:13:24.411704 kubelet[3184]: I0813 07:13:24.411692 3184 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 07:13:24.411900 kubelet[3184]: I0813 07:13:24.411874 3184 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 07:13:24.412097 kubelet[3184]: E0813 07:13:24.412055 3184 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:13:24.473754 kubelet[3184]: I0813 07:13:24.473599 3184 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 07:13:24.473754 kubelet[3184]: I0813 07:13:24.473625 3184 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 07:13:24.473754 kubelet[3184]: I0813 07:13:24.473649 3184 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:13:24.475741 kubelet[3184]: I0813 07:13:24.474946 3184 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 07:13:24.475741 kubelet[3184]: I0813 07:13:24.474969 3184 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 07:13:24.475741 kubelet[3184]: I0813 07:13:24.474992 3184 policy_none.go:49] "None policy: Start" Aug 13 07:13:24.475741 kubelet[3184]: I0813 07:13:24.475007 3184 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 07:13:24.475741 kubelet[3184]: I0813 07:13:24.475023 3184 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:13:24.475741 kubelet[3184]: I0813 07:13:24.475149 3184 state_mem.go:75] "Updated machine memory state" Aug 13 07:13:24.483206 kubelet[3184]: E0813 07:13:24.483169 3184 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 07:13:24.483502 kubelet[3184]: I0813 07:13:24.483408 3184 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:13:24.483502 kubelet[3184]: I0813 07:13:24.483430 3184 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:13:24.484327 kubelet[3184]: I0813 07:13:24.484300 3184 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:13:24.489660 kubelet[3184]: E0813 07:13:24.489623 3184 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 07:13:24.520626 kubelet[3184]: I0813 07:13:24.516205 3184 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-16-138" Aug 13 07:13:24.520626 kubelet[3184]: I0813 07:13:24.519897 3184 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:24.520626 kubelet[3184]: I0813 07:13:24.520545 3184 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:24.598385 kubelet[3184]: I0813 07:13:24.598349 3184 kubelet_node_status.go:75] "Attempting to register node" node="ip-172-31-16-138" Aug 13 07:13:24.610550 kubelet[3184]: I0813 07:13:24.610484 3184 kubelet_node_status.go:124] "Node was previously registered" node="ip-172-31-16-138" Aug 13 07:13:24.611738 kubelet[3184]: I0813 07:13:24.611341 3184 kubelet_node_status.go:78] "Successfully registered node" node="ip-172-31-16-138" Aug 13 07:13:24.674488 kubelet[3184]: I0813 07:13:24.674279 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9e680469fcafab1b016eee57455e5fba-k8s-certs\") pod \"kube-apiserver-ip-172-31-16-138\" (UID: \"9e680469fcafab1b016eee57455e5fba\") " pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:24.674488 kubelet[3184]: I0813 07:13:24.674327 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-ca-certs\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:24.674488 kubelet[3184]: I0813 07:13:24.674349 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:24.674488 kubelet[3184]: I0813 07:13:24.674382 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-k8s-certs\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:24.674831 kubelet[3184]: I0813 07:13:24.674793 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:24.674884 kubelet[3184]: I0813 07:13:24.674853 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/285c9ec3cdd51491cd921923dc3172ca-kubeconfig\") pod \"kube-scheduler-ip-172-31-16-138\" (UID: \"285c9ec3cdd51491cd921923dc3172ca\") " pod="kube-system/kube-scheduler-ip-172-31-16-138" Aug 13 07:13:24.674884 kubelet[3184]: I0813 07:13:24.674873 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9e680469fcafab1b016eee57455e5fba-ca-certs\") pod \"kube-apiserver-ip-172-31-16-138\" (UID: \"9e680469fcafab1b016eee57455e5fba\") " pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:24.675186 kubelet[3184]: I0813 07:13:24.674890 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9e680469fcafab1b016eee57455e5fba-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-16-138\" (UID: \"9e680469fcafab1b016eee57455e5fba\") " pod="kube-system/kube-apiserver-ip-172-31-16-138" Aug 13 07:13:24.675186 kubelet[3184]: I0813 07:13:24.674915 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/85fcb035caa1c4a4ffa6d58b654622c7-kubeconfig\") pod \"kube-controller-manager-ip-172-31-16-138\" (UID: \"85fcb035caa1c4a4ffa6d58b654622c7\") " pod="kube-system/kube-controller-manager-ip-172-31-16-138" Aug 13 07:13:25.316509 kubelet[3184]: I0813 07:13:25.316458 3184 apiserver.go:52] "Watching apiserver" Aug 13 07:13:25.372043 kubelet[3184]: I0813 07:13:25.372002 3184 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 07:13:25.582864 kubelet[3184]: I0813 07:13:25.582537 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-16-138" podStartSLOduration=1.58250242 podStartE2EDuration="1.58250242s" podCreationTimestamp="2025-08-13 07:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:13:25.534524209 +0000 UTC m=+1.329323400" watchObservedRunningTime="2025-08-13 07:13:25.58250242 +0000 UTC m=+1.377301612" Aug 13 07:13:25.604464 kubelet[3184]: I0813 07:13:25.604115 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-16-138" podStartSLOduration=1.604098548 podStartE2EDuration="1.604098548s" podCreationTimestamp="2025-08-13 07:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:13:25.585215591 +0000 UTC m=+1.380014786" watchObservedRunningTime="2025-08-13 07:13:25.604098548 +0000 UTC m=+1.398897720" Aug 13 07:13:25.615929 kubelet[3184]: I0813 07:13:25.615874 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-16-138" podStartSLOduration=1.615859855 podStartE2EDuration="1.615859855s" podCreationTimestamp="2025-08-13 07:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:13:25.605261107 +0000 UTC m=+1.400060291" watchObservedRunningTime="2025-08-13 07:13:25.615859855 +0000 UTC m=+1.410659043" Aug 13 07:13:29.075601 kubelet[3184]: I0813 07:13:29.075548 3184 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 07:13:29.076689 containerd[1991]: time="2025-08-13T07:13:29.076647763Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 07:13:29.077138 kubelet[3184]: I0813 07:13:29.077113 3184 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 07:13:29.718963 systemd[1]: Created slice kubepods-besteffort-pod2346b4be_3bf7_407a_a701_82b905c2a380.slice - libcontainer container kubepods-besteffort-pod2346b4be_3bf7_407a_a701_82b905c2a380.slice. Aug 13 07:13:29.816667 kubelet[3184]: I0813 07:13:29.816512 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2346b4be-3bf7-407a-a701-82b905c2a380-xtables-lock\") pod \"kube-proxy-ctxp4\" (UID: \"2346b4be-3bf7-407a-a701-82b905c2a380\") " pod="kube-system/kube-proxy-ctxp4" Aug 13 07:13:29.816667 kubelet[3184]: I0813 07:13:29.816565 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/2346b4be-3bf7-407a-a701-82b905c2a380-kube-proxy\") pod \"kube-proxy-ctxp4\" (UID: \"2346b4be-3bf7-407a-a701-82b905c2a380\") " pod="kube-system/kube-proxy-ctxp4" Aug 13 07:13:29.816667 kubelet[3184]: I0813 07:13:29.816589 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2346b4be-3bf7-407a-a701-82b905c2a380-lib-modules\") pod \"kube-proxy-ctxp4\" (UID: \"2346b4be-3bf7-407a-a701-82b905c2a380\") " pod="kube-system/kube-proxy-ctxp4" Aug 13 07:13:29.816667 kubelet[3184]: I0813 07:13:29.816610 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzswg\" (UniqueName: \"kubernetes.io/projected/2346b4be-3bf7-407a-a701-82b905c2a380-kube-api-access-xzswg\") pod \"kube-proxy-ctxp4\" (UID: \"2346b4be-3bf7-407a-a701-82b905c2a380\") " pod="kube-system/kube-proxy-ctxp4" Aug 13 07:13:29.925424 kubelet[3184]: E0813 07:13:29.925247 3184 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Aug 13 07:13:29.925424 kubelet[3184]: E0813 07:13:29.925341 3184 projected.go:194] Error preparing data for projected volume kube-api-access-xzswg for pod kube-system/kube-proxy-ctxp4: configmap "kube-root-ca.crt" not found Aug 13 07:13:29.925424 kubelet[3184]: E0813 07:13:29.925421 3184 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2346b4be-3bf7-407a-a701-82b905c2a380-kube-api-access-xzswg podName:2346b4be-3bf7-407a-a701-82b905c2a380 nodeName:}" failed. No retries permitted until 2025-08-13 07:13:30.425401824 +0000 UTC m=+6.220200992 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-xzswg" (UniqueName: "kubernetes.io/projected/2346b4be-3bf7-407a-a701-82b905c2a380-kube-api-access-xzswg") pod "kube-proxy-ctxp4" (UID: "2346b4be-3bf7-407a-a701-82b905c2a380") : configmap "kube-root-ca.crt" not found Aug 13 07:13:30.359397 systemd[1]: Created slice kubepods-besteffort-pod4d6b3b58_943f_461d_88ec_65792ad2cb2a.slice - libcontainer container kubepods-besteffort-pod4d6b3b58_943f_461d_88ec_65792ad2cb2a.slice. Aug 13 07:13:30.421568 kubelet[3184]: I0813 07:13:30.421480 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6262\" (UniqueName: \"kubernetes.io/projected/4d6b3b58-943f-461d-88ec-65792ad2cb2a-kube-api-access-r6262\") pod \"tigera-operator-747864d56d-qjxwk\" (UID: \"4d6b3b58-943f-461d-88ec-65792ad2cb2a\") " pod="tigera-operator/tigera-operator-747864d56d-qjxwk" Aug 13 07:13:30.421568 kubelet[3184]: I0813 07:13:30.421539 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/4d6b3b58-943f-461d-88ec-65792ad2cb2a-var-lib-calico\") pod \"tigera-operator-747864d56d-qjxwk\" (UID: \"4d6b3b58-943f-461d-88ec-65792ad2cb2a\") " pod="tigera-operator/tigera-operator-747864d56d-qjxwk" Aug 13 07:13:30.630327 containerd[1991]: time="2025-08-13T07:13:30.630211321Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ctxp4,Uid:2346b4be-3bf7-407a-a701-82b905c2a380,Namespace:kube-system,Attempt:0,}" Aug 13 07:13:30.663455 containerd[1991]: time="2025-08-13T07:13:30.663047895Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-747864d56d-qjxwk,Uid:4d6b3b58-943f-461d-88ec-65792ad2cb2a,Namespace:tigera-operator,Attempt:0,}" Aug 13 07:13:30.669222 containerd[1991]: time="2025-08-13T07:13:30.669062959Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:30.670061 containerd[1991]: time="2025-08-13T07:13:30.669850129Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:30.670061 containerd[1991]: time="2025-08-13T07:13:30.669879362Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:30.670061 containerd[1991]: time="2025-08-13T07:13:30.669982273Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:30.707005 systemd[1]: Started cri-containerd-70fca4648fbdf68d556a17a8f81c79e0d439543a9e7c76eb4ef59d268aa105ce.scope - libcontainer container 70fca4648fbdf68d556a17a8f81c79e0d439543a9e7c76eb4ef59d268aa105ce. Aug 13 07:13:30.725124 containerd[1991]: time="2025-08-13T07:13:30.724910729Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:30.725124 containerd[1991]: time="2025-08-13T07:13:30.724989212Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:30.725124 containerd[1991]: time="2025-08-13T07:13:30.725026066Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:30.726665 containerd[1991]: time="2025-08-13T07:13:30.725139890Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:30.753158 containerd[1991]: time="2025-08-13T07:13:30.753019130Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ctxp4,Uid:2346b4be-3bf7-407a-a701-82b905c2a380,Namespace:kube-system,Attempt:0,} returns sandbox id \"70fca4648fbdf68d556a17a8f81c79e0d439543a9e7c76eb4ef59d268aa105ce\"" Aug 13 07:13:30.765950 containerd[1991]: time="2025-08-13T07:13:30.765860260Z" level=info msg="CreateContainer within sandbox \"70fca4648fbdf68d556a17a8f81c79e0d439543a9e7c76eb4ef59d268aa105ce\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 07:13:30.774041 systemd[1]: Started cri-containerd-82c797d7bf524d80e78eefa702fa5883838c1da6eee42c50cfb3cb5740b9e790.scope - libcontainer container 82c797d7bf524d80e78eefa702fa5883838c1da6eee42c50cfb3cb5740b9e790. Aug 13 07:13:30.841752 containerd[1991]: time="2025-08-13T07:13:30.841228471Z" level=info msg="CreateContainer within sandbox \"70fca4648fbdf68d556a17a8f81c79e0d439543a9e7c76eb4ef59d268aa105ce\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"42a06cc35f8d0f87e7d164a4191a2f14c7a716f16dab647e6d533a0cefa5dbf1\"" Aug 13 07:13:30.844203 containerd[1991]: time="2025-08-13T07:13:30.843951933Z" level=info msg="StartContainer for \"42a06cc35f8d0f87e7d164a4191a2f14c7a716f16dab647e6d533a0cefa5dbf1\"" Aug 13 07:13:30.853483 containerd[1991]: time="2025-08-13T07:13:30.853431079Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-747864d56d-qjxwk,Uid:4d6b3b58-943f-461d-88ec-65792ad2cb2a,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"82c797d7bf524d80e78eefa702fa5883838c1da6eee42c50cfb3cb5740b9e790\"" Aug 13 07:13:30.858489 containerd[1991]: time="2025-08-13T07:13:30.858439039Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Aug 13 07:13:30.887047 systemd[1]: Started cri-containerd-42a06cc35f8d0f87e7d164a4191a2f14c7a716f16dab647e6d533a0cefa5dbf1.scope - libcontainer container 42a06cc35f8d0f87e7d164a4191a2f14c7a716f16dab647e6d533a0cefa5dbf1. Aug 13 07:13:30.927263 containerd[1991]: time="2025-08-13T07:13:30.927200659Z" level=info msg="StartContainer for \"42a06cc35f8d0f87e7d164a4191a2f14c7a716f16dab647e6d533a0cefa5dbf1\" returns successfully" Aug 13 07:13:31.490657 kubelet[3184]: I0813 07:13:31.490408 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-ctxp4" podStartSLOduration=2.49038603 podStartE2EDuration="2.49038603s" podCreationTimestamp="2025-08-13 07:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:13:31.489790621 +0000 UTC m=+7.284589811" watchObservedRunningTime="2025-08-13 07:13:31.49038603 +0000 UTC m=+7.285185222" Aug 13 07:13:32.251168 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount842163589.mount: Deactivated successfully. Aug 13 07:13:33.241153 containerd[1991]: time="2025-08-13T07:13:33.240401031Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:33.241681 containerd[1991]: time="2025-08-13T07:13:33.241644250Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.3: active requests=0, bytes read=25056543" Aug 13 07:13:33.242522 containerd[1991]: time="2025-08-13T07:13:33.242486676Z" level=info msg="ImageCreate event name:\"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:33.245275 containerd[1991]: time="2025-08-13T07:13:33.245236068Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.3\" with image id \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\", repo tag \"quay.io/tigera/operator:v1.38.3\", repo digest \"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\", size \"25052538\" in 2.386717031s" Aug 13 07:13:33.245275 containerd[1991]: time="2025-08-13T07:13:33.245275800Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\"" Aug 13 07:13:33.247381 containerd[1991]: time="2025-08-13T07:13:33.246408477Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:33.251611 containerd[1991]: time="2025-08-13T07:13:33.251436398Z" level=info msg="CreateContainer within sandbox \"82c797d7bf524d80e78eefa702fa5883838c1da6eee42c50cfb3cb5740b9e790\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Aug 13 07:13:33.270493 containerd[1991]: time="2025-08-13T07:13:33.270436299Z" level=info msg="CreateContainer within sandbox \"82c797d7bf524d80e78eefa702fa5883838c1da6eee42c50cfb3cb5740b9e790\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"6a75fd35f5f4969511d0e9a41210e68097ad3df3e1eeb2cf68b86f240ade3b0e\"" Aug 13 07:13:33.272686 containerd[1991]: time="2025-08-13T07:13:33.271873618Z" level=info msg="StartContainer for \"6a75fd35f5f4969511d0e9a41210e68097ad3df3e1eeb2cf68b86f240ade3b0e\"" Aug 13 07:13:33.314943 systemd[1]: Started cri-containerd-6a75fd35f5f4969511d0e9a41210e68097ad3df3e1eeb2cf68b86f240ade3b0e.scope - libcontainer container 6a75fd35f5f4969511d0e9a41210e68097ad3df3e1eeb2cf68b86f240ade3b0e. Aug 13 07:13:33.350124 containerd[1991]: time="2025-08-13T07:13:33.349974849Z" level=info msg="StartContainer for \"6a75fd35f5f4969511d0e9a41210e68097ad3df3e1eeb2cf68b86f240ade3b0e\" returns successfully" Aug 13 07:13:34.112139 update_engine[1963]: I20250813 07:13:34.112033 1963 update_attempter.cc:509] Updating boot flags... Aug 13 07:13:34.188849 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (3545) Aug 13 07:13:35.681857 kubelet[3184]: I0813 07:13:35.681641 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-747864d56d-qjxwk" podStartSLOduration=3.290969181 podStartE2EDuration="5.681601641s" podCreationTimestamp="2025-08-13 07:13:30 +0000 UTC" firstStartedPulling="2025-08-13 07:13:30.856045668 +0000 UTC m=+6.650844851" lastFinishedPulling="2025-08-13 07:13:33.24667814 +0000 UTC m=+9.041477311" observedRunningTime="2025-08-13 07:13:33.485141429 +0000 UTC m=+9.279940617" watchObservedRunningTime="2025-08-13 07:13:35.681601641 +0000 UTC m=+11.476400835" Aug 13 07:13:40.528957 sudo[2301]: pam_unix(sudo:session): session closed for user root Aug 13 07:13:40.553392 sshd[2298]: pam_unix(sshd:session): session closed for user core Aug 13 07:13:40.560108 systemd[1]: sshd@6-172.31.16.138:22-147.75.109.163:40190.service: Deactivated successfully. Aug 13 07:13:40.566270 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 07:13:40.568470 systemd[1]: session-7.scope: Consumed 6.791s CPU time, 145.8M memory peak, 0B memory swap peak. Aug 13 07:13:40.575434 systemd-logind[1961]: Session 7 logged out. Waiting for processes to exit. Aug 13 07:13:40.582648 systemd-logind[1961]: Removed session 7. Aug 13 07:13:46.053635 systemd[1]: Created slice kubepods-besteffort-pod49b2b521_ef4b_411d_b5ab_411935703227.slice - libcontainer container kubepods-besteffort-pod49b2b521_ef4b_411d_b5ab_411935703227.slice. Aug 13 07:13:46.143869 kubelet[3184]: I0813 07:13:46.143811 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/49b2b521-ef4b-411d-b5ab-411935703227-typha-certs\") pod \"calico-typha-6cb4f746f7-2cwck\" (UID: \"49b2b521-ef4b-411d-b5ab-411935703227\") " pod="calico-system/calico-typha-6cb4f746f7-2cwck" Aug 13 07:13:46.144492 kubelet[3184]: I0813 07:13:46.143920 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p259l\" (UniqueName: \"kubernetes.io/projected/49b2b521-ef4b-411d-b5ab-411935703227-kube-api-access-p259l\") pod \"calico-typha-6cb4f746f7-2cwck\" (UID: \"49b2b521-ef4b-411d-b5ab-411935703227\") " pod="calico-system/calico-typha-6cb4f746f7-2cwck" Aug 13 07:13:46.144492 kubelet[3184]: I0813 07:13:46.143952 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49b2b521-ef4b-411d-b5ab-411935703227-tigera-ca-bundle\") pod \"calico-typha-6cb4f746f7-2cwck\" (UID: \"49b2b521-ef4b-411d-b5ab-411935703227\") " pod="calico-system/calico-typha-6cb4f746f7-2cwck" Aug 13 07:13:46.364505 containerd[1991]: time="2025-08-13T07:13:46.363388916Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6cb4f746f7-2cwck,Uid:49b2b521-ef4b-411d-b5ab-411935703227,Namespace:calico-system,Attempt:0,}" Aug 13 07:13:46.419387 containerd[1991]: time="2025-08-13T07:13:46.419055705Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:46.419387 containerd[1991]: time="2025-08-13T07:13:46.419129342Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:46.419387 containerd[1991]: time="2025-08-13T07:13:46.419145028Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:46.419387 containerd[1991]: time="2025-08-13T07:13:46.419246649Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:46.442752 kubelet[3184]: E0813 07:13:46.439634 3184 reflector.go:200] "Failed to watch" err="failed to list *v1.ConfigMap: configmaps \"cni-config\" is forbidden: User \"system:node:ip-172-31-16-138\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"calico-system\": no relationship found between node 'ip-172-31-16-138' and this object" logger="UnhandledError" reflector="object-\"calico-system\"/\"cni-config\"" type="*v1.ConfigMap" Aug 13 07:13:46.442752 kubelet[3184]: E0813 07:13:46.439967 3184 reflector.go:200] "Failed to watch" err="failed to list *v1.Secret: secrets \"node-certs\" is forbidden: User \"system:node:ip-172-31-16-138\" cannot list resource \"secrets\" in API group \"\" in the namespace \"calico-system\": no relationship found between node 'ip-172-31-16-138' and this object" logger="UnhandledError" reflector="object-\"calico-system\"/\"node-certs\"" type="*v1.Secret" Aug 13 07:13:46.442752 kubelet[3184]: I0813 07:13:46.440958 3184 status_manager.go:895] "Failed to get status for pod" podUID="b0ec4c4b-b560-410c-95a8-bf3d6e92aa75" pod="calico-system/calico-node-fp8ds" err="pods \"calico-node-fp8ds\" is forbidden: User \"system:node:ip-172-31-16-138\" cannot get resource \"pods\" in API group \"\" in the namespace \"calico-system\": no relationship found between node 'ip-172-31-16-138' and this object" Aug 13 07:13:46.446919 systemd[1]: Created slice kubepods-besteffort-podb0ec4c4b_b560_410c_95a8_bf3d6e92aa75.slice - libcontainer container kubepods-besteffort-podb0ec4c4b_b560_410c_95a8_bf3d6e92aa75.slice. Aug 13 07:13:46.512540 systemd[1]: Started cri-containerd-0522f1e406a65e7bec5a78cf95b487f406cbff88bbaf80234af4d6555655f5c0.scope - libcontainer container 0522f1e406a65e7bec5a78cf95b487f406cbff88bbaf80234af4d6555655f5c0. Aug 13 07:13:46.547913 kubelet[3184]: I0813 07:13:46.547875 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-cni-net-dir\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.547913 kubelet[3184]: I0813 07:13:46.547916 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-var-lib-calico\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548086 kubelet[3184]: I0813 07:13:46.547933 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-xtables-lock\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548086 kubelet[3184]: I0813 07:13:46.547949 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-var-run-calico\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548086 kubelet[3184]: I0813 07:13:46.547968 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlwdp\" (UniqueName: \"kubernetes.io/projected/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-kube-api-access-mlwdp\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548086 kubelet[3184]: I0813 07:13:46.547987 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-cni-log-dir\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548086 kubelet[3184]: I0813 07:13:46.548003 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-policysync\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548209 kubelet[3184]: I0813 07:13:46.548022 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-cni-bin-dir\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548209 kubelet[3184]: I0813 07:13:46.548047 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-tigera-ca-bundle\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548209 kubelet[3184]: I0813 07:13:46.548064 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-node-certs\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548209 kubelet[3184]: I0813 07:13:46.548084 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-flexvol-driver-host\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.548209 kubelet[3184]: I0813 07:13:46.548100 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-lib-modules\") pod \"calico-node-fp8ds\" (UID: \"b0ec4c4b-b560-410c-95a8-bf3d6e92aa75\") " pod="calico-system/calico-node-fp8ds" Aug 13 07:13:46.616584 containerd[1991]: time="2025-08-13T07:13:46.615446196Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6cb4f746f7-2cwck,Uid:49b2b521-ef4b-411d-b5ab-411935703227,Namespace:calico-system,Attempt:0,} returns sandbox id \"0522f1e406a65e7bec5a78cf95b487f406cbff88bbaf80234af4d6555655f5c0\"" Aug 13 07:13:46.627615 containerd[1991]: time="2025-08-13T07:13:46.627568256Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Aug 13 07:13:46.666055 kubelet[3184]: E0813 07:13:46.666020 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.666055 kubelet[3184]: W0813 07:13:46.666047 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.666400 kubelet[3184]: E0813 07:13:46.666084 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.670330 kubelet[3184]: E0813 07:13:46.670303 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.670330 kubelet[3184]: W0813 07:13:46.670326 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.670479 kubelet[3184]: E0813 07:13:46.670365 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.726281 kubelet[3184]: E0813 07:13:46.726023 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:46.816279 kubelet[3184]: E0813 07:13:46.816246 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.820813 kubelet[3184]: W0813 07:13:46.816525 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.820813 kubelet[3184]: E0813 07:13:46.816578 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.820813 kubelet[3184]: E0813 07:13:46.819571 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.820813 kubelet[3184]: W0813 07:13:46.819590 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.820813 kubelet[3184]: E0813 07:13:46.819615 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.821476 kubelet[3184]: E0813 07:13:46.821232 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.821476 kubelet[3184]: W0813 07:13:46.821255 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.821476 kubelet[3184]: E0813 07:13:46.821287 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.836903 kubelet[3184]: E0813 07:13:46.836867 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.837087 kubelet[3184]: W0813 07:13:46.837069 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.837186 kubelet[3184]: E0813 07:13:46.837171 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.837616 kubelet[3184]: E0813 07:13:46.837600 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.837712 kubelet[3184]: W0813 07:13:46.837701 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.837825 kubelet[3184]: E0813 07:13:46.837813 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.838129 kubelet[3184]: E0813 07:13:46.838117 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.838214 kubelet[3184]: W0813 07:13:46.838203 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.838290 kubelet[3184]: E0813 07:13:46.838278 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.838542 kubelet[3184]: E0813 07:13:46.838531 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.838627 kubelet[3184]: W0813 07:13:46.838617 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.838701 kubelet[3184]: E0813 07:13:46.838690 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.839971 kubelet[3184]: E0813 07:13:46.839952 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.840093 kubelet[3184]: W0813 07:13:46.840081 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.840185 kubelet[3184]: E0813 07:13:46.840173 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.840502 kubelet[3184]: E0813 07:13:46.840491 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.840632 kubelet[3184]: W0813 07:13:46.840573 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.840632 kubelet[3184]: E0813 07:13:46.840589 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.841277 kubelet[3184]: E0813 07:13:46.840928 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.841277 kubelet[3184]: W0813 07:13:46.840940 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.841277 kubelet[3184]: E0813 07:13:46.840952 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.842070 kubelet[3184]: E0813 07:13:46.841941 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.842070 kubelet[3184]: W0813 07:13:46.841956 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.842070 kubelet[3184]: E0813 07:13:46.841971 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.843007 kubelet[3184]: E0813 07:13:46.842847 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.843007 kubelet[3184]: W0813 07:13:46.842863 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.843007 kubelet[3184]: E0813 07:13:46.842876 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.844133 kubelet[3184]: E0813 07:13:46.843818 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.844133 kubelet[3184]: W0813 07:13:46.843833 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.844133 kubelet[3184]: E0813 07:13:46.843847 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.844970 kubelet[3184]: E0813 07:13:46.844439 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.844970 kubelet[3184]: W0813 07:13:46.844454 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.844970 kubelet[3184]: E0813 07:13:46.844468 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.845962 kubelet[3184]: E0813 07:13:46.845372 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.845962 kubelet[3184]: W0813 07:13:46.845386 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.845962 kubelet[3184]: E0813 07:13:46.845399 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.846537 kubelet[3184]: E0813 07:13:46.846387 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.846537 kubelet[3184]: W0813 07:13:46.846402 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.846537 kubelet[3184]: E0813 07:13:46.846418 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.847943 kubelet[3184]: E0813 07:13:46.847825 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.847943 kubelet[3184]: W0813 07:13:46.847841 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.847943 kubelet[3184]: E0813 07:13:46.847856 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.849567 kubelet[3184]: E0813 07:13:46.848297 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.849567 kubelet[3184]: W0813 07:13:46.848311 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.849567 kubelet[3184]: E0813 07:13:46.848323 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.850268 kubelet[3184]: E0813 07:13:46.850047 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.850268 kubelet[3184]: W0813 07:13:46.850065 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.850268 kubelet[3184]: E0813 07:13:46.850081 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.850711 kubelet[3184]: E0813 07:13:46.850489 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.850711 kubelet[3184]: W0813 07:13:46.850503 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.850711 kubelet[3184]: E0813 07:13:46.850516 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.853032 kubelet[3184]: E0813 07:13:46.852970 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.853032 kubelet[3184]: W0813 07:13:46.852997 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.853389 kubelet[3184]: E0813 07:13:46.853022 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.853389 kubelet[3184]: I0813 07:13:46.853252 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a57f8240-352f-45e2-8291-fe7645c155a3-socket-dir\") pod \"csi-node-driver-782zn\" (UID: \"a57f8240-352f-45e2-8291-fe7645c155a3\") " pod="calico-system/csi-node-driver-782zn" Aug 13 07:13:46.853943 kubelet[3184]: E0813 07:13:46.853752 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.853943 kubelet[3184]: W0813 07:13:46.853768 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.853943 kubelet[3184]: E0813 07:13:46.853784 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.853943 kubelet[3184]: I0813 07:13:46.853825 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/a57f8240-352f-45e2-8291-fe7645c155a3-varrun\") pod \"csi-node-driver-782zn\" (UID: \"a57f8240-352f-45e2-8291-fe7645c155a3\") " pod="calico-system/csi-node-driver-782zn" Aug 13 07:13:46.854487 kubelet[3184]: E0813 07:13:46.854333 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.854487 kubelet[3184]: W0813 07:13:46.854349 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.854487 kubelet[3184]: E0813 07:13:46.854363 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.854487 kubelet[3184]: I0813 07:13:46.854390 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2nlk\" (UniqueName: \"kubernetes.io/projected/a57f8240-352f-45e2-8291-fe7645c155a3-kube-api-access-h2nlk\") pod \"csi-node-driver-782zn\" (UID: \"a57f8240-352f-45e2-8291-fe7645c155a3\") " pod="calico-system/csi-node-driver-782zn" Aug 13 07:13:46.856112 kubelet[3184]: E0813 07:13:46.855221 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.856112 kubelet[3184]: W0813 07:13:46.855239 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.856112 kubelet[3184]: E0813 07:13:46.855253 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.856112 kubelet[3184]: I0813 07:13:46.855401 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a57f8240-352f-45e2-8291-fe7645c155a3-kubelet-dir\") pod \"csi-node-driver-782zn\" (UID: \"a57f8240-352f-45e2-8291-fe7645c155a3\") " pod="calico-system/csi-node-driver-782zn" Aug 13 07:13:46.857245 kubelet[3184]: E0813 07:13:46.856613 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.857245 kubelet[3184]: W0813 07:13:46.856635 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.857245 kubelet[3184]: E0813 07:13:46.856651 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.857875 kubelet[3184]: E0813 07:13:46.857464 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.857875 kubelet[3184]: W0813 07:13:46.857479 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.857875 kubelet[3184]: E0813 07:13:46.857493 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.859447 kubelet[3184]: E0813 07:13:46.858763 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.859447 kubelet[3184]: W0813 07:13:46.858780 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.859447 kubelet[3184]: E0813 07:13:46.858795 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.860688 kubelet[3184]: E0813 07:13:46.859702 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.860688 kubelet[3184]: W0813 07:13:46.859736 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.860688 kubelet[3184]: E0813 07:13:46.859752 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.861116 kubelet[3184]: I0813 07:13:46.860921 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a57f8240-352f-45e2-8291-fe7645c155a3-registration-dir\") pod \"csi-node-driver-782zn\" (UID: \"a57f8240-352f-45e2-8291-fe7645c155a3\") " pod="calico-system/csi-node-driver-782zn" Aug 13 07:13:46.861374 kubelet[3184]: E0813 07:13:46.861252 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.861374 kubelet[3184]: W0813 07:13:46.861265 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.861374 kubelet[3184]: E0813 07:13:46.861280 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.862597 kubelet[3184]: E0813 07:13:46.862389 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.862597 kubelet[3184]: W0813 07:13:46.862405 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.862597 kubelet[3184]: E0813 07:13:46.862420 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.865204 kubelet[3184]: E0813 07:13:46.864973 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.865204 kubelet[3184]: W0813 07:13:46.864994 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.865204 kubelet[3184]: E0813 07:13:46.865014 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.866545 kubelet[3184]: E0813 07:13:46.866419 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.866545 kubelet[3184]: W0813 07:13:46.866437 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.866545 kubelet[3184]: E0813 07:13:46.866454 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.869394 kubelet[3184]: E0813 07:13:46.869271 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.869394 kubelet[3184]: W0813 07:13:46.869316 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.869394 kubelet[3184]: E0813 07:13:46.869335 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.873095 kubelet[3184]: E0813 07:13:46.871865 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.873095 kubelet[3184]: W0813 07:13:46.871892 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.873095 kubelet[3184]: E0813 07:13:46.871912 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.873095 kubelet[3184]: E0813 07:13:46.872307 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.873095 kubelet[3184]: W0813 07:13:46.872320 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.873095 kubelet[3184]: E0813 07:13:46.872335 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.964145 kubelet[3184]: E0813 07:13:46.964103 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.964145 kubelet[3184]: W0813 07:13:46.964138 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.964368 kubelet[3184]: E0813 07:13:46.964174 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.964896 kubelet[3184]: E0813 07:13:46.964701 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.965050 kubelet[3184]: W0813 07:13:46.964898 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.965050 kubelet[3184]: E0813 07:13:46.964922 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.965251 kubelet[3184]: E0813 07:13:46.965241 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.965464 kubelet[3184]: W0813 07:13:46.965254 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.965464 kubelet[3184]: E0813 07:13:46.965268 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.965650 kubelet[3184]: E0813 07:13:46.965629 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.965650 kubelet[3184]: W0813 07:13:46.965649 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.965806 kubelet[3184]: E0813 07:13:46.965666 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.966039 kubelet[3184]: E0813 07:13:46.966018 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.966039 kubelet[3184]: W0813 07:13:46.966037 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.966188 kubelet[3184]: E0813 07:13:46.966051 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.966307 kubelet[3184]: E0813 07:13:46.966286 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.966307 kubelet[3184]: W0813 07:13:46.966305 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.966496 kubelet[3184]: E0813 07:13:46.966319 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.966558 kubelet[3184]: E0813 07:13:46.966550 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.966620 kubelet[3184]: W0813 07:13:46.966561 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.966620 kubelet[3184]: E0813 07:13:46.966574 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.966864 kubelet[3184]: E0813 07:13:46.966841 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.966864 kubelet[3184]: W0813 07:13:46.966863 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.967145 kubelet[3184]: E0813 07:13:46.966876 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.967238 kubelet[3184]: E0813 07:13:46.967220 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.967284 kubelet[3184]: W0813 07:13:46.967239 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.967284 kubelet[3184]: E0813 07:13:46.967253 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.967559 kubelet[3184]: E0813 07:13:46.967537 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.967559 kubelet[3184]: W0813 07:13:46.967555 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.967677 kubelet[3184]: E0813 07:13:46.967572 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.967978 kubelet[3184]: E0813 07:13:46.967957 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.967978 kubelet[3184]: W0813 07:13:46.967976 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.968106 kubelet[3184]: E0813 07:13:46.967990 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.968317 kubelet[3184]: E0813 07:13:46.968298 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.968317 kubelet[3184]: W0813 07:13:46.968314 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.968427 kubelet[3184]: E0813 07:13:46.968327 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.968683 kubelet[3184]: E0813 07:13:46.968664 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.968683 kubelet[3184]: W0813 07:13:46.968681 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.968815 kubelet[3184]: E0813 07:13:46.968694 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.969049 kubelet[3184]: E0813 07:13:46.969030 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.969049 kubelet[3184]: W0813 07:13:46.969048 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.969156 kubelet[3184]: E0813 07:13:46.969061 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.969423 kubelet[3184]: E0813 07:13:46.969403 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.969423 kubelet[3184]: W0813 07:13:46.969421 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.969523 kubelet[3184]: E0813 07:13:46.969435 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.969812 kubelet[3184]: E0813 07:13:46.969753 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.969812 kubelet[3184]: W0813 07:13:46.969777 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.969812 kubelet[3184]: E0813 07:13:46.969790 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.970126 kubelet[3184]: E0813 07:13:46.970107 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.970126 kubelet[3184]: W0813 07:13:46.970125 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.970234 kubelet[3184]: E0813 07:13:46.970138 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.970451 kubelet[3184]: E0813 07:13:46.970432 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.970451 kubelet[3184]: W0813 07:13:46.970449 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.970564 kubelet[3184]: E0813 07:13:46.970462 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.970854 kubelet[3184]: E0813 07:13:46.970834 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.970854 kubelet[3184]: W0813 07:13:46.970852 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.971080 kubelet[3184]: E0813 07:13:46.970865 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.971254 kubelet[3184]: E0813 07:13:46.971234 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.971254 kubelet[3184]: W0813 07:13:46.971252 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.971397 kubelet[3184]: E0813 07:13:46.971266 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.971510 kubelet[3184]: E0813 07:13:46.971483 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.971510 kubelet[3184]: W0813 07:13:46.971508 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.971663 kubelet[3184]: E0813 07:13:46.971521 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.971899 kubelet[3184]: E0813 07:13:46.971875 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.971899 kubelet[3184]: W0813 07:13:46.971896 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.971986 kubelet[3184]: E0813 07:13:46.971910 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.973868 kubelet[3184]: E0813 07:13:46.972410 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.973868 kubelet[3184]: W0813 07:13:46.972423 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.973868 kubelet[3184]: E0813 07:13:46.972467 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.973868 kubelet[3184]: E0813 07:13:46.972839 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.973868 kubelet[3184]: W0813 07:13:46.972851 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.973868 kubelet[3184]: E0813 07:13:46.972888 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.973868 kubelet[3184]: E0813 07:13:46.973219 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.973868 kubelet[3184]: W0813 07:13:46.973229 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.973868 kubelet[3184]: E0813 07:13:46.973242 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:46.984572 kubelet[3184]: E0813 07:13:46.984535 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:46.984572 kubelet[3184]: W0813 07:13:46.984566 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:46.984794 kubelet[3184]: E0813 07:13:46.984591 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:47.662236 kubelet[3184]: E0813 07:13:47.662177 3184 secret.go:189] Couldn't get secret calico-system/node-certs: failed to sync secret cache: timed out waiting for the condition Aug 13 07:13:47.662623 kubelet[3184]: E0813 07:13:47.662305 3184 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-node-certs podName:b0ec4c4b-b560-410c-95a8-bf3d6e92aa75 nodeName:}" failed. No retries permitted until 2025-08-13 07:13:48.162282488 +0000 UTC m=+23.957081664 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-certs" (UniqueName: "kubernetes.io/secret/b0ec4c4b-b560-410c-95a8-bf3d6e92aa75-node-certs") pod "calico-node-fp8ds" (UID: "b0ec4c4b-b560-410c-95a8-bf3d6e92aa75") : failed to sync secret cache: timed out waiting for the condition Aug 13 07:13:47.679414 kubelet[3184]: E0813 07:13:47.679378 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:47.679414 kubelet[3184]: W0813 07:13:47.679411 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:47.679586 kubelet[3184]: E0813 07:13:47.679432 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:47.781018 kubelet[3184]: E0813 07:13:47.780910 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:47.781546 kubelet[3184]: W0813 07:13:47.781076 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:47.781546 kubelet[3184]: E0813 07:13:47.781131 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:47.882678 kubelet[3184]: E0813 07:13:47.882608 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:47.882678 kubelet[3184]: W0813 07:13:47.882810 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:47.882678 kubelet[3184]: E0813 07:13:47.882914 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:47.971744 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1294878344.mount: Deactivated successfully. Aug 13 07:13:47.987219 kubelet[3184]: E0813 07:13:47.987061 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:47.987219 kubelet[3184]: W0813 07:13:47.987091 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:47.987219 kubelet[3184]: E0813 07:13:47.987119 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.088702 kubelet[3184]: E0813 07:13:48.088670 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:48.088702 kubelet[3184]: W0813 07:13:48.088698 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:48.088980 kubelet[3184]: E0813 07:13:48.088734 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.190140 kubelet[3184]: E0813 07:13:48.189979 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:48.190140 kubelet[3184]: W0813 07:13:48.190005 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:48.190140 kubelet[3184]: E0813 07:13:48.190030 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.191075 kubelet[3184]: E0813 07:13:48.190877 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:48.191075 kubelet[3184]: W0813 07:13:48.190896 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:48.191075 kubelet[3184]: E0813 07:13:48.190915 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.191714 kubelet[3184]: E0813 07:13:48.191694 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:48.191806 kubelet[3184]: W0813 07:13:48.191715 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:48.191806 kubelet[3184]: E0813 07:13:48.191787 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.192284 kubelet[3184]: E0813 07:13:48.192228 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:48.192284 kubelet[3184]: W0813 07:13:48.192243 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:48.192284 kubelet[3184]: E0813 07:13:48.192259 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.192713 kubelet[3184]: E0813 07:13:48.192687 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:48.192829 kubelet[3184]: W0813 07:13:48.192711 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:48.192878 kubelet[3184]: E0813 07:13:48.192856 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.203802 kubelet[3184]: E0813 07:13:48.203340 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:48.203802 kubelet[3184]: W0813 07:13:48.203368 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:48.203802 kubelet[3184]: E0813 07:13:48.203393 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:48.257584 containerd[1991]: time="2025-08-13T07:13:48.257443505Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-fp8ds,Uid:b0ec4c4b-b560-410c-95a8-bf3d6e92aa75,Namespace:calico-system,Attempt:0,}" Aug 13 07:13:48.321553 containerd[1991]: time="2025-08-13T07:13:48.321024375Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:48.321553 containerd[1991]: time="2025-08-13T07:13:48.321103116Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:48.321553 containerd[1991]: time="2025-08-13T07:13:48.321127813Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:48.321553 containerd[1991]: time="2025-08-13T07:13:48.321241005Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:48.375539 systemd[1]: Started cri-containerd-89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2.scope - libcontainer container 89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2. Aug 13 07:13:48.415449 kubelet[3184]: E0813 07:13:48.415192 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:48.442097 containerd[1991]: time="2025-08-13T07:13:48.441696616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-fp8ds,Uid:b0ec4c4b-b560-410c-95a8-bf3d6e92aa75,Namespace:calico-system,Attempt:0,} returns sandbox id \"89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2\"" Aug 13 07:13:49.245619 containerd[1991]: time="2025-08-13T07:13:49.245567800Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:49.247829 containerd[1991]: time="2025-08-13T07:13:49.247766775Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.2: active requests=0, bytes read=35233364" Aug 13 07:13:49.249891 containerd[1991]: time="2025-08-13T07:13:49.249830380Z" level=info msg="ImageCreate event name:\"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:49.253895 containerd[1991]: time="2025-08-13T07:13:49.253374195Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:49.254470 containerd[1991]: time="2025-08-13T07:13:49.254425700Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.2\" with image id \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\", size \"35233218\" in 2.626815803s" Aug 13 07:13:49.254564 containerd[1991]: time="2025-08-13T07:13:49.254474172Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\"" Aug 13 07:13:49.274477 containerd[1991]: time="2025-08-13T07:13:49.273854377Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Aug 13 07:13:49.302122 containerd[1991]: time="2025-08-13T07:13:49.302054034Z" level=info msg="CreateContainer within sandbox \"0522f1e406a65e7bec5a78cf95b487f406cbff88bbaf80234af4d6555655f5c0\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Aug 13 07:13:49.345223 containerd[1991]: time="2025-08-13T07:13:49.345171106Z" level=info msg="CreateContainer within sandbox \"0522f1e406a65e7bec5a78cf95b487f406cbff88bbaf80234af4d6555655f5c0\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"aebbfa7270063dad6984592b2764d6ff7bd15b5016a30084154b8e4e93879d42\"" Aug 13 07:13:49.352127 containerd[1991]: time="2025-08-13T07:13:49.351973855Z" level=info msg="StartContainer for \"aebbfa7270063dad6984592b2764d6ff7bd15b5016a30084154b8e4e93879d42\"" Aug 13 07:13:49.421921 systemd[1]: Started cri-containerd-aebbfa7270063dad6984592b2764d6ff7bd15b5016a30084154b8e4e93879d42.scope - libcontainer container aebbfa7270063dad6984592b2764d6ff7bd15b5016a30084154b8e4e93879d42. Aug 13 07:13:49.522576 containerd[1991]: time="2025-08-13T07:13:49.522348841Z" level=info msg="StartContainer for \"aebbfa7270063dad6984592b2764d6ff7bd15b5016a30084154b8e4e93879d42\" returns successfully" Aug 13 07:13:49.673375 kubelet[3184]: I0813 07:13:49.669661 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6cb4f746f7-2cwck" podStartSLOduration=2.030205299 podStartE2EDuration="4.669629965s" podCreationTimestamp="2025-08-13 07:13:45 +0000 UTC" firstStartedPulling="2025-08-13 07:13:46.622307069 +0000 UTC m=+22.417106237" lastFinishedPulling="2025-08-13 07:13:49.26173172 +0000 UTC m=+25.056530903" observedRunningTime="2025-08-13 07:13:49.659646866 +0000 UTC m=+25.454446063" watchObservedRunningTime="2025-08-13 07:13:49.669629965 +0000 UTC m=+25.464429155" Aug 13 07:13:49.685351 kubelet[3184]: E0813 07:13:49.685306 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.685351 kubelet[3184]: W0813 07:13:49.685345 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.685560 kubelet[3184]: E0813 07:13:49.685372 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.685647 kubelet[3184]: E0813 07:13:49.685626 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.685647 kubelet[3184]: W0813 07:13:49.685646 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.685793 kubelet[3184]: E0813 07:13:49.685660 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.685911 kubelet[3184]: E0813 07:13:49.685890 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.685911 kubelet[3184]: W0813 07:13:49.685910 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.686014 kubelet[3184]: E0813 07:13:49.685923 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.686800 kubelet[3184]: E0813 07:13:49.686143 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.686800 kubelet[3184]: W0813 07:13:49.686156 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.686800 kubelet[3184]: E0813 07:13:49.686168 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.686800 kubelet[3184]: E0813 07:13:49.686399 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.686800 kubelet[3184]: W0813 07:13:49.686414 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.686800 kubelet[3184]: E0813 07:13:49.686426 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.687362 kubelet[3184]: E0813 07:13:49.686819 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.687362 kubelet[3184]: W0813 07:13:49.686831 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.687362 kubelet[3184]: E0813 07:13:49.686844 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.687362 kubelet[3184]: E0813 07:13:49.687081 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.687362 kubelet[3184]: W0813 07:13:49.687091 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.687362 kubelet[3184]: E0813 07:13:49.687103 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.687628 kubelet[3184]: E0813 07:13:49.687390 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.687628 kubelet[3184]: W0813 07:13:49.687400 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.687628 kubelet[3184]: E0813 07:13:49.687412 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.688875 kubelet[3184]: E0813 07:13:49.687796 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.688875 kubelet[3184]: W0813 07:13:49.687807 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.688875 kubelet[3184]: E0813 07:13:49.687819 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.689095 kubelet[3184]: E0813 07:13:49.689073 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.689095 kubelet[3184]: W0813 07:13:49.689095 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.689221 kubelet[3184]: E0813 07:13:49.689111 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.689736 kubelet[3184]: E0813 07:13:49.689332 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.689736 kubelet[3184]: W0813 07:13:49.689344 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.689736 kubelet[3184]: E0813 07:13:49.689356 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.689736 kubelet[3184]: E0813 07:13:49.689565 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.689736 kubelet[3184]: W0813 07:13:49.689575 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.689736 kubelet[3184]: E0813 07:13:49.689587 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.690066 kubelet[3184]: E0813 07:13:49.689869 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.690066 kubelet[3184]: W0813 07:13:49.689879 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.690066 kubelet[3184]: E0813 07:13:49.689891 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.690198 kubelet[3184]: E0813 07:13:49.690097 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.690198 kubelet[3184]: W0813 07:13:49.690106 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.690198 kubelet[3184]: E0813 07:13:49.690117 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.690353 kubelet[3184]: E0813 07:13:49.690316 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.690353 kubelet[3184]: W0813 07:13:49.690325 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.690353 kubelet[3184]: E0813 07:13:49.690336 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.702757 kubelet[3184]: E0813 07:13:49.701033 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.702757 kubelet[3184]: W0813 07:13:49.701089 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.702757 kubelet[3184]: E0813 07:13:49.701118 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.704147 kubelet[3184]: E0813 07:13:49.704116 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.704147 kubelet[3184]: W0813 07:13:49.704148 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.704317 kubelet[3184]: E0813 07:13:49.704171 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.705743 kubelet[3184]: E0813 07:13:49.704601 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.705743 kubelet[3184]: W0813 07:13:49.704618 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.705743 kubelet[3184]: E0813 07:13:49.704634 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.705743 kubelet[3184]: E0813 07:13:49.704909 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.705743 kubelet[3184]: W0813 07:13:49.704920 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.705743 kubelet[3184]: E0813 07:13:49.704934 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.706242 kubelet[3184]: E0813 07:13:49.706216 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.706242 kubelet[3184]: W0813 07:13:49.706240 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.706371 kubelet[3184]: E0813 07:13:49.706256 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.706518 kubelet[3184]: E0813 07:13:49.706495 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.706518 kubelet[3184]: W0813 07:13:49.706516 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.706629 kubelet[3184]: E0813 07:13:49.706529 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.707761 kubelet[3184]: E0813 07:13:49.706746 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.707761 kubelet[3184]: W0813 07:13:49.706758 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.707761 kubelet[3184]: E0813 07:13:49.706769 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.710247 kubelet[3184]: E0813 07:13:49.710217 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.710247 kubelet[3184]: W0813 07:13:49.710243 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.710428 kubelet[3184]: E0813 07:13:49.710262 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.710566 kubelet[3184]: E0813 07:13:49.710543 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.710566 kubelet[3184]: W0813 07:13:49.710564 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.710700 kubelet[3184]: E0813 07:13:49.710579 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.728104 kubelet[3184]: E0813 07:13:49.727841 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.728104 kubelet[3184]: W0813 07:13:49.727872 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.728104 kubelet[3184]: E0813 07:13:49.727899 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.732427 kubelet[3184]: E0813 07:13:49.732381 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.732427 kubelet[3184]: W0813 07:13:49.732422 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.732684 kubelet[3184]: E0813 07:13:49.732451 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.733590 kubelet[3184]: E0813 07:13:49.733561 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.733590 kubelet[3184]: W0813 07:13:49.733585 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.733934 kubelet[3184]: E0813 07:13:49.733605 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.733934 kubelet[3184]: E0813 07:13:49.733893 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.735467 kubelet[3184]: W0813 07:13:49.734755 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.735467 kubelet[3184]: E0813 07:13:49.734784 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.735467 kubelet[3184]: E0813 07:13:49.735068 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.735467 kubelet[3184]: W0813 07:13:49.735079 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.735467 kubelet[3184]: E0813 07:13:49.735094 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.738427 kubelet[3184]: E0813 07:13:49.737797 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.738427 kubelet[3184]: W0813 07:13:49.737815 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.738427 kubelet[3184]: E0813 07:13:49.737832 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.739075 kubelet[3184]: E0813 07:13:49.738811 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.739075 kubelet[3184]: W0813 07:13:49.738827 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.739075 kubelet[3184]: E0813 07:13:49.738843 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.739607 kubelet[3184]: E0813 07:13:49.739586 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.739607 kubelet[3184]: W0813 07:13:49.739606 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.739765 kubelet[3184]: E0813 07:13:49.739621 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:49.753453 kubelet[3184]: E0813 07:13:49.753385 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:49.753453 kubelet[3184]: W0813 07:13:49.753415 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:49.753453 kubelet[3184]: E0813 07:13:49.753443 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.418076 kubelet[3184]: E0813 07:13:50.418025 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:50.578658 kubelet[3184]: I0813 07:13:50.578271 3184 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:13:50.593818 containerd[1991]: time="2025-08-13T07:13:50.593747667Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:50.595319 kubelet[3184]: E0813 07:13:50.595285 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.595319 kubelet[3184]: W0813 07:13:50.595312 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.595471 kubelet[3184]: E0813 07:13:50.595333 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.595922 containerd[1991]: time="2025-08-13T07:13:50.595855241Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2: active requests=0, bytes read=4446956" Aug 13 07:13:50.596218 kubelet[3184]: E0813 07:13:50.596190 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.596218 kubelet[3184]: W0813 07:13:50.596211 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.596344 kubelet[3184]: E0813 07:13:50.596228 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.596480 kubelet[3184]: E0813 07:13:50.596466 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.596611 kubelet[3184]: W0813 07:13:50.596524 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.596611 kubelet[3184]: E0813 07:13:50.596551 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.597101 kubelet[3184]: E0813 07:13:50.597001 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.597101 kubelet[3184]: W0813 07:13:50.597013 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.597101 kubelet[3184]: E0813 07:13:50.597023 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.599365 kubelet[3184]: E0813 07:13:50.597438 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.599365 kubelet[3184]: W0813 07:13:50.597449 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.599365 kubelet[3184]: E0813 07:13:50.597459 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.599365 kubelet[3184]: E0813 07:13:50.598959 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.599365 kubelet[3184]: W0813 07:13:50.598975 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.599365 kubelet[3184]: E0813 07:13:50.598991 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.599365 kubelet[3184]: E0813 07:13:50.599201 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.599365 kubelet[3184]: W0813 07:13:50.599210 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.599365 kubelet[3184]: E0813 07:13:50.599218 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.599651 containerd[1991]: time="2025-08-13T07:13:50.598056305Z" level=info msg="ImageCreate event name:\"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:50.599782 kubelet[3184]: E0813 07:13:50.599771 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.599924 kubelet[3184]: W0813 07:13:50.599829 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.599924 kubelet[3184]: E0813 07:13:50.599842 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.600101 kubelet[3184]: E0813 07:13:50.600093 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.600164 kubelet[3184]: W0813 07:13:50.600155 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.600240 kubelet[3184]: E0813 07:13:50.600204 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.600435 kubelet[3184]: E0813 07:13:50.600426 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.600498 kubelet[3184]: W0813 07:13:50.600489 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.600548 kubelet[3184]: E0813 07:13:50.600540 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.600939 kubelet[3184]: E0813 07:13:50.600927 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.601002 kubelet[3184]: W0813 07:13:50.600994 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.601067 kubelet[3184]: E0813 07:13:50.601058 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.601482 kubelet[3184]: E0813 07:13:50.601471 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.601544 kubelet[3184]: W0813 07:13:50.601535 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.601610 kubelet[3184]: E0813 07:13:50.601600 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.602853 kubelet[3184]: E0813 07:13:50.602743 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.602853 kubelet[3184]: W0813 07:13:50.602756 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.602853 kubelet[3184]: E0813 07:13:50.602767 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.603335 kubelet[3184]: E0813 07:13:50.603043 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.603335 kubelet[3184]: W0813 07:13:50.603052 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.603335 kubelet[3184]: E0813 07:13:50.603062 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.603335 kubelet[3184]: E0813 07:13:50.603235 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.603335 kubelet[3184]: W0813 07:13:50.603244 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.603335 kubelet[3184]: E0813 07:13:50.603252 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.608427 containerd[1991]: time="2025-08-13T07:13:50.608168279Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:50.609242 containerd[1991]: time="2025-08-13T07:13:50.609174199Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" with image id \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\", size \"5939619\" in 1.335278367s" Aug 13 07:13:50.609441 containerd[1991]: time="2025-08-13T07:13:50.609341113Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\"" Aug 13 07:13:50.610783 kubelet[3184]: E0813 07:13:50.610761 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.610783 kubelet[3184]: W0813 07:13:50.610782 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.611053 kubelet[3184]: E0813 07:13:50.610800 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.611770 kubelet[3184]: E0813 07:13:50.611190 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.611770 kubelet[3184]: W0813 07:13:50.611210 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.611770 kubelet[3184]: E0813 07:13:50.611223 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.611944 kubelet[3184]: E0813 07:13:50.611923 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.611944 kubelet[3184]: W0813 07:13:50.611944 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.612030 kubelet[3184]: E0813 07:13:50.611955 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.612694 kubelet[3184]: E0813 07:13:50.612314 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.612694 kubelet[3184]: W0813 07:13:50.612327 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.612694 kubelet[3184]: E0813 07:13:50.612337 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.612694 kubelet[3184]: E0813 07:13:50.612506 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.612694 kubelet[3184]: W0813 07:13:50.612512 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.612694 kubelet[3184]: E0813 07:13:50.612533 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.612694 kubelet[3184]: E0813 07:13:50.612670 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.612694 kubelet[3184]: W0813 07:13:50.612675 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.612694 kubelet[3184]: E0813 07:13:50.612681 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.613098 kubelet[3184]: E0813 07:13:50.612873 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.613098 kubelet[3184]: W0813 07:13:50.612879 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.613098 kubelet[3184]: E0813 07:13:50.612886 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.613242 kubelet[3184]: E0813 07:13:50.613198 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.613242 kubelet[3184]: W0813 07:13:50.613223 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.613242 kubelet[3184]: E0813 07:13:50.613231 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.614133 kubelet[3184]: E0813 07:13:50.614113 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.614133 kubelet[3184]: W0813 07:13:50.614130 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.614223 kubelet[3184]: E0813 07:13:50.614142 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.614385 kubelet[3184]: E0813 07:13:50.614366 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.614385 kubelet[3184]: W0813 07:13:50.614379 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.614445 kubelet[3184]: E0813 07:13:50.614398 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.614633 kubelet[3184]: E0813 07:13:50.614609 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.614633 kubelet[3184]: W0813 07:13:50.614629 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.614703 kubelet[3184]: E0813 07:13:50.614642 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.615104 kubelet[3184]: E0813 07:13:50.615082 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.615104 kubelet[3184]: W0813 07:13:50.615098 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.615173 kubelet[3184]: E0813 07:13:50.615108 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.616118 kubelet[3184]: E0813 07:13:50.616086 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.616118 kubelet[3184]: W0813 07:13:50.616101 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.616118 kubelet[3184]: E0813 07:13:50.616112 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.616396 kubelet[3184]: E0813 07:13:50.616383 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.616396 kubelet[3184]: W0813 07:13:50.616394 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.616484 kubelet[3184]: E0813 07:13:50.616402 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.616598 kubelet[3184]: E0813 07:13:50.616582 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.616598 kubelet[3184]: W0813 07:13:50.616594 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.616706 kubelet[3184]: E0813 07:13:50.616601 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.617312 kubelet[3184]: E0813 07:13:50.616776 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.617312 kubelet[3184]: W0813 07:13:50.616782 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.617312 kubelet[3184]: E0813 07:13:50.616789 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.617312 kubelet[3184]: E0813 07:13:50.616947 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.617312 kubelet[3184]: W0813 07:13:50.616953 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.617312 kubelet[3184]: E0813 07:13:50.616960 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.617312 kubelet[3184]: E0813 07:13:50.617240 3184 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 07:13:50.617312 kubelet[3184]: W0813 07:13:50.617258 3184 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 07:13:50.617312 kubelet[3184]: E0813 07:13:50.617269 3184 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 07:13:50.624173 containerd[1991]: time="2025-08-13T07:13:50.624106166Z" level=info msg="CreateContainer within sandbox \"89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Aug 13 07:13:50.670592 containerd[1991]: time="2025-08-13T07:13:50.668646134Z" level=info msg="CreateContainer within sandbox \"89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b\"" Aug 13 07:13:50.670592 containerd[1991]: time="2025-08-13T07:13:50.670397186Z" level=info msg="StartContainer for \"982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b\"" Aug 13 07:13:50.734851 systemd[1]: Started cri-containerd-982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b.scope - libcontainer container 982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b. Aug 13 07:13:50.781973 containerd[1991]: time="2025-08-13T07:13:50.781239175Z" level=info msg="StartContainer for \"982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b\" returns successfully" Aug 13 07:13:50.792992 systemd[1]: cri-containerd-982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b.scope: Deactivated successfully. Aug 13 07:13:50.828552 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b-rootfs.mount: Deactivated successfully. Aug 13 07:13:50.932268 containerd[1991]: time="2025-08-13T07:13:50.910705910Z" level=info msg="shim disconnected" id=982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b namespace=k8s.io Aug 13 07:13:50.932268 containerd[1991]: time="2025-08-13T07:13:50.931886157Z" level=warning msg="cleaning up after shim disconnected" id=982c18ba04137a65a424ec9cf57ec7e6cb6c5fef23a438b25b3ff70cec51031b namespace=k8s.io Aug 13 07:13:50.932268 containerd[1991]: time="2025-08-13T07:13:50.931908378Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:13:50.949927 containerd[1991]: time="2025-08-13T07:13:50.949863309Z" level=warning msg="cleanup warnings time=\"2025-08-13T07:13:50Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Aug 13 07:13:51.579930 containerd[1991]: time="2025-08-13T07:13:51.579885407Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Aug 13 07:13:52.414305 kubelet[3184]: E0813 07:13:52.414154 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:54.458914 kubelet[3184]: E0813 07:13:54.458857 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:55.939569 containerd[1991]: time="2025-08-13T07:13:55.939493746Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:55.941802 containerd[1991]: time="2025-08-13T07:13:55.941476040Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.2: active requests=0, bytes read=70436221" Aug 13 07:13:55.945319 containerd[1991]: time="2025-08-13T07:13:55.945245106Z" level=info msg="ImageCreate event name:\"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:55.948637 containerd[1991]: time="2025-08-13T07:13:55.948574893Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:13:55.949685 containerd[1991]: time="2025-08-13T07:13:55.949144230Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.2\" with image id \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\", size \"71928924\" in 4.36922194s" Aug 13 07:13:55.949685 containerd[1991]: time="2025-08-13T07:13:55.949180112Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\"" Aug 13 07:13:55.956211 containerd[1991]: time="2025-08-13T07:13:55.956156705Z" level=info msg="CreateContainer within sandbox \"89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Aug 13 07:13:56.014017 containerd[1991]: time="2025-08-13T07:13:56.013940330Z" level=info msg="CreateContainer within sandbox \"89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d\"" Aug 13 07:13:56.015148 containerd[1991]: time="2025-08-13T07:13:56.014744387Z" level=info msg="StartContainer for \"0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d\"" Aug 13 07:13:56.056494 systemd[1]: run-containerd-runc-k8s.io-0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d-runc.KUdOlo.mount: Deactivated successfully. Aug 13 07:13:56.063972 systemd[1]: Started cri-containerd-0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d.scope - libcontainer container 0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d. Aug 13 07:13:56.107328 containerd[1991]: time="2025-08-13T07:13:56.107253307Z" level=info msg="StartContainer for \"0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d\" returns successfully" Aug 13 07:13:56.413953 kubelet[3184]: E0813 07:13:56.413899 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:56.877293 systemd[1]: cri-containerd-0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d.scope: Deactivated successfully. Aug 13 07:13:56.943205 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d-rootfs.mount: Deactivated successfully. Aug 13 07:13:56.972445 kubelet[3184]: I0813 07:13:56.972399 3184 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Aug 13 07:13:57.247405 systemd[1]: Created slice kubepods-burstable-pod9f23824c_ed26_4c1b_bcd8_13d31489a3f2.slice - libcontainer container kubepods-burstable-pod9f23824c_ed26_4c1b_bcd8_13d31489a3f2.slice. Aug 13 07:13:57.262941 systemd[1]: Created slice kubepods-besteffort-podaf80a1b7_708d_43e5_8d99_c39cc31d2f2e.slice - libcontainer container kubepods-besteffort-podaf80a1b7_708d_43e5_8d99_c39cc31d2f2e.slice. Aug 13 07:13:57.273167 kubelet[3184]: I0813 07:13:57.273032 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqkgs\" (UniqueName: \"kubernetes.io/projected/9f23824c-ed26-4c1b-bcd8-13d31489a3f2-kube-api-access-dqkgs\") pod \"coredns-674b8bbfcf-gbtfq\" (UID: \"9f23824c-ed26-4c1b-bcd8-13d31489a3f2\") " pod="kube-system/coredns-674b8bbfcf-gbtfq" Aug 13 07:13:57.273167 kubelet[3184]: I0813 07:13:57.273122 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b69ns\" (UniqueName: \"kubernetes.io/projected/0e7198d4-7857-481c-b5fc-d03f1f0d0953-kube-api-access-b69ns\") pod \"whisker-768487784d-kgpj7\" (UID: \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\") " pod="calico-system/whisker-768487784d-kgpj7" Aug 13 07:13:57.273652 kubelet[3184]: I0813 07:13:57.273445 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-backend-key-pair\") pod \"whisker-768487784d-kgpj7\" (UID: \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\") " pod="calico-system/whisker-768487784d-kgpj7" Aug 13 07:13:57.273652 kubelet[3184]: I0813 07:13:57.273507 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zz5s\" (UniqueName: \"kubernetes.io/projected/af80a1b7-708d-43e5-8d99-c39cc31d2f2e-kube-api-access-7zz5s\") pod \"calico-kube-controllers-5bf57cfb9b-z778g\" (UID: \"af80a1b7-708d-43e5-8d99-c39cc31d2f2e\") " pod="calico-system/calico-kube-controllers-5bf57cfb9b-z778g" Aug 13 07:13:57.273652 kubelet[3184]: I0813 07:13:57.273537 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/0550883e-aad9-45ab-a0d8-997147ab1f77-calico-apiserver-certs\") pod \"calico-apiserver-645bc4db8-2qtgq\" (UID: \"0550883e-aad9-45ab-a0d8-997147ab1f77\") " pod="calico-apiserver/calico-apiserver-645bc4db8-2qtgq" Aug 13 07:13:57.273652 kubelet[3184]: I0813 07:13:57.273596 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-ca-bundle\") pod \"whisker-768487784d-kgpj7\" (UID: \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\") " pod="calico-system/whisker-768487784d-kgpj7" Aug 13 07:13:57.273652 kubelet[3184]: I0813 07:13:57.273626 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af80a1b7-708d-43e5-8d99-c39cc31d2f2e-tigera-ca-bundle\") pod \"calico-kube-controllers-5bf57cfb9b-z778g\" (UID: \"af80a1b7-708d-43e5-8d99-c39cc31d2f2e\") " pod="calico-system/calico-kube-controllers-5bf57cfb9b-z778g" Aug 13 07:13:57.274847 containerd[1991]: time="2025-08-13T07:13:57.274537694Z" level=info msg="shim disconnected" id=0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d namespace=k8s.io Aug 13 07:13:57.276792 kubelet[3184]: I0813 07:13:57.274694 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f23824c-ed26-4c1b-bcd8-13d31489a3f2-config-volume\") pod \"coredns-674b8bbfcf-gbtfq\" (UID: \"9f23824c-ed26-4c1b-bcd8-13d31489a3f2\") " pod="kube-system/coredns-674b8bbfcf-gbtfq" Aug 13 07:13:57.276792 kubelet[3184]: I0813 07:13:57.274809 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drhg5\" (UniqueName: \"kubernetes.io/projected/0550883e-aad9-45ab-a0d8-997147ab1f77-kube-api-access-drhg5\") pod \"calico-apiserver-645bc4db8-2qtgq\" (UID: \"0550883e-aad9-45ab-a0d8-997147ab1f77\") " pod="calico-apiserver/calico-apiserver-645bc4db8-2qtgq" Aug 13 07:13:57.276974 containerd[1991]: time="2025-08-13T07:13:57.274641773Z" level=warning msg="cleaning up after shim disconnected" id=0f7b78f970786e94e1ecfda1a6b1c3c443545cc40979352ca35fad9ff85a813d namespace=k8s.io Aug 13 07:13:57.276974 containerd[1991]: time="2025-08-13T07:13:57.274901990Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:13:57.293156 systemd[1]: Created slice kubepods-besteffort-pod0e7198d4_7857_481c_b5fc_d03f1f0d0953.slice - libcontainer container kubepods-besteffort-pod0e7198d4_7857_481c_b5fc_d03f1f0d0953.slice. Aug 13 07:13:57.309556 systemd[1]: Created slice kubepods-burstable-pod6b61b836_afb1_4539_8d8f_b832203fca86.slice - libcontainer container kubepods-burstable-pod6b61b836_afb1_4539_8d8f_b832203fca86.slice. Aug 13 07:13:57.332239 systemd[1]: Created slice kubepods-besteffort-pode3e699d1_02d9_4ab9_afc8_37fa42f38432.slice - libcontainer container kubepods-besteffort-pode3e699d1_02d9_4ab9_afc8_37fa42f38432.slice. Aug 13 07:13:57.345661 systemd[1]: Created slice kubepods-besteffort-pod0550883e_aad9_45ab_a0d8_997147ab1f77.slice - libcontainer container kubepods-besteffort-pod0550883e_aad9_45ab_a0d8_997147ab1f77.slice. Aug 13 07:13:57.351369 systemd[1]: Created slice kubepods-besteffort-podb381d26d_9591_43bd_8a36_a31d1d40c8e1.slice - libcontainer container kubepods-besteffort-podb381d26d_9591_43bd_8a36_a31d1d40c8e1.slice. Aug 13 07:13:57.375806 kubelet[3184]: I0813 07:13:57.375672 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/b381d26d-9591-43bd-8a36-a31d1d40c8e1-goldmane-key-pair\") pod \"goldmane-768f4c5c69-lw9wb\" (UID: \"b381d26d-9591-43bd-8a36-a31d1d40c8e1\") " pod="calico-system/goldmane-768f4c5c69-lw9wb" Aug 13 07:13:57.376087 kubelet[3184]: I0813 07:13:57.375897 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b381d26d-9591-43bd-8a36-a31d1d40c8e1-goldmane-ca-bundle\") pod \"goldmane-768f4c5c69-lw9wb\" (UID: \"b381d26d-9591-43bd-8a36-a31d1d40c8e1\") " pod="calico-system/goldmane-768f4c5c69-lw9wb" Aug 13 07:13:57.376087 kubelet[3184]: I0813 07:13:57.375929 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dnzc\" (UniqueName: \"kubernetes.io/projected/b381d26d-9591-43bd-8a36-a31d1d40c8e1-kube-api-access-8dnzc\") pod \"goldmane-768f4c5c69-lw9wb\" (UID: \"b381d26d-9591-43bd-8a36-a31d1d40c8e1\") " pod="calico-system/goldmane-768f4c5c69-lw9wb" Aug 13 07:13:57.376087 kubelet[3184]: I0813 07:13:57.375995 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx6gz\" (UniqueName: \"kubernetes.io/projected/6b61b836-afb1-4539-8d8f-b832203fca86-kube-api-access-gx6gz\") pod \"coredns-674b8bbfcf-jkxj9\" (UID: \"6b61b836-afb1-4539-8d8f-b832203fca86\") " pod="kube-system/coredns-674b8bbfcf-jkxj9" Aug 13 07:13:57.376087 kubelet[3184]: I0813 07:13:57.376082 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b61b836-afb1-4539-8d8f-b832203fca86-config-volume\") pod \"coredns-674b8bbfcf-jkxj9\" (UID: \"6b61b836-afb1-4539-8d8f-b832203fca86\") " pod="kube-system/coredns-674b8bbfcf-jkxj9" Aug 13 07:13:57.378066 kubelet[3184]: I0813 07:13:57.376274 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b381d26d-9591-43bd-8a36-a31d1d40c8e1-config\") pod \"goldmane-768f4c5c69-lw9wb\" (UID: \"b381d26d-9591-43bd-8a36-a31d1d40c8e1\") " pod="calico-system/goldmane-768f4c5c69-lw9wb" Aug 13 07:13:57.378066 kubelet[3184]: I0813 07:13:57.376317 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/e3e699d1-02d9-4ab9-afc8-37fa42f38432-calico-apiserver-certs\") pod \"calico-apiserver-645bc4db8-wmhdx\" (UID: \"e3e699d1-02d9-4ab9-afc8-37fa42f38432\") " pod="calico-apiserver/calico-apiserver-645bc4db8-wmhdx" Aug 13 07:13:57.378066 kubelet[3184]: I0813 07:13:57.376340 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbllp\" (UniqueName: \"kubernetes.io/projected/e3e699d1-02d9-4ab9-afc8-37fa42f38432-kube-api-access-sbllp\") pod \"calico-apiserver-645bc4db8-wmhdx\" (UID: \"e3e699d1-02d9-4ab9-afc8-37fa42f38432\") " pod="calico-apiserver/calico-apiserver-645bc4db8-wmhdx" Aug 13 07:13:57.554414 containerd[1991]: time="2025-08-13T07:13:57.554201060Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gbtfq,Uid:9f23824c-ed26-4c1b-bcd8-13d31489a3f2,Namespace:kube-system,Attempt:0,}" Aug 13 07:13:57.582365 containerd[1991]: time="2025-08-13T07:13:57.581498073Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf57cfb9b-z778g,Uid:af80a1b7-708d-43e5-8d99-c39cc31d2f2e,Namespace:calico-system,Attempt:0,}" Aug 13 07:13:57.597613 containerd[1991]: time="2025-08-13T07:13:57.597230599Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Aug 13 07:13:57.602592 containerd[1991]: time="2025-08-13T07:13:57.602266053Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-768487784d-kgpj7,Uid:0e7198d4-7857-481c-b5fc-d03f1f0d0953,Namespace:calico-system,Attempt:0,}" Aug 13 07:13:57.622394 containerd[1991]: time="2025-08-13T07:13:57.622090236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-jkxj9,Uid:6b61b836-afb1-4539-8d8f-b832203fca86,Namespace:kube-system,Attempt:0,}" Aug 13 07:13:57.645138 containerd[1991]: time="2025-08-13T07:13:57.645017919Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-wmhdx,Uid:e3e699d1-02d9-4ab9-afc8-37fa42f38432,Namespace:calico-apiserver,Attempt:0,}" Aug 13 07:13:57.669685 containerd[1991]: time="2025-08-13T07:13:57.669645001Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-lw9wb,Uid:b381d26d-9591-43bd-8a36-a31d1d40c8e1,Namespace:calico-system,Attempt:0,}" Aug 13 07:13:57.670814 containerd[1991]: time="2025-08-13T07:13:57.670490090Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-2qtgq,Uid:0550883e-aad9-45ab-a0d8-997147ab1f77,Namespace:calico-apiserver,Attempt:0,}" Aug 13 07:13:58.132430 containerd[1991]: time="2025-08-13T07:13:58.132335442Z" level=error msg="Failed to destroy network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.141592 containerd[1991]: time="2025-08-13T07:13:58.141320348Z" level=error msg="encountered an error cleaning up failed sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.141592 containerd[1991]: time="2025-08-13T07:13:58.141472896Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-2qtgq,Uid:0550883e-aad9-45ab-a0d8-997147ab1f77,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.178985 kubelet[3184]: E0813 07:13:58.144837 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.191637 containerd[1991]: time="2025-08-13T07:13:58.191559812Z" level=error msg="Failed to destroy network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.192202 containerd[1991]: time="2025-08-13T07:13:58.192169478Z" level=error msg="encountered an error cleaning up failed sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.193444 containerd[1991]: time="2025-08-13T07:13:58.192359716Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-768487784d-kgpj7,Uid:0e7198d4-7857-481c-b5fc-d03f1f0d0953,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.196292 kubelet[3184]: E0813 07:13:58.196085 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.201497 kubelet[3184]: E0813 07:13:58.201157 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-645bc4db8-2qtgq" Aug 13 07:13:58.201497 kubelet[3184]: E0813 07:13:58.201278 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-645bc4db8-2qtgq" Aug 13 07:13:58.201497 kubelet[3184]: E0813 07:13:58.201375 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-645bc4db8-2qtgq_calico-apiserver(0550883e-aad9-45ab-a0d8-997147ab1f77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-645bc4db8-2qtgq_calico-apiserver(0550883e-aad9-45ab-a0d8-997147ab1f77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-645bc4db8-2qtgq" podUID="0550883e-aad9-45ab-a0d8-997147ab1f77" Aug 13 07:13:58.203638 kubelet[3184]: E0813 07:13:58.201100 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-768487784d-kgpj7" Aug 13 07:13:58.203638 kubelet[3184]: E0813 07:13:58.203136 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-768487784d-kgpj7" Aug 13 07:13:58.203638 kubelet[3184]: E0813 07:13:58.203207 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-768487784d-kgpj7_calico-system(0e7198d4-7857-481c-b5fc-d03f1f0d0953)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-768487784d-kgpj7_calico-system(0e7198d4-7857-481c-b5fc-d03f1f0d0953)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-768487784d-kgpj7" podUID="0e7198d4-7857-481c-b5fc-d03f1f0d0953" Aug 13 07:13:58.203963 containerd[1991]: time="2025-08-13T07:13:58.203483846Z" level=error msg="Failed to destroy network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.205205 containerd[1991]: time="2025-08-13T07:13:58.205122763Z" level=error msg="encountered an error cleaning up failed sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.207406 containerd[1991]: time="2025-08-13T07:13:58.207290703Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-wmhdx,Uid:e3e699d1-02d9-4ab9-afc8-37fa42f38432,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.207901 kubelet[3184]: E0813 07:13:58.207825 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.208182 kubelet[3184]: E0813 07:13:58.208021 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-645bc4db8-wmhdx" Aug 13 07:13:58.208182 kubelet[3184]: E0813 07:13:58.208052 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-645bc4db8-wmhdx" Aug 13 07:13:58.208182 kubelet[3184]: E0813 07:13:58.208122 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-645bc4db8-wmhdx_calico-apiserver(e3e699d1-02d9-4ab9-afc8-37fa42f38432)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-645bc4db8-wmhdx_calico-apiserver(e3e699d1-02d9-4ab9-afc8-37fa42f38432)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-645bc4db8-wmhdx" podUID="e3e699d1-02d9-4ab9-afc8-37fa42f38432" Aug 13 07:13:58.225844 containerd[1991]: time="2025-08-13T07:13:58.225707340Z" level=error msg="Failed to destroy network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.231223 containerd[1991]: time="2025-08-13T07:13:58.231163301Z" level=error msg="encountered an error cleaning up failed sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.231363 containerd[1991]: time="2025-08-13T07:13:58.231244461Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gbtfq,Uid:9f23824c-ed26-4c1b-bcd8-13d31489a3f2,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.231762 kubelet[3184]: E0813 07:13:58.231575 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.231762 kubelet[3184]: E0813 07:13:58.231656 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-gbtfq" Aug 13 07:13:58.231762 kubelet[3184]: E0813 07:13:58.231682 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-gbtfq" Aug 13 07:13:58.231989 containerd[1991]: time="2025-08-13T07:13:58.231699631Z" level=error msg="Failed to destroy network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.232585 kubelet[3184]: E0813 07:13:58.232254 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-gbtfq_kube-system(9f23824c-ed26-4c1b-bcd8-13d31489a3f2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-gbtfq_kube-system(9f23824c-ed26-4c1b-bcd8-13d31489a3f2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-gbtfq" podUID="9f23824c-ed26-4c1b-bcd8-13d31489a3f2" Aug 13 07:13:58.235057 containerd[1991]: time="2025-08-13T07:13:58.235012271Z" level=error msg="encountered an error cleaning up failed sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.235167 containerd[1991]: time="2025-08-13T07:13:58.235084753Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-lw9wb,Uid:b381d26d-9591-43bd-8a36-a31d1d40c8e1,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.235341 kubelet[3184]: E0813 07:13:58.235289 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.235449 kubelet[3184]: E0813 07:13:58.235364 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-768f4c5c69-lw9wb" Aug 13 07:13:58.235449 kubelet[3184]: E0813 07:13:58.235402 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-768f4c5c69-lw9wb" Aug 13 07:13:58.235566 kubelet[3184]: E0813 07:13:58.235467 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-768f4c5c69-lw9wb_calico-system(b381d26d-9591-43bd-8a36-a31d1d40c8e1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-768f4c5c69-lw9wb_calico-system(b381d26d-9591-43bd-8a36-a31d1d40c8e1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-768f4c5c69-lw9wb" podUID="b381d26d-9591-43bd-8a36-a31d1d40c8e1" Aug 13 07:13:58.240342 containerd[1991]: time="2025-08-13T07:13:58.240297919Z" level=error msg="Failed to destroy network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.240673 containerd[1991]: time="2025-08-13T07:13:58.240626559Z" level=error msg="encountered an error cleaning up failed sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.240791 containerd[1991]: time="2025-08-13T07:13:58.240706702Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-jkxj9,Uid:6b61b836-afb1-4539-8d8f-b832203fca86,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.240902 containerd[1991]: time="2025-08-13T07:13:58.240864405Z" level=error msg="Failed to destroy network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.241420 kubelet[3184]: E0813 07:13:58.241372 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.241514 kubelet[3184]: E0813 07:13:58.241443 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-jkxj9" Aug 13 07:13:58.241514 kubelet[3184]: E0813 07:13:58.241471 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-jkxj9" Aug 13 07:13:58.241609 kubelet[3184]: E0813 07:13:58.241539 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-jkxj9_kube-system(6b61b836-afb1-4539-8d8f-b832203fca86)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-jkxj9_kube-system(6b61b836-afb1-4539-8d8f-b832203fca86)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-jkxj9" podUID="6b61b836-afb1-4539-8d8f-b832203fca86" Aug 13 07:13:58.242469 containerd[1991]: time="2025-08-13T07:13:58.242428462Z" level=error msg="encountered an error cleaning up failed sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.242691 containerd[1991]: time="2025-08-13T07:13:58.242504233Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf57cfb9b-z778g,Uid:af80a1b7-708d-43e5-8d99-c39cc31d2f2e,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.244189 kubelet[3184]: E0813 07:13:58.242680 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.244189 kubelet[3184]: E0813 07:13:58.242769 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf57cfb9b-z778g" Aug 13 07:13:58.244189 kubelet[3184]: E0813 07:13:58.242802 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf57cfb9b-z778g" Aug 13 07:13:58.244307 kubelet[3184]: E0813 07:13:58.242855 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bf57cfb9b-z778g_calico-system(af80a1b7-708d-43e5-8d99-c39cc31d2f2e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bf57cfb9b-z778g_calico-system(af80a1b7-708d-43e5-8d99-c39cc31d2f2e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf57cfb9b-z778g" podUID="af80a1b7-708d-43e5-8d99-c39cc31d2f2e" Aug 13 07:13:58.423896 systemd[1]: Created slice kubepods-besteffort-poda57f8240_352f_45e2_8291_fe7645c155a3.slice - libcontainer container kubepods-besteffort-poda57f8240_352f_45e2_8291_fe7645c155a3.slice. Aug 13 07:13:58.430204 containerd[1991]: time="2025-08-13T07:13:58.430154051Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-782zn,Uid:a57f8240-352f-45e2-8291-fe7645c155a3,Namespace:calico-system,Attempt:0,}" Aug 13 07:13:58.511465 containerd[1991]: time="2025-08-13T07:13:58.511376465Z" level=error msg="Failed to destroy network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.512170 containerd[1991]: time="2025-08-13T07:13:58.511848449Z" level=error msg="encountered an error cleaning up failed sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.512170 containerd[1991]: time="2025-08-13T07:13:58.511912541Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-782zn,Uid:a57f8240-352f-45e2-8291-fe7645c155a3,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.512995 kubelet[3184]: E0813 07:13:58.512949 3184 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.513096 kubelet[3184]: E0813 07:13:58.513018 3184 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-782zn" Aug 13 07:13:58.513096 kubelet[3184]: E0813 07:13:58.513046 3184 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-782zn" Aug 13 07:13:58.513180 kubelet[3184]: E0813 07:13:58.513117 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-782zn_calico-system(a57f8240-352f-45e2-8291-fe7645c155a3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-782zn_calico-system(a57f8240-352f-45e2-8291-fe7645c155a3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:58.515987 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9-shm.mount: Deactivated successfully. Aug 13 07:13:58.604491 kubelet[3184]: I0813 07:13:58.604454 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:13:58.606245 kubelet[3184]: I0813 07:13:58.605776 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:13:58.632831 kubelet[3184]: I0813 07:13:58.632766 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:13:58.634890 kubelet[3184]: I0813 07:13:58.634560 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:13:58.636440 kubelet[3184]: I0813 07:13:58.636404 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:13:58.655588 kubelet[3184]: I0813 07:13:58.654707 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:13:58.660499 kubelet[3184]: I0813 07:13:58.660471 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:13:58.667473 containerd[1991]: time="2025-08-13T07:13:58.667414433Z" level=info msg="StopPodSandbox for \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\"" Aug 13 07:13:58.678323 containerd[1991]: time="2025-08-13T07:13:58.668307978Z" level=info msg="StopPodSandbox for \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\"" Aug 13 07:13:58.678738 containerd[1991]: time="2025-08-13T07:13:58.678674836Z" level=info msg="StopPodSandbox for \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\"" Aug 13 07:13:58.683079 containerd[1991]: time="2025-08-13T07:13:58.683027314Z" level=info msg="Ensure that sandbox b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9 in task-service has been cleanup successfully" Aug 13 07:13:58.683515 containerd[1991]: time="2025-08-13T07:13:58.683476065Z" level=info msg="Ensure that sandbox baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873 in task-service has been cleanup successfully" Aug 13 07:13:58.686617 containerd[1991]: time="2025-08-13T07:13:58.682203508Z" level=info msg="StopPodSandbox for \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\"" Aug 13 07:13:58.687286 containerd[1991]: time="2025-08-13T07:13:58.682261289Z" level=info msg="StopPodSandbox for \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\"" Aug 13 07:13:58.688165 containerd[1991]: time="2025-08-13T07:13:58.688052502Z" level=info msg="Ensure that sandbox 1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8 in task-service has been cleanup successfully" Aug 13 07:13:58.709063 containerd[1991]: time="2025-08-13T07:13:58.708183468Z" level=info msg="Ensure that sandbox b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce in task-service has been cleanup successfully" Aug 13 07:13:58.709405 containerd[1991]: time="2025-08-13T07:13:58.682283040Z" level=info msg="StopPodSandbox for \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\"" Aug 13 07:13:58.710563 containerd[1991]: time="2025-08-13T07:13:58.710514501Z" level=info msg="Ensure that sandbox b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69 in task-service has been cleanup successfully" Aug 13 07:13:58.722177 containerd[1991]: time="2025-08-13T07:13:58.682763228Z" level=info msg="StopPodSandbox for \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\"" Aug 13 07:13:58.723779 containerd[1991]: time="2025-08-13T07:13:58.723241118Z" level=info msg="Ensure that sandbox 641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1 in task-service has been cleanup successfully" Aug 13 07:13:58.733085 containerd[1991]: time="2025-08-13T07:13:58.682768149Z" level=info msg="Ensure that sandbox ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea in task-service has been cleanup successfully" Aug 13 07:13:58.768096 kubelet[3184]: I0813 07:13:58.768063 3184 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:13:58.770085 containerd[1991]: time="2025-08-13T07:13:58.770035862Z" level=info msg="StopPodSandbox for \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\"" Aug 13 07:13:58.770285 containerd[1991]: time="2025-08-13T07:13:58.770258770Z" level=info msg="Ensure that sandbox 9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7 in task-service has been cleanup successfully" Aug 13 07:13:58.849920 containerd[1991]: time="2025-08-13T07:13:58.849764133Z" level=error msg="StopPodSandbox for \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\" failed" error="failed to destroy network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.852077 kubelet[3184]: E0813 07:13:58.850093 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:13:58.863973 kubelet[3184]: E0813 07:13:58.852108 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9"} Aug 13 07:13:58.863973 kubelet[3184]: E0813 07:13:58.863854 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"a57f8240-352f-45e2-8291-fe7645c155a3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.863973 kubelet[3184]: E0813 07:13:58.863892 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"a57f8240-352f-45e2-8291-fe7645c155a3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-782zn" podUID="a57f8240-352f-45e2-8291-fe7645c155a3" Aug 13 07:13:58.865454 containerd[1991]: time="2025-08-13T07:13:58.865205586Z" level=error msg="StopPodSandbox for \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\" failed" error="failed to destroy network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.865673 kubelet[3184]: E0813 07:13:58.865458 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:13:58.865673 kubelet[3184]: E0813 07:13:58.865509 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8"} Aug 13 07:13:58.865673 kubelet[3184]: E0813 07:13:58.865566 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"6b61b836-afb1-4539-8d8f-b832203fca86\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.865673 kubelet[3184]: E0813 07:13:58.865604 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"6b61b836-afb1-4539-8d8f-b832203fca86\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-jkxj9" podUID="6b61b836-afb1-4539-8d8f-b832203fca86" Aug 13 07:13:58.897826 containerd[1991]: time="2025-08-13T07:13:58.897672267Z" level=error msg="StopPodSandbox for \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\" failed" error="failed to destroy network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.898459 kubelet[3184]: E0813 07:13:58.897979 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:13:58.898459 kubelet[3184]: E0813 07:13:58.898046 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69"} Aug 13 07:13:58.898459 kubelet[3184]: E0813 07:13:58.898087 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"9f23824c-ed26-4c1b-bcd8-13d31489a3f2\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.898459 kubelet[3184]: E0813 07:13:58.898122 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"9f23824c-ed26-4c1b-bcd8-13d31489a3f2\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-gbtfq" podUID="9f23824c-ed26-4c1b-bcd8-13d31489a3f2" Aug 13 07:13:58.908204 containerd[1991]: time="2025-08-13T07:13:58.908151823Z" level=error msg="StopPodSandbox for \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\" failed" error="failed to destroy network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.913324 kubelet[3184]: E0813 07:13:58.913280 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:13:58.913590 kubelet[3184]: E0813 07:13:58.913556 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873"} Aug 13 07:13:58.913791 kubelet[3184]: E0813 07:13:58.913706 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"0550883e-aad9-45ab-a0d8-997147ab1f77\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.914019 kubelet[3184]: E0813 07:13:58.913994 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"0550883e-aad9-45ab-a0d8-997147ab1f77\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-645bc4db8-2qtgq" podUID="0550883e-aad9-45ab-a0d8-997147ab1f77" Aug 13 07:13:58.924488 containerd[1991]: time="2025-08-13T07:13:58.923821545Z" level=error msg="StopPodSandbox for \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\" failed" error="failed to destroy network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.924648 kubelet[3184]: E0813 07:13:58.924084 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:13:58.924648 kubelet[3184]: E0813 07:13:58.924142 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce"} Aug 13 07:13:58.924648 kubelet[3184]: E0813 07:13:58.924192 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"e3e699d1-02d9-4ab9-afc8-37fa42f38432\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.924648 kubelet[3184]: E0813 07:13:58.924227 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"e3e699d1-02d9-4ab9-afc8-37fa42f38432\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-645bc4db8-wmhdx" podUID="e3e699d1-02d9-4ab9-afc8-37fa42f38432" Aug 13 07:13:58.937876 containerd[1991]: time="2025-08-13T07:13:58.936824709Z" level=error msg="StopPodSandbox for \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\" failed" error="failed to destroy network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.938836 kubelet[3184]: E0813 07:13:58.938627 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:13:58.938997 kubelet[3184]: E0813 07:13:58.938695 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7"} Aug 13 07:13:58.938997 kubelet[3184]: E0813 07:13:58.938953 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"b381d26d-9591-43bd-8a36-a31d1d40c8e1\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.940756 kubelet[3184]: E0813 07:13:58.939083 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"b381d26d-9591-43bd-8a36-a31d1d40c8e1\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-768f4c5c69-lw9wb" podUID="b381d26d-9591-43bd-8a36-a31d1d40c8e1" Aug 13 07:13:58.940756 kubelet[3184]: E0813 07:13:58.940345 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:13:58.940756 kubelet[3184]: E0813 07:13:58.940413 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea"} Aug 13 07:13:58.940756 kubelet[3184]: E0813 07:13:58.940480 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"af80a1b7-708d-43e5-8d99-c39cc31d2f2e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.941029 containerd[1991]: time="2025-08-13T07:13:58.940053088Z" level=error msg="StopPodSandbox for \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\" failed" error="failed to destroy network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.941851 kubelet[3184]: E0813 07:13:58.940514 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"af80a1b7-708d-43e5-8d99-c39cc31d2f2e\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf57cfb9b-z778g" podUID="af80a1b7-708d-43e5-8d99-c39cc31d2f2e" Aug 13 07:13:58.943514 containerd[1991]: time="2025-08-13T07:13:58.943466715Z" level=error msg="StopPodSandbox for \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\" failed" error="failed to destroy network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 07:13:58.943809 kubelet[3184]: E0813 07:13:58.943763 3184 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:13:58.943875 kubelet[3184]: E0813 07:13:58.943810 3184 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1"} Aug 13 07:13:58.943875 kubelet[3184]: E0813 07:13:58.943847 3184 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 07:13:58.943990 kubelet[3184]: E0813 07:13:58.943881 3184 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-768487784d-kgpj7" podUID="0e7198d4-7857-481c-b5fc-d03f1f0d0953" Aug 13 07:14:01.334914 kubelet[3184]: I0813 07:14:01.334829 3184 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:14:06.526630 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2985863284.mount: Deactivated successfully. Aug 13 07:14:06.620805 containerd[1991]: time="2025-08-13T07:14:06.611367571Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.2: active requests=0, bytes read=158500163" Aug 13 07:14:06.623573 containerd[1991]: time="2025-08-13T07:14:06.623442834Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.2\" with image id \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\", size \"158500025\" in 9.026170508s" Aug 13 07:14:06.623573 containerd[1991]: time="2025-08-13T07:14:06.623483027Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\"" Aug 13 07:14:06.626135 containerd[1991]: time="2025-08-13T07:14:06.626050288Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:06.676848 containerd[1991]: time="2025-08-13T07:14:06.675968672Z" level=info msg="ImageCreate event name:\"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:06.676848 containerd[1991]: time="2025-08-13T07:14:06.676783957Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:06.687122 containerd[1991]: time="2025-08-13T07:14:06.687051122Z" level=info msg="CreateContainer within sandbox \"89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Aug 13 07:14:06.781849 containerd[1991]: time="2025-08-13T07:14:06.781684436Z" level=info msg="CreateContainer within sandbox \"89e04320ccfdb69886270b2599a125f53e8d33be1f0faf61a56189220c58f1a2\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"e43e4129f43e4c771fab16dda9150b1a02fac1afe380e68b3a0e72bca1fb6fe8\"" Aug 13 07:14:06.787205 containerd[1991]: time="2025-08-13T07:14:06.787160751Z" level=info msg="StartContainer for \"e43e4129f43e4c771fab16dda9150b1a02fac1afe380e68b3a0e72bca1fb6fe8\"" Aug 13 07:14:06.907984 systemd[1]: Started cri-containerd-e43e4129f43e4c771fab16dda9150b1a02fac1afe380e68b3a0e72bca1fb6fe8.scope - libcontainer container e43e4129f43e4c771fab16dda9150b1a02fac1afe380e68b3a0e72bca1fb6fe8. Aug 13 07:14:06.977755 containerd[1991]: time="2025-08-13T07:14:06.975898010Z" level=info msg="StartContainer for \"e43e4129f43e4c771fab16dda9150b1a02fac1afe380e68b3a0e72bca1fb6fe8\" returns successfully" Aug 13 07:14:07.389394 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Aug 13 07:14:07.389793 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Aug 13 07:14:07.773437 containerd[1991]: time="2025-08-13T07:14:07.773243506Z" level=info msg="StopPodSandbox for \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\"" Aug 13 07:14:07.940470 kubelet[3184]: I0813 07:14:07.935359 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-fp8ds" podStartSLOduration=3.755195698 podStartE2EDuration="21.931216482s" podCreationTimestamp="2025-08-13 07:13:46 +0000 UTC" firstStartedPulling="2025-08-13 07:13:48.448912282 +0000 UTC m=+24.243711457" lastFinishedPulling="2025-08-13 07:14:06.62493307 +0000 UTC m=+42.419732241" observedRunningTime="2025-08-13 07:14:07.87726946 +0000 UTC m=+43.672068651" watchObservedRunningTime="2025-08-13 07:14:07.931216482 +0000 UTC m=+43.726015674" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:07.926 [INFO][4530] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:07.930 [INFO][4530] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" iface="eth0" netns="/var/run/netns/cni-11c5dc59-9539-1a47-2f59-e4db57a6ccea" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:07.931 [INFO][4530] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" iface="eth0" netns="/var/run/netns/cni-11c5dc59-9539-1a47-2f59-e4db57a6ccea" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:07.933 [INFO][4530] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" iface="eth0" netns="/var/run/netns/cni-11c5dc59-9539-1a47-2f59-e4db57a6ccea" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:07.933 [INFO][4530] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:07.933 [INFO][4530] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:08.359 [INFO][4537] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:08.364 [INFO][4537] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:08.364 [INFO][4537] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:08.380 [WARNING][4537] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:08.380 [INFO][4537] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:08.383 [INFO][4537] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:08.391885 containerd[1991]: 2025-08-13 07:14:08.388 [INFO][4530] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:08.398627 systemd[1]: run-netns-cni\x2d11c5dc59\x2d9539\x2d1a47\x2d2f59\x2de4db57a6ccea.mount: Deactivated successfully. Aug 13 07:14:08.423968 containerd[1991]: time="2025-08-13T07:14:08.423654918Z" level=info msg="TearDown network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\" successfully" Aug 13 07:14:08.423968 containerd[1991]: time="2025-08-13T07:14:08.423703372Z" level=info msg="StopPodSandbox for \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\" returns successfully" Aug 13 07:14:08.641600 kubelet[3184]: I0813 07:14:08.641543 3184 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-ca-bundle\") pod \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\" (UID: \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\") " Aug 13 07:14:08.643164 kubelet[3184]: I0813 07:14:08.641657 3184 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b69ns\" (UniqueName: \"kubernetes.io/projected/0e7198d4-7857-481c-b5fc-d03f1f0d0953-kube-api-access-b69ns\") pod \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\" (UID: \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\") " Aug 13 07:14:08.643164 kubelet[3184]: I0813 07:14:08.641690 3184 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-backend-key-pair\") pod \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\" (UID: \"0e7198d4-7857-481c-b5fc-d03f1f0d0953\") " Aug 13 07:14:08.674385 kubelet[3184]: I0813 07:14:08.667120 3184 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "0e7198d4-7857-481c-b5fc-d03f1f0d0953" (UID: "0e7198d4-7857-481c-b5fc-d03f1f0d0953"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Aug 13 07:14:08.692517 kubelet[3184]: I0813 07:14:08.692161 3184 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "0e7198d4-7857-481c-b5fc-d03f1f0d0953" (UID: "0e7198d4-7857-481c-b5fc-d03f1f0d0953"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Aug 13 07:14:08.692300 systemd[1]: var-lib-kubelet-pods-0e7198d4\x2d7857\x2d481c\x2db5fc\x2dd03f1f0d0953-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2db69ns.mount: Deactivated successfully. Aug 13 07:14:08.698450 kubelet[3184]: I0813 07:14:08.696969 3184 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e7198d4-7857-481c-b5fc-d03f1f0d0953-kube-api-access-b69ns" (OuterVolumeSpecName: "kube-api-access-b69ns") pod "0e7198d4-7857-481c-b5fc-d03f1f0d0953" (UID: "0e7198d4-7857-481c-b5fc-d03f1f0d0953"). InnerVolumeSpecName "kube-api-access-b69ns". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 07:14:08.698254 systemd[1]: var-lib-kubelet-pods-0e7198d4\x2d7857\x2d481c\x2db5fc\x2dd03f1f0d0953-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Aug 13 07:14:08.743139 kubelet[3184]: I0813 07:14:08.743051 3184 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-ca-bundle\") on node \"ip-172-31-16-138\" DevicePath \"\"" Aug 13 07:14:08.743139 kubelet[3184]: I0813 07:14:08.743090 3184 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-b69ns\" (UniqueName: \"kubernetes.io/projected/0e7198d4-7857-481c-b5fc-d03f1f0d0953-kube-api-access-b69ns\") on node \"ip-172-31-16-138\" DevicePath \"\"" Aug 13 07:14:08.743139 kubelet[3184]: I0813 07:14:08.743101 3184 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/0e7198d4-7857-481c-b5fc-d03f1f0d0953-whisker-backend-key-pair\") on node \"ip-172-31-16-138\" DevicePath \"\"" Aug 13 07:14:08.875481 systemd[1]: Removed slice kubepods-besteffort-pod0e7198d4_7857_481c_b5fc_d03f1f0d0953.slice - libcontainer container kubepods-besteffort-pod0e7198d4_7857_481c_b5fc_d03f1f0d0953.slice. Aug 13 07:14:09.081765 systemd[1]: Created slice kubepods-besteffort-pod0a4ce8c5_e1af_478f_beca_ececc3618ab5.slice - libcontainer container kubepods-besteffort-pod0a4ce8c5_e1af_478f_beca_ececc3618ab5.slice. Aug 13 07:14:09.148654 kubelet[3184]: I0813 07:14:09.148525 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a4ce8c5-e1af-478f-beca-ececc3618ab5-whisker-ca-bundle\") pod \"whisker-7fbf94ddb4-jrjql\" (UID: \"0a4ce8c5-e1af-478f-beca-ececc3618ab5\") " pod="calico-system/whisker-7fbf94ddb4-jrjql" Aug 13 07:14:09.148654 kubelet[3184]: I0813 07:14:09.148582 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/0a4ce8c5-e1af-478f-beca-ececc3618ab5-whisker-backend-key-pair\") pod \"whisker-7fbf94ddb4-jrjql\" (UID: \"0a4ce8c5-e1af-478f-beca-ececc3618ab5\") " pod="calico-system/whisker-7fbf94ddb4-jrjql" Aug 13 07:14:09.148654 kubelet[3184]: I0813 07:14:09.148599 3184 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hktj5\" (UniqueName: \"kubernetes.io/projected/0a4ce8c5-e1af-478f-beca-ececc3618ab5-kube-api-access-hktj5\") pod \"whisker-7fbf94ddb4-jrjql\" (UID: \"0a4ce8c5-e1af-478f-beca-ececc3618ab5\") " pod="calico-system/whisker-7fbf94ddb4-jrjql" Aug 13 07:14:09.388833 containerd[1991]: time="2025-08-13T07:14:09.387895450Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7fbf94ddb4-jrjql,Uid:0a4ce8c5-e1af-478f-beca-ececc3618ab5,Namespace:calico-system,Attempt:0,}" Aug 13 07:14:09.446310 containerd[1991]: time="2025-08-13T07:14:09.445627387Z" level=info msg="StopPodSandbox for \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\"" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.594 [INFO][4734] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.595 [INFO][4734] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" iface="eth0" netns="/var/run/netns/cni-74076f13-f0e0-7553-08e1-001bcb51f233" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.595 [INFO][4734] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" iface="eth0" netns="/var/run/netns/cni-74076f13-f0e0-7553-08e1-001bcb51f233" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.595 [INFO][4734] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" iface="eth0" netns="/var/run/netns/cni-74076f13-f0e0-7553-08e1-001bcb51f233" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.596 [INFO][4734] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.596 [INFO][4734] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.683 [INFO][4745] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.684 [INFO][4745] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.753 [INFO][4745] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.768 [WARNING][4745] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.769 [INFO][4745] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.771 [INFO][4745] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:09.790845 containerd[1991]: 2025-08-13 07:14:09.778 [INFO][4734] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:09.790845 containerd[1991]: time="2025-08-13T07:14:09.786827243Z" level=info msg="TearDown network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\" successfully" Aug 13 07:14:09.790845 containerd[1991]: time="2025-08-13T07:14:09.786861854Z" level=info msg="StopPodSandbox for \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\" returns successfully" Aug 13 07:14:09.799995 containerd[1991]: time="2025-08-13T07:14:09.797412358Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-2qtgq,Uid:0550883e-aad9-45ab-a0d8-997147ab1f77,Namespace:calico-apiserver,Attempt:1,}" Aug 13 07:14:09.792675 systemd[1]: run-netns-cni\x2d74076f13\x2df0e0\x2d7553\x2d08e1\x2d001bcb51f233.mount: Deactivated successfully. Aug 13 07:14:09.794393 (udev-worker)[4507]: Network interface NamePolicy= disabled on kernel command line. Aug 13 07:14:09.798513 systemd-networkd[1818]: calif17e658e0c0: Link UP Aug 13 07:14:09.798873 systemd-networkd[1818]: calif17e658e0c0: Gained carrier Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.544 [INFO][4716] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.586 [INFO][4716] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0 whisker-7fbf94ddb4- calico-system 0a4ce8c5-e1af-478f-beca-ececc3618ab5 911 0 2025-08-13 07:14:08 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:7fbf94ddb4 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ip-172-31-16-138 whisker-7fbf94ddb4-jrjql eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] calif17e658e0c0 [] [] }} ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.587 [INFO][4716] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.671 [INFO][4747] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" HandleID="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Workload="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.672 [INFO][4747] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" HandleID="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Workload="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003725c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-16-138", "pod":"whisker-7fbf94ddb4-jrjql", "timestamp":"2025-08-13 07:14:09.671760563 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.672 [INFO][4747] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.672 [INFO][4747] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.672 [INFO][4747] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.689 [INFO][4747] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.710 [INFO][4747] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.720 [INFO][4747] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.723 [INFO][4747] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.727 [INFO][4747] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.727 [INFO][4747] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.731 [INFO][4747] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7 Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.739 [INFO][4747] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.753 [INFO][4747] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.1/26] block=192.168.68.0/26 handle="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.753 [INFO][4747] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.1/26] handle="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" host="ip-172-31-16-138" Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.753 [INFO][4747] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:09.850040 containerd[1991]: 2025-08-13 07:14:09.753 [INFO][4747] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.1/26] IPv6=[] ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" HandleID="k8s-pod-network.c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Workload="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" Aug 13 07:14:09.852358 containerd[1991]: 2025-08-13 07:14:09.761 [INFO][4716] cni-plugin/k8s.go 418: Populated endpoint ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0", GenerateName:"whisker-7fbf94ddb4-", Namespace:"calico-system", SelfLink:"", UID:"0a4ce8c5-e1af-478f-beca-ececc3618ab5", ResourceVersion:"911", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 14, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7fbf94ddb4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"whisker-7fbf94ddb4-jrjql", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.68.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif17e658e0c0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:09.852358 containerd[1991]: 2025-08-13 07:14:09.762 [INFO][4716] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.1/32] ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" Aug 13 07:14:09.852358 containerd[1991]: 2025-08-13 07:14:09.762 [INFO][4716] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif17e658e0c0 ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" Aug 13 07:14:09.852358 containerd[1991]: 2025-08-13 07:14:09.789 [INFO][4716] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" Aug 13 07:14:09.852358 containerd[1991]: 2025-08-13 07:14:09.789 [INFO][4716] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0", GenerateName:"whisker-7fbf94ddb4-", Namespace:"calico-system", SelfLink:"", UID:"0a4ce8c5-e1af-478f-beca-ececc3618ab5", ResourceVersion:"911", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 14, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7fbf94ddb4", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7", Pod:"whisker-7fbf94ddb4-jrjql", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.68.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"calif17e658e0c0", MAC:"5e:20:62:d7:d9:e4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:09.852358 containerd[1991]: 2025-08-13 07:14:09.837 [INFO][4716] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7" Namespace="calico-system" Pod="whisker-7fbf94ddb4-jrjql" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--7fbf94ddb4--jrjql-eth0" Aug 13 07:14:09.912147 kernel: bpftool[4808]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Aug 13 07:14:09.960582 containerd[1991]: time="2025-08-13T07:14:09.958205214Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:09.960769 containerd[1991]: time="2025-08-13T07:14:09.960641952Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:09.960769 containerd[1991]: time="2025-08-13T07:14:09.960710964Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:09.965752 containerd[1991]: time="2025-08-13T07:14:09.963911930Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:10.036008 systemd[1]: Started cri-containerd-c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7.scope - libcontainer container c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7. Aug 13 07:14:10.195262 systemd-networkd[1818]: cali84771e8c5e6: Link UP Aug 13 07:14:10.204987 systemd-networkd[1818]: cali84771e8c5e6: Gained carrier Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:09.962 [INFO][4786] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0 calico-apiserver-645bc4db8- calico-apiserver 0550883e-aad9-45ab-a0d8-997147ab1f77 914 0 2025-08-13 07:13:41 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:645bc4db8 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-16-138 calico-apiserver-645bc4db8-2qtgq eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali84771e8c5e6 [] [] }} ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:09.963 [INFO][4786] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.084 [INFO][4850] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" HandleID="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.085 [INFO][4850] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" HandleID="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003fa120), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-16-138", "pod":"calico-apiserver-645bc4db8-2qtgq", "timestamp":"2025-08-13 07:14:10.084194776 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.085 [INFO][4850] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.085 [INFO][4850] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.085 [INFO][4850] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.098 [INFO][4850] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.106 [INFO][4850] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.117 [INFO][4850] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.120 [INFO][4850] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.130 [INFO][4850] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.130 [INFO][4850] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.134 [INFO][4850] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.154 [INFO][4850] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.176 [INFO][4850] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.2/26] block=192.168.68.0/26 handle="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.176 [INFO][4850] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.2/26] handle="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" host="ip-172-31-16-138" Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.176 [INFO][4850] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:10.251162 containerd[1991]: 2025-08-13 07:14:10.176 [INFO][4850] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.2/26] IPv6=[] ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" HandleID="k8s-pod-network.ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:10.252238 containerd[1991]: 2025-08-13 07:14:10.183 [INFO][4786] cni-plugin/k8s.go 418: Populated endpoint ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"0550883e-aad9-45ab-a0d8-997147ab1f77", ResourceVersion:"914", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"calico-apiserver-645bc4db8-2qtgq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali84771e8c5e6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:10.252238 containerd[1991]: 2025-08-13 07:14:10.184 [INFO][4786] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.2/32] ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:10.252238 containerd[1991]: 2025-08-13 07:14:10.184 [INFO][4786] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali84771e8c5e6 ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:10.252238 containerd[1991]: 2025-08-13 07:14:10.201 [INFO][4786] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:10.252238 containerd[1991]: 2025-08-13 07:14:10.202 [INFO][4786] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"0550883e-aad9-45ab-a0d8-997147ab1f77", ResourceVersion:"914", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db", Pod:"calico-apiserver-645bc4db8-2qtgq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali84771e8c5e6", MAC:"be:e0:8d:bb:c0:b9", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:10.252238 containerd[1991]: 2025-08-13 07:14:10.240 [INFO][4786] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-2qtgq" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:10.269747 containerd[1991]: time="2025-08-13T07:14:10.268638901Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7fbf94ddb4-jrjql,Uid:0a4ce8c5-e1af-478f-beca-ececc3618ab5,Namespace:calico-system,Attempt:0,} returns sandbox id \"c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7\"" Aug 13 07:14:10.294067 containerd[1991]: time="2025-08-13T07:14:10.294017689Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Aug 13 07:14:10.343981 containerd[1991]: time="2025-08-13T07:14:10.335412726Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:10.343981 containerd[1991]: time="2025-08-13T07:14:10.335488132Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:10.343981 containerd[1991]: time="2025-08-13T07:14:10.335529833Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:10.343981 containerd[1991]: time="2025-08-13T07:14:10.335638875Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:10.364015 systemd[1]: Started cri-containerd-ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db.scope - libcontainer container ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db. Aug 13 07:14:10.416111 containerd[1991]: time="2025-08-13T07:14:10.416030807Z" level=info msg="StopPodSandbox for \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\"" Aug 13 07:14:10.430008 containerd[1991]: time="2025-08-13T07:14:10.429889760Z" level=info msg="StopPodSandbox for \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\"" Aug 13 07:14:10.509397 kubelet[3184]: I0813 07:14:10.509333 3184 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e7198d4-7857-481c-b5fc-d03f1f0d0953" path="/var/lib/kubelet/pods/0e7198d4-7857-481c-b5fc-d03f1f0d0953/volumes" Aug 13 07:14:10.552409 containerd[1991]: time="2025-08-13T07:14:10.552258482Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-2qtgq,Uid:0550883e-aad9-45ab-a0d8-997147ab1f77,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db\"" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.688 [INFO][4955] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.693 [INFO][4955] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" iface="eth0" netns="/var/run/netns/cni-9aa3b48d-3b02-1298-c6be-64cc92ca1960" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.693 [INFO][4955] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" iface="eth0" netns="/var/run/netns/cni-9aa3b48d-3b02-1298-c6be-64cc92ca1960" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.694 [INFO][4955] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" iface="eth0" netns="/var/run/netns/cni-9aa3b48d-3b02-1298-c6be-64cc92ca1960" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.694 [INFO][4955] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.694 [INFO][4955] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.742 [INFO][4969] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.743 [INFO][4969] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.743 [INFO][4969] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.775 [WARNING][4969] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.776 [INFO][4969] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.801 [INFO][4969] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:10.811978 containerd[1991]: 2025-08-13 07:14:10.805 [INFO][4955] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:10.816155 containerd[1991]: time="2025-08-13T07:14:10.813293043Z" level=info msg="TearDown network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\" successfully" Aug 13 07:14:10.816155 containerd[1991]: time="2025-08-13T07:14:10.813331986Z" level=info msg="StopPodSandbox for \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\" returns successfully" Aug 13 07:14:10.819761 containerd[1991]: time="2025-08-13T07:14:10.818063408Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-jkxj9,Uid:6b61b836-afb1-4539-8d8f-b832203fca86,Namespace:kube-system,Attempt:1,}" Aug 13 07:14:10.821256 systemd[1]: run-netns-cni\x2d9aa3b48d\x2d3b02\x2d1298\x2dc6be\x2d64cc92ca1960.mount: Deactivated successfully. Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.752 [INFO][4945] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.752 [INFO][4945] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" iface="eth0" netns="/var/run/netns/cni-ef93ad16-a709-cfc8-1741-ee458a6b5cbf" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.753 [INFO][4945] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" iface="eth0" netns="/var/run/netns/cni-ef93ad16-a709-cfc8-1741-ee458a6b5cbf" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.753 [INFO][4945] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" iface="eth0" netns="/var/run/netns/cni-ef93ad16-a709-cfc8-1741-ee458a6b5cbf" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.753 [INFO][4945] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.753 [INFO][4945] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.846 [INFO][4978] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.846 [INFO][4978] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.846 [INFO][4978] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.891 [WARNING][4978] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.891 [INFO][4978] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.896 [INFO][4978] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:10.913078 containerd[1991]: 2025-08-13 07:14:10.903 [INFO][4945] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:10.916372 containerd[1991]: time="2025-08-13T07:14:10.913028690Z" level=info msg="TearDown network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\" successfully" Aug 13 07:14:10.916372 containerd[1991]: time="2025-08-13T07:14:10.913849259Z" level=info msg="StopPodSandbox for \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\" returns successfully" Aug 13 07:14:10.920756 containerd[1991]: time="2025-08-13T07:14:10.920532981Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gbtfq,Uid:9f23824c-ed26-4c1b-bcd8-13d31489a3f2,Namespace:kube-system,Attempt:1,}" Aug 13 07:14:10.924502 systemd[1]: run-netns-cni\x2def93ad16\x2da709\x2dcfc8\x2d1741\x2dee458a6b5cbf.mount: Deactivated successfully. Aug 13 07:14:10.990701 systemd-networkd[1818]: vxlan.calico: Link UP Aug 13 07:14:10.990712 systemd-networkd[1818]: vxlan.calico: Gained carrier Aug 13 07:14:11.132231 (udev-worker)[4509]: Network interface NamePolicy= disabled on kernel command line. Aug 13 07:14:11.192923 systemd-networkd[1818]: calia07661b75a4: Link UP Aug 13 07:14:11.196425 systemd-networkd[1818]: calia07661b75a4: Gained carrier Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:10.981 [INFO][4988] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0 coredns-674b8bbfcf- kube-system 6b61b836-afb1-4539-8d8f-b832203fca86 928 0 2025-08-13 07:13:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-16-138 coredns-674b8bbfcf-jkxj9 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calia07661b75a4 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:10.982 [INFO][4988] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.069 [INFO][5003] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" HandleID="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.069 [INFO][5003] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" HandleID="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d1960), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-16-138", "pod":"coredns-674b8bbfcf-jkxj9", "timestamp":"2025-08-13 07:14:11.067649538 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.070 [INFO][5003] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.070 [INFO][5003] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.070 [INFO][5003] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.080 [INFO][5003] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.095 [INFO][5003] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.104 [INFO][5003] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.114 [INFO][5003] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.133 [INFO][5003] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.135 [INFO][5003] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.138 [INFO][5003] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886 Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.150 [INFO][5003] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.162 [INFO][5003] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.3/26] block=192.168.68.0/26 handle="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.163 [INFO][5003] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.3/26] handle="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" host="ip-172-31-16-138" Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.163 [INFO][5003] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:11.238712 containerd[1991]: 2025-08-13 07:14:11.163 [INFO][5003] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.3/26] IPv6=[] ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" HandleID="k8s-pod-network.3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:11.242657 containerd[1991]: 2025-08-13 07:14:11.178 [INFO][4988] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"6b61b836-afb1-4539-8d8f-b832203fca86", ResourceVersion:"928", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"coredns-674b8bbfcf-jkxj9", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia07661b75a4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:11.242657 containerd[1991]: 2025-08-13 07:14:11.178 [INFO][4988] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.3/32] ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:11.242657 containerd[1991]: 2025-08-13 07:14:11.178 [INFO][4988] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia07661b75a4 ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:11.242657 containerd[1991]: 2025-08-13 07:14:11.199 [INFO][4988] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:11.242657 containerd[1991]: 2025-08-13 07:14:11.209 [INFO][4988] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"6b61b836-afb1-4539-8d8f-b832203fca86", ResourceVersion:"928", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886", Pod:"coredns-674b8bbfcf-jkxj9", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia07661b75a4", MAC:"66:d9:d3:45:27:29", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:11.242657 containerd[1991]: 2025-08-13 07:14:11.228 [INFO][4988] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886" Namespace="kube-system" Pod="coredns-674b8bbfcf-jkxj9" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:11.280693 containerd[1991]: time="2025-08-13T07:14:11.280339446Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:11.280693 containerd[1991]: time="2025-08-13T07:14:11.280416954Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:11.280693 containerd[1991]: time="2025-08-13T07:14:11.280458656Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:11.280693 containerd[1991]: time="2025-08-13T07:14:11.280571150Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:11.313783 systemd[1]: Started cri-containerd-3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886.scope - libcontainer container 3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886. Aug 13 07:14:11.341100 systemd-networkd[1818]: cali7cded0b467d: Link UP Aug 13 07:14:11.341453 systemd-networkd[1818]: cali7cded0b467d: Gained carrier Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.069 [INFO][5004] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0 coredns-674b8bbfcf- kube-system 9f23824c-ed26-4c1b-bcd8-13d31489a3f2 929 0 2025-08-13 07:13:30 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-16-138 coredns-674b8bbfcf-gbtfq eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali7cded0b467d [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.069 [INFO][5004] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.204 [INFO][5026] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" HandleID="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.207 [INFO][5026] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" HandleID="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000357ba0), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-16-138", "pod":"coredns-674b8bbfcf-gbtfq", "timestamp":"2025-08-13 07:14:11.204374934 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.207 [INFO][5026] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.208 [INFO][5026] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.208 [INFO][5026] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.235 [INFO][5026] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.249 [INFO][5026] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.260 [INFO][5026] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.266 [INFO][5026] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.273 [INFO][5026] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.273 [INFO][5026] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.278 [INFO][5026] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90 Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.291 [INFO][5026] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.314 [INFO][5026] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.4/26] block=192.168.68.0/26 handle="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.315 [INFO][5026] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.4/26] handle="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" host="ip-172-31-16-138" Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.315 [INFO][5026] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:11.409331 containerd[1991]: 2025-08-13 07:14:11.315 [INFO][5026] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.4/26] IPv6=[] ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" HandleID="k8s-pod-network.9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:11.415377 containerd[1991]: 2025-08-13 07:14:11.330 [INFO][5004] cni-plugin/k8s.go 418: Populated endpoint ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"9f23824c-ed26-4c1b-bcd8-13d31489a3f2", ResourceVersion:"929", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"coredns-674b8bbfcf-gbtfq", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7cded0b467d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:11.415377 containerd[1991]: 2025-08-13 07:14:11.331 [INFO][5004] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.4/32] ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:11.415377 containerd[1991]: 2025-08-13 07:14:11.331 [INFO][5004] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali7cded0b467d ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:11.415377 containerd[1991]: 2025-08-13 07:14:11.341 [INFO][5004] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:11.415377 containerd[1991]: 2025-08-13 07:14:11.345 [INFO][5004] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"9f23824c-ed26-4c1b-bcd8-13d31489a3f2", ResourceVersion:"929", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90", Pod:"coredns-674b8bbfcf-gbtfq", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7cded0b467d", MAC:"e2:b5:c3:70:c6:35", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:11.415377 containerd[1991]: 2025-08-13 07:14:11.392 [INFO][5004] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90" Namespace="kube-system" Pod="coredns-674b8bbfcf-gbtfq" WorkloadEndpoint="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:11.417109 containerd[1991]: time="2025-08-13T07:14:11.415482193Z" level=info msg="StopPodSandbox for \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\"" Aug 13 07:14:11.418557 containerd[1991]: time="2025-08-13T07:14:11.417858035Z" level=info msg="StopPodSandbox for \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\"" Aug 13 07:14:11.440951 systemd-networkd[1818]: calif17e658e0c0: Gained IPv6LL Aug 13 07:14:11.469763 containerd[1991]: time="2025-08-13T07:14:11.469419637Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-jkxj9,Uid:6b61b836-afb1-4539-8d8f-b832203fca86,Namespace:kube-system,Attempt:1,} returns sandbox id \"3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886\"" Aug 13 07:14:11.481505 containerd[1991]: time="2025-08-13T07:14:11.481216031Z" level=info msg="CreateContainer within sandbox \"3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:14:11.533952 containerd[1991]: time="2025-08-13T07:14:11.533899320Z" level=info msg="CreateContainer within sandbox \"3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9714daf5177e7c083b73c8f9ad00649ee91ec78892de388e56ced1da2bae8c71\"" Aug 13 07:14:11.536125 containerd[1991]: time="2025-08-13T07:14:11.536083609Z" level=info msg="StartContainer for \"9714daf5177e7c083b73c8f9ad00649ee91ec78892de388e56ced1da2bae8c71\"" Aug 13 07:14:11.561750 containerd[1991]: time="2025-08-13T07:14:11.561147291Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:11.561750 containerd[1991]: time="2025-08-13T07:14:11.561236306Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:11.561750 containerd[1991]: time="2025-08-13T07:14:11.561260058Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:11.561750 containerd[1991]: time="2025-08-13T07:14:11.561393740Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:11.601970 systemd[1]: Started cri-containerd-9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90.scope - libcontainer container 9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90. Aug 13 07:14:11.683414 systemd[1]: Started cri-containerd-9714daf5177e7c083b73c8f9ad00649ee91ec78892de388e56ced1da2bae8c71.scope - libcontainer container 9714daf5177e7c083b73c8f9ad00649ee91ec78892de388e56ced1da2bae8c71. Aug 13 07:14:11.818750 containerd[1991]: time="2025-08-13T07:14:11.817493236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-gbtfq,Uid:9f23824c-ed26-4c1b-bcd8-13d31489a3f2,Namespace:kube-system,Attempt:1,} returns sandbox id \"9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90\"" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.670 [INFO][5130] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.675 [INFO][5130] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" iface="eth0" netns="/var/run/netns/cni-9e171201-fbe3-5abe-9bbe-7b393ed62df2" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.677 [INFO][5130] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" iface="eth0" netns="/var/run/netns/cni-9e171201-fbe3-5abe-9bbe-7b393ed62df2" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.678 [INFO][5130] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" iface="eth0" netns="/var/run/netns/cni-9e171201-fbe3-5abe-9bbe-7b393ed62df2" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.679 [INFO][5130] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.679 [INFO][5130] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.787 [INFO][5187] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.787 [INFO][5187] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.787 [INFO][5187] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.809 [WARNING][5187] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.809 [INFO][5187] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.814 [INFO][5187] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:11.847444 containerd[1991]: 2025-08-13 07:14:11.829 [INFO][5130] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:11.852137 containerd[1991]: time="2025-08-13T07:14:11.848837647Z" level=info msg="TearDown network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\" successfully" Aug 13 07:14:11.852137 containerd[1991]: time="2025-08-13T07:14:11.848876972Z" level=info msg="StopPodSandbox for \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\" returns successfully" Aug 13 07:14:11.855714 systemd[1]: run-netns-cni\x2d9e171201\x2dfbe3\x2d5abe\x2d9bbe\x2d7b393ed62df2.mount: Deactivated successfully. Aug 13 07:14:11.891601 containerd[1991]: time="2025-08-13T07:14:11.891050504Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf57cfb9b-z778g,Uid:af80a1b7-708d-43e5-8d99-c39cc31d2f2e,Namespace:calico-system,Attempt:1,}" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.679 [INFO][5114] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.679 [INFO][5114] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" iface="eth0" netns="/var/run/netns/cni-698ac5f9-373c-ac9a-4245-4738acad9b40" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.683 [INFO][5114] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" iface="eth0" netns="/var/run/netns/cni-698ac5f9-373c-ac9a-4245-4738acad9b40" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.687 [INFO][5114] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" iface="eth0" netns="/var/run/netns/cni-698ac5f9-373c-ac9a-4245-4738acad9b40" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.687 [INFO][5114] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.687 [INFO][5114] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.856 [INFO][5191] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.857 [INFO][5191] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.857 [INFO][5191] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.870 [WARNING][5191] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.870 [INFO][5191] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.876 [INFO][5191] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:11.892700 containerd[1991]: 2025-08-13 07:14:11.885 [INFO][5114] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:11.895812 containerd[1991]: time="2025-08-13T07:14:11.894614334Z" level=info msg="TearDown network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\" successfully" Aug 13 07:14:11.895812 containerd[1991]: time="2025-08-13T07:14:11.894863529Z" level=info msg="StopPodSandbox for \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\" returns successfully" Aug 13 07:14:11.903709 systemd[1]: run-netns-cni\x2d698ac5f9\x2d373c\x2dac9a\x2d4245\x2d4738acad9b40.mount: Deactivated successfully. Aug 13 07:14:11.921786 containerd[1991]: time="2025-08-13T07:14:11.921707769Z" level=info msg="StartContainer for \"9714daf5177e7c083b73c8f9ad00649ee91ec78892de388e56ced1da2bae8c71\" returns successfully" Aug 13 07:14:11.948681 containerd[1991]: time="2025-08-13T07:14:11.948554206Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-wmhdx,Uid:e3e699d1-02d9-4ab9-afc8-37fa42f38432,Namespace:calico-apiserver,Attempt:1,}" Aug 13 07:14:11.970087 containerd[1991]: time="2025-08-13T07:14:11.970034545Z" level=info msg="CreateContainer within sandbox \"9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:14:12.017597 systemd-networkd[1818]: cali84771e8c5e6: Gained IPv6LL Aug 13 07:14:12.099048 containerd[1991]: time="2025-08-13T07:14:12.098264289Z" level=info msg="CreateContainer within sandbox \"9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"29598f0fde264983218807c9d223daa99609602f31892c0585e7a04c7defc3e9\"" Aug 13 07:14:12.105961 containerd[1991]: time="2025-08-13T07:14:12.105922071Z" level=info msg="StartContainer for \"29598f0fde264983218807c9d223daa99609602f31892c0585e7a04c7defc3e9\"" Aug 13 07:14:12.293068 systemd[1]: Started cri-containerd-29598f0fde264983218807c9d223daa99609602f31892c0585e7a04c7defc3e9.scope - libcontainer container 29598f0fde264983218807c9d223daa99609602f31892c0585e7a04c7defc3e9. Aug 13 07:14:12.426542 containerd[1991]: time="2025-08-13T07:14:12.425797226Z" level=info msg="StopPodSandbox for \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\"" Aug 13 07:14:12.447907 containerd[1991]: time="2025-08-13T07:14:12.447679999Z" level=info msg="StartContainer for \"29598f0fde264983218807c9d223daa99609602f31892c0585e7a04c7defc3e9\" returns successfully" Aug 13 07:14:12.579100 systemd-networkd[1818]: calie2f9c874d52: Link UP Aug 13 07:14:12.596079 systemd-networkd[1818]: calie2f9c874d52: Gained carrier Aug 13 07:14:12.716760 kubelet[3184]: I0813 07:14:12.716177 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-jkxj9" podStartSLOduration=42.716143749 podStartE2EDuration="42.716143749s" podCreationTimestamp="2025-08-13 07:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:14:12.042807856 +0000 UTC m=+47.837607050" watchObservedRunningTime="2025-08-13 07:14:12.716143749 +0000 UTC m=+48.510942940" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.111 [INFO][5224] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0 calico-kube-controllers-5bf57cfb9b- calico-system af80a1b7-708d-43e5-8d99-c39cc31d2f2e 941 0 2025-08-13 07:13:46 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5bf57cfb9b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ip-172-31-16-138 calico-kube-controllers-5bf57cfb9b-z778g eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calie2f9c874d52 [] [] }} ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.114 [INFO][5224] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.290 [INFO][5258] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" HandleID="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.291 [INFO][5258] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" HandleID="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000374640), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-16-138", "pod":"calico-kube-controllers-5bf57cfb9b-z778g", "timestamp":"2025-08-13 07:14:12.29076457 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.291 [INFO][5258] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.291 [INFO][5258] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.292 [INFO][5258] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.379 [INFO][5258] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.402 [INFO][5258] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.438 [INFO][5258] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.445 [INFO][5258] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.460 [INFO][5258] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.460 [INFO][5258] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.468 [INFO][5258] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530 Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.507 [INFO][5258] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.551 [INFO][5258] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.5/26] block=192.168.68.0/26 handle="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.551 [INFO][5258] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.5/26] handle="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" host="ip-172-31-16-138" Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.551 [INFO][5258] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:12.732932 containerd[1991]: 2025-08-13 07:14:12.552 [INFO][5258] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.5/26] IPv6=[] ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" HandleID="k8s-pod-network.3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:12.735354 containerd[1991]: 2025-08-13 07:14:12.561 [INFO][5224] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0", GenerateName:"calico-kube-controllers-5bf57cfb9b-", Namespace:"calico-system", SelfLink:"", UID:"af80a1b7-708d-43e5-8d99-c39cc31d2f2e", ResourceVersion:"941", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bf57cfb9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"calico-kube-controllers-5bf57cfb9b-z778g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie2f9c874d52", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:12.735354 containerd[1991]: 2025-08-13 07:14:12.562 [INFO][5224] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.5/32] ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:12.735354 containerd[1991]: 2025-08-13 07:14:12.562 [INFO][5224] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie2f9c874d52 ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:12.735354 containerd[1991]: 2025-08-13 07:14:12.602 [INFO][5224] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:12.735354 containerd[1991]: 2025-08-13 07:14:12.614 [INFO][5224] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0", GenerateName:"calico-kube-controllers-5bf57cfb9b-", Namespace:"calico-system", SelfLink:"", UID:"af80a1b7-708d-43e5-8d99-c39cc31d2f2e", ResourceVersion:"941", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bf57cfb9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530", Pod:"calico-kube-controllers-5bf57cfb9b-z778g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie2f9c874d52", MAC:"56:04:ed:63:1f:83", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:12.735354 containerd[1991]: 2025-08-13 07:14:12.720 [INFO][5224] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530" Namespace="calico-system" Pod="calico-kube-controllers-5bf57cfb9b-z778g" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:12.867256 containerd[1991]: time="2025-08-13T07:14:12.862574064Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:12.867256 containerd[1991]: time="2025-08-13T07:14:12.865858377Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:12.867256 containerd[1991]: time="2025-08-13T07:14:12.865885617Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:12.867256 containerd[1991]: time="2025-08-13T07:14:12.866018904Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:12.913804 containerd[1991]: time="2025-08-13T07:14:12.913757130Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:12.914200 systemd-networkd[1818]: calia07661b75a4: Gained IPv6LL Aug 13 07:14:12.916022 systemd-networkd[1818]: vxlan.calico: Gained IPv6LL Aug 13 07:14:12.919301 containerd[1991]: time="2025-08-13T07:14:12.914154879Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.2: active requests=0, bytes read=4661207" Aug 13 07:14:12.929058 containerd[1991]: time="2025-08-13T07:14:12.927907793Z" level=info msg="ImageCreate event name:\"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:12.938218 containerd[1991]: time="2025-08-13T07:14:12.936956510Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:12.942555 containerd[1991]: time="2025-08-13T07:14:12.940381393Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.2\" with image id \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\", size \"6153902\" in 2.646309879s" Aug 13 07:14:12.942555 containerd[1991]: time="2025-08-13T07:14:12.940428512Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\"" Aug 13 07:14:12.970187 containerd[1991]: time="2025-08-13T07:14:12.970134909Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 07:14:12.972050 systemd[1]: Started cri-containerd-3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530.scope - libcontainer container 3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530. Aug 13 07:14:13.046252 containerd[1991]: time="2025-08-13T07:14:13.043335858Z" level=info msg="CreateContainer within sandbox \"c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Aug 13 07:14:13.053782 systemd-networkd[1818]: cali12081f8f9f2: Link UP Aug 13 07:14:13.055562 systemd-networkd[1818]: cali12081f8f9f2: Gained carrier Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.343 [INFO][5239] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0 calico-apiserver-645bc4db8- calico-apiserver e3e699d1-02d9-4ab9-afc8-37fa42f38432 942 0 2025-08-13 07:13:41 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:645bc4db8 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-16-138 calico-apiserver-645bc4db8-wmhdx eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali12081f8f9f2 [] [] }} ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.344 [INFO][5239] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.537 [INFO][5297] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" HandleID="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.537 [INFO][5297] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" HandleID="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000103e00), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-16-138", "pod":"calico-apiserver-645bc4db8-wmhdx", "timestamp":"2025-08-13 07:14:12.537021049 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.537 [INFO][5297] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.555 [INFO][5297] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.555 [INFO][5297] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.632 [INFO][5297] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.747 [INFO][5297] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.843 [INFO][5297] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.854 [INFO][5297] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.863 [INFO][5297] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.864 [INFO][5297] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.867 [INFO][5297] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876 Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.885 [INFO][5297] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.964 [INFO][5297] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.6/26] block=192.168.68.0/26 handle="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.964 [INFO][5297] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.6/26] handle="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" host="ip-172-31-16-138" Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.964 [INFO][5297] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:13.110046 containerd[1991]: 2025-08-13 07:14:12.964 [INFO][5297] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.6/26] IPv6=[] ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" HandleID="k8s-pod-network.687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:13.111049 containerd[1991]: 2025-08-13 07:14:13.008 [INFO][5239] cni-plugin/k8s.go 418: Populated endpoint ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"e3e699d1-02d9-4ab9-afc8-37fa42f38432", ResourceVersion:"942", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"calico-apiserver-645bc4db8-wmhdx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali12081f8f9f2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:13.111049 containerd[1991]: 2025-08-13 07:14:13.010 [INFO][5239] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.6/32] ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:13.111049 containerd[1991]: 2025-08-13 07:14:13.017 [INFO][5239] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali12081f8f9f2 ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:13.111049 containerd[1991]: 2025-08-13 07:14:13.046 [INFO][5239] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:13.111049 containerd[1991]: 2025-08-13 07:14:13.047 [INFO][5239] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"e3e699d1-02d9-4ab9-afc8-37fa42f38432", ResourceVersion:"942", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876", Pod:"calico-apiserver-645bc4db8-wmhdx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali12081f8f9f2", MAC:"56:9e:94:a4:93:00", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:13.111049 containerd[1991]: 2025-08-13 07:14:13.091 [INFO][5239] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876" Namespace="calico-apiserver" Pod="calico-apiserver-645bc4db8-wmhdx" WorkloadEndpoint="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:13.150354 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2604214104.mount: Deactivated successfully. Aug 13 07:14:13.160905 containerd[1991]: time="2025-08-13T07:14:13.160842176Z" level=info msg="CreateContainer within sandbox \"c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"d4344b696998980fc954bd5a52785ae61ab38138ac5de84654cc8c9c81ae99c0\"" Aug 13 07:14:13.164754 containerd[1991]: time="2025-08-13T07:14:13.163321140Z" level=info msg="StartContainer for \"d4344b696998980fc954bd5a52785ae61ab38138ac5de84654cc8c9c81ae99c0\"" Aug 13 07:14:13.218012 containerd[1991]: time="2025-08-13T07:14:13.217856193Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:13.218178 containerd[1991]: time="2025-08-13T07:14:13.218020713Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:13.218178 containerd[1991]: time="2025-08-13T07:14:13.218094177Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:13.219687 containerd[1991]: time="2025-08-13T07:14:13.218234559Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:13.232977 systemd[1]: Started cri-containerd-d4344b696998980fc954bd5a52785ae61ab38138ac5de84654cc8c9c81ae99c0.scope - libcontainer container d4344b696998980fc954bd5a52785ae61ab38138ac5de84654cc8c9c81ae99c0. Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:12.881 [INFO][5321] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:12.882 [INFO][5321] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" iface="eth0" netns="/var/run/netns/cni-16e8807e-cda4-8d34-212c-b3e82cb08a23" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:12.882 [INFO][5321] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" iface="eth0" netns="/var/run/netns/cni-16e8807e-cda4-8d34-212c-b3e82cb08a23" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:12.882 [INFO][5321] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" iface="eth0" netns="/var/run/netns/cni-16e8807e-cda4-8d34-212c-b3e82cb08a23" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:12.883 [INFO][5321] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:12.883 [INFO][5321] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:13.071 [INFO][5369] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:13.071 [INFO][5369] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:13.072 [INFO][5369] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:13.163 [WARNING][5369] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:13.163 [INFO][5369] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:13.198 [INFO][5369] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:13.237077 containerd[1991]: 2025-08-13 07:14:13.223 [INFO][5321] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:13.237789 containerd[1991]: time="2025-08-13T07:14:13.237396500Z" level=info msg="TearDown network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\" successfully" Aug 13 07:14:13.237789 containerd[1991]: time="2025-08-13T07:14:13.237431245Z" level=info msg="StopPodSandbox for \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\" returns successfully" Aug 13 07:14:13.244606 containerd[1991]: time="2025-08-13T07:14:13.240670998Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-lw9wb,Uid:b381d26d-9591-43bd-8a36-a31d1d40c8e1,Namespace:calico-system,Attempt:1,}" Aug 13 07:14:13.244772 kubelet[3184]: I0813 07:14:13.242249 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-gbtfq" podStartSLOduration=43.242224984 podStartE2EDuration="43.242224984s" podCreationTimestamp="2025-08-13 07:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:14:13.164695608 +0000 UTC m=+48.959494798" watchObservedRunningTime="2025-08-13 07:14:13.242224984 +0000 UTC m=+49.037024179" Aug 13 07:14:13.261970 systemd[1]: Started cri-containerd-687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876.scope - libcontainer container 687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876. Aug 13 07:14:13.362097 systemd-networkd[1818]: cali7cded0b467d: Gained IPv6LL Aug 13 07:14:13.534978 systemd-networkd[1818]: caliad0836b7584: Link UP Aug 13 07:14:13.538300 systemd-networkd[1818]: caliad0836b7584: Gained carrier Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.354 [INFO][5443] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0 goldmane-768f4c5c69- calico-system b381d26d-9591-43bd-8a36-a31d1d40c8e1 961 0 2025-08-13 07:13:45 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:768f4c5c69 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ip-172-31-16-138 goldmane-768f4c5c69-lw9wb eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] caliad0836b7584 [] [] }} ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.355 [INFO][5443] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.430 [INFO][5466] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" HandleID="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.430 [INFO][5466] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" HandleID="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f6c0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-16-138", "pod":"goldmane-768f4c5c69-lw9wb", "timestamp":"2025-08-13 07:14:13.430333504 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.431 [INFO][5466] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.431 [INFO][5466] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.431 [INFO][5466] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.446 [INFO][5466] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.457 [INFO][5466] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.467 [INFO][5466] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.474 [INFO][5466] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.483 [INFO][5466] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.483 [INFO][5466] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.485 [INFO][5466] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.498 [INFO][5466] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.519 [INFO][5466] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.7/26] block=192.168.68.0/26 handle="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.519 [INFO][5466] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.7/26] handle="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" host="ip-172-31-16-138" Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.519 [INFO][5466] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:13.575405 containerd[1991]: 2025-08-13 07:14:13.519 [INFO][5466] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.7/26] IPv6=[] ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" HandleID="k8s-pod-network.1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.580210 containerd[1991]: 2025-08-13 07:14:13.526 [INFO][5443] cni-plugin/k8s.go 418: Populated endpoint ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b381d26d-9591-43bd-8a36-a31d1d40c8e1", ResourceVersion:"961", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"goldmane-768f4c5c69-lw9wb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"caliad0836b7584", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:13.580210 containerd[1991]: 2025-08-13 07:14:13.527 [INFO][5443] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.7/32] ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.580210 containerd[1991]: 2025-08-13 07:14:13.528 [INFO][5443] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliad0836b7584 ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.580210 containerd[1991]: 2025-08-13 07:14:13.540 [INFO][5443] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.580210 containerd[1991]: 2025-08-13 07:14:13.540 [INFO][5443] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b381d26d-9591-43bd-8a36-a31d1d40c8e1", ResourceVersion:"961", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f", Pod:"goldmane-768f4c5c69-lw9wb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"caliad0836b7584", MAC:"da:c1:c9:23:c8:cd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:13.580210 containerd[1991]: 2025-08-13 07:14:13.567 [INFO][5443] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f" Namespace="calico-system" Pod="goldmane-768f4c5c69-lw9wb" WorkloadEndpoint="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:13.632555 containerd[1991]: time="2025-08-13T07:14:13.632270572Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:13.632555 containerd[1991]: time="2025-08-13T07:14:13.632412698Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:13.632555 containerd[1991]: time="2025-08-13T07:14:13.632442546Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:13.634114 containerd[1991]: time="2025-08-13T07:14:13.632593033Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:13.641576 systemd[1]: run-netns-cni\x2d16e8807e\x2dcda4\x2d8d34\x2d212c\x2db3e82cb08a23.mount: Deactivated successfully. Aug 13 07:14:13.717472 systemd[1]: Started sshd@7-172.31.16.138:22-147.75.109.163:32888.service - OpenSSH per-connection server daemon (147.75.109.163:32888). Aug 13 07:14:13.748029 systemd[1]: Started cri-containerd-1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f.scope - libcontainer container 1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f. Aug 13 07:14:13.805851 containerd[1991]: time="2025-08-13T07:14:13.804981624Z" level=info msg="StartContainer for \"d4344b696998980fc954bd5a52785ae61ab38138ac5de84654cc8c9c81ae99c0\" returns successfully" Aug 13 07:14:14.014330 sshd[5524]: Accepted publickey for core from 147.75.109.163 port 32888 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:14.020303 sshd[5524]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:14.043793 systemd-logind[1961]: New session 8 of user core. Aug 13 07:14:14.048454 systemd[1]: Started session-8.scope - Session 8 of User core. Aug 13 07:14:14.072749 containerd[1991]: time="2025-08-13T07:14:14.072138514Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf57cfb9b-z778g,Uid:af80a1b7-708d-43e5-8d99-c39cc31d2f2e,Namespace:calico-system,Attempt:1,} returns sandbox id \"3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530\"" Aug 13 07:14:14.128512 containerd[1991]: time="2025-08-13T07:14:14.128112156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-645bc4db8-wmhdx,Uid:e3e699d1-02d9-4ab9-afc8-37fa42f38432,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876\"" Aug 13 07:14:14.193075 systemd-networkd[1818]: cali12081f8f9f2: Gained IPv6LL Aug 13 07:14:14.264598 containerd[1991]: time="2025-08-13T07:14:14.264543077Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-lw9wb,Uid:b381d26d-9591-43bd-8a36-a31d1d40c8e1,Namespace:calico-system,Attempt:1,} returns sandbox id \"1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f\"" Aug 13 07:14:14.418823 containerd[1991]: time="2025-08-13T07:14:14.416893014Z" level=info msg="StopPodSandbox for \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\"" Aug 13 07:14:14.513991 systemd-networkd[1818]: calie2f9c874d52: Gained IPv6LL Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.525 [INFO][5600] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.525 [INFO][5600] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" iface="eth0" netns="/var/run/netns/cni-f5016ab8-1324-d325-1c63-dd1bf6e2a216" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.530 [INFO][5600] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" iface="eth0" netns="/var/run/netns/cni-f5016ab8-1324-d325-1c63-dd1bf6e2a216" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.537 [INFO][5600] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" iface="eth0" netns="/var/run/netns/cni-f5016ab8-1324-d325-1c63-dd1bf6e2a216" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.537 [INFO][5600] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.537 [INFO][5600] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.647 [INFO][5608] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.649 [INFO][5608] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.649 [INFO][5608] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.688 [WARNING][5608] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.688 [INFO][5608] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.693 [INFO][5608] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:14.708197 containerd[1991]: 2025-08-13 07:14:14.702 [INFO][5600] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:14.713875 containerd[1991]: time="2025-08-13T07:14:14.712855594Z" level=info msg="TearDown network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\" successfully" Aug 13 07:14:14.713875 containerd[1991]: time="2025-08-13T07:14:14.712899489Z" level=info msg="StopPodSandbox for \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\" returns successfully" Aug 13 07:14:14.716418 containerd[1991]: time="2025-08-13T07:14:14.716356594Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-782zn,Uid:a57f8240-352f-45e2-8291-fe7645c155a3,Namespace:calico-system,Attempt:1,}" Aug 13 07:14:14.718232 systemd[1]: run-netns-cni\x2df5016ab8\x2d1324\x2dd325\x2d1c63\x2ddd1bf6e2a216.mount: Deactivated successfully. Aug 13 07:14:15.335168 systemd-networkd[1818]: calic82a766ceec: Link UP Aug 13 07:14:15.335530 systemd-networkd[1818]: calic82a766ceec: Gained carrier Aug 13 07:14:15.345013 systemd-networkd[1818]: caliad0836b7584: Gained IPv6LL Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:14.950 [INFO][5615] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0 csi-node-driver- calico-system a57f8240-352f-45e2-8291-fe7645c155a3 1020 0 2025-08-13 07:13:46 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:8967bcb6f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ip-172-31-16-138 csi-node-driver-782zn eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calic82a766ceec [] [] }} ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:14.951 [INFO][5615] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.168 [INFO][5628] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" HandleID="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.168 [INFO][5628] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" HandleID="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5b80), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-16-138", "pod":"csi-node-driver-782zn", "timestamp":"2025-08-13 07:14:15.163602316 +0000 UTC"}, Hostname:"ip-172-31-16-138", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.168 [INFO][5628] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.169 [INFO][5628] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.169 [INFO][5628] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-16-138' Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.197 [INFO][5628] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.215 [INFO][5628] ipam/ipam.go 394: Looking up existing affinities for host host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.232 [INFO][5628] ipam/ipam.go 511: Trying affinity for 192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.239 [INFO][5628] ipam/ipam.go 158: Attempting to load block cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.250 [INFO][5628] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.68.0/26 host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.250 [INFO][5628] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.68.0/26 handle="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.258 [INFO][5628] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482 Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.275 [INFO][5628] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.68.0/26 handle="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.302 [INFO][5628] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.68.8/26] block=192.168.68.0/26 handle="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.304 [INFO][5628] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.68.8/26] handle="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" host="ip-172-31-16-138" Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.304 [INFO][5628] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:15.413466 containerd[1991]: 2025-08-13 07:14:15.305 [INFO][5628] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.68.8/26] IPv6=[] ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" HandleID="k8s-pod-network.89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:15.421211 containerd[1991]: 2025-08-13 07:14:15.315 [INFO][5615] cni-plugin/k8s.go 418: Populated endpoint ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a57f8240-352f-45e2-8291-fe7645c155a3", ResourceVersion:"1020", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"", Pod:"csi-node-driver-782zn", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic82a766ceec", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:15.421211 containerd[1991]: 2025-08-13 07:14:15.315 [INFO][5615] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.68.8/32] ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:15.421211 containerd[1991]: 2025-08-13 07:14:15.317 [INFO][5615] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic82a766ceec ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:15.421211 containerd[1991]: 2025-08-13 07:14:15.336 [INFO][5615] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:15.421211 containerd[1991]: 2025-08-13 07:14:15.337 [INFO][5615] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a57f8240-352f-45e2-8291-fe7645c155a3", ResourceVersion:"1020", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482", Pod:"csi-node-driver-782zn", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic82a766ceec", MAC:"ee:48:3b:c7:0b:bd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:15.421211 containerd[1991]: 2025-08-13 07:14:15.408 [INFO][5615] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482" Namespace="calico-system" Pod="csi-node-driver-782zn" WorkloadEndpoint="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:15.633153 containerd[1991]: time="2025-08-13T07:14:15.631513465Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:14:15.633153 containerd[1991]: time="2025-08-13T07:14:15.632862642Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:14:15.633153 containerd[1991]: time="2025-08-13T07:14:15.632893232Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:15.636249 containerd[1991]: time="2025-08-13T07:14:15.633122840Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:14:15.695239 sshd[5524]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:15.707639 systemd[1]: sshd@7-172.31.16.138:22-147.75.109.163:32888.service: Deactivated successfully. Aug 13 07:14:15.730945 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 07:14:15.736546 systemd-logind[1961]: Session 8 logged out. Waiting for processes to exit. Aug 13 07:14:15.749006 systemd[1]: Started cri-containerd-89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482.scope - libcontainer container 89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482. Aug 13 07:14:15.752100 systemd-logind[1961]: Removed session 8. Aug 13 07:14:15.843324 containerd[1991]: time="2025-08-13T07:14:15.842485302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-782zn,Uid:a57f8240-352f-45e2-8291-fe7645c155a3,Namespace:calico-system,Attempt:1,} returns sandbox id \"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482\"" Aug 13 07:14:16.998100 containerd[1991]: time="2025-08-13T07:14:16.998046058Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:16.999425 containerd[1991]: time="2025-08-13T07:14:16.999337404Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=47317977" Aug 13 07:14:17.008299 containerd[1991]: time="2025-08-13T07:14:17.008230399Z" level=info msg="ImageCreate event name:\"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:17.054003 containerd[1991]: time="2025-08-13T07:14:17.053160568Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:17.054003 containerd[1991]: time="2025-08-13T07:14:17.053860553Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 4.083672463s" Aug 13 07:14:17.054003 containerd[1991]: time="2025-08-13T07:14:17.053892400Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Aug 13 07:14:17.055086 containerd[1991]: time="2025-08-13T07:14:17.055047447Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Aug 13 07:14:17.065863 containerd[1991]: time="2025-08-13T07:14:17.065822417Z" level=info msg="CreateContainer within sandbox \"ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 07:14:17.088102 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3373381503.mount: Deactivated successfully. Aug 13 07:14:17.095299 containerd[1991]: time="2025-08-13T07:14:17.094997726Z" level=info msg="CreateContainer within sandbox \"ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"6b9d777651092d1238da2309fa5b60703e72c5d6fc5e6ef8426ae74710d13c79\"" Aug 13 07:14:17.096753 containerd[1991]: time="2025-08-13T07:14:17.096467643Z" level=info msg="StartContainer for \"6b9d777651092d1238da2309fa5b60703e72c5d6fc5e6ef8426ae74710d13c79\"" Aug 13 07:14:17.152985 systemd[1]: Started cri-containerd-6b9d777651092d1238da2309fa5b60703e72c5d6fc5e6ef8426ae74710d13c79.scope - libcontainer container 6b9d777651092d1238da2309fa5b60703e72c5d6fc5e6ef8426ae74710d13c79. Aug 13 07:14:17.229037 containerd[1991]: time="2025-08-13T07:14:17.228985656Z" level=info msg="StartContainer for \"6b9d777651092d1238da2309fa5b60703e72c5d6fc5e6ef8426ae74710d13c79\" returns successfully" Aug 13 07:14:17.265165 systemd-networkd[1818]: calic82a766ceec: Gained IPv6LL Aug 13 07:14:18.286766 kubelet[3184]: I0813 07:14:18.286508 3184 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:14:19.739094 ntpd[1953]: Listen normally on 7 vxlan.calico 192.168.68.0:123 Aug 13 07:14:19.739863 ntpd[1953]: Listen normally on 8 calif17e658e0c0 [fe80::ecee:eeff:feee:eeee%4]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 7 vxlan.calico 192.168.68.0:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 8 calif17e658e0c0 [fe80::ecee:eeff:feee:eeee%4]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 9 cali84771e8c5e6 [fe80::ecee:eeff:feee:eeee%5]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 10 vxlan.calico [fe80::6415:b6ff:fe27:afcb%6]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 11 calia07661b75a4 [fe80::ecee:eeff:feee:eeee%9]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 12 cali7cded0b467d [fe80::ecee:eeff:feee:eeee%10]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 13 calie2f9c874d52 [fe80::ecee:eeff:feee:eeee%11]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 14 cali12081f8f9f2 [fe80::ecee:eeff:feee:eeee%12]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 15 caliad0836b7584 [fe80::ecee:eeff:feee:eeee%13]:123 Aug 13 07:14:19.742443 ntpd[1953]: 13 Aug 07:14:19 ntpd[1953]: Listen normally on 16 calic82a766ceec [fe80::ecee:eeff:feee:eeee%14]:123 Aug 13 07:14:19.739956 ntpd[1953]: Listen normally on 9 cali84771e8c5e6 [fe80::ecee:eeff:feee:eeee%5]:123 Aug 13 07:14:19.740007 ntpd[1953]: Listen normally on 10 vxlan.calico [fe80::6415:b6ff:fe27:afcb%6]:123 Aug 13 07:14:19.740048 ntpd[1953]: Listen normally on 11 calia07661b75a4 [fe80::ecee:eeff:feee:eeee%9]:123 Aug 13 07:14:19.740101 ntpd[1953]: Listen normally on 12 cali7cded0b467d [fe80::ecee:eeff:feee:eeee%10]:123 Aug 13 07:14:19.740145 ntpd[1953]: Listen normally on 13 calie2f9c874d52 [fe80::ecee:eeff:feee:eeee%11]:123 Aug 13 07:14:19.740223 ntpd[1953]: Listen normally on 14 cali12081f8f9f2 [fe80::ecee:eeff:feee:eeee%12]:123 Aug 13 07:14:19.740254 ntpd[1953]: Listen normally on 15 caliad0836b7584 [fe80::ecee:eeff:feee:eeee%13]:123 Aug 13 07:14:19.740319 ntpd[1953]: Listen normally on 16 calic82a766ceec [fe80::ecee:eeff:feee:eeee%14]:123 Aug 13 07:14:19.782394 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1787254411.mount: Deactivated successfully. Aug 13 07:14:19.834404 containerd[1991]: time="2025-08-13T07:14:19.834232572Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.2: active requests=0, bytes read=33083477" Aug 13 07:14:19.834834 containerd[1991]: time="2025-08-13T07:14:19.834435139Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:19.848341 containerd[1991]: time="2025-08-13T07:14:19.848204862Z" level=info msg="ImageCreate event name:\"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:19.852397 containerd[1991]: time="2025-08-13T07:14:19.852343217Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:19.854176 containerd[1991]: time="2025-08-13T07:14:19.853361424Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" with image id \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\", size \"33083307\" in 2.79827735s" Aug 13 07:14:19.854176 containerd[1991]: time="2025-08-13T07:14:19.853393951Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\"" Aug 13 07:14:19.856460 containerd[1991]: time="2025-08-13T07:14:19.856213417Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Aug 13 07:14:19.863687 containerd[1991]: time="2025-08-13T07:14:19.863620115Z" level=info msg="CreateContainer within sandbox \"c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Aug 13 07:14:19.910240 containerd[1991]: time="2025-08-13T07:14:19.910156551Z" level=info msg="CreateContainer within sandbox \"c7d58e498d67be869d57d67285872de2402f1f0e35e672089be01233405bcba7\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"7d9752b22846acda8ad696ade20808e2fe7536aa9d2b08f04d52f6f63a8c6e3f\"" Aug 13 07:14:19.928190 containerd[1991]: time="2025-08-13T07:14:19.927125565Z" level=info msg="StartContainer for \"7d9752b22846acda8ad696ade20808e2fe7536aa9d2b08f04d52f6f63a8c6e3f\"" Aug 13 07:14:19.989935 systemd[1]: run-containerd-runc-k8s.io-7d9752b22846acda8ad696ade20808e2fe7536aa9d2b08f04d52f6f63a8c6e3f-runc.gCj1ZV.mount: Deactivated successfully. Aug 13 07:14:20.017495 systemd[1]: Started cri-containerd-7d9752b22846acda8ad696ade20808e2fe7536aa9d2b08f04d52f6f63a8c6e3f.scope - libcontainer container 7d9752b22846acda8ad696ade20808e2fe7536aa9d2b08f04d52f6f63a8c6e3f. Aug 13 07:14:20.091462 containerd[1991]: time="2025-08-13T07:14:20.091230005Z" level=info msg="StartContainer for \"7d9752b22846acda8ad696ade20808e2fe7536aa9d2b08f04d52f6f63a8c6e3f\" returns successfully" Aug 13 07:14:20.538616 kubelet[3184]: I0813 07:14:20.446091 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-645bc4db8-2qtgq" podStartSLOduration=32.941559563 podStartE2EDuration="39.437534781s" podCreationTimestamp="2025-08-13 07:13:41 +0000 UTC" firstStartedPulling="2025-08-13 07:14:10.558941098 +0000 UTC m=+46.353740266" lastFinishedPulling="2025-08-13 07:14:17.054916316 +0000 UTC m=+52.849715484" observedRunningTime="2025-08-13 07:14:17.294467806 +0000 UTC m=+53.089266996" watchObservedRunningTime="2025-08-13 07:14:20.437534781 +0000 UTC m=+56.232333973" Aug 13 07:14:20.539139 kubelet[3184]: I0813 07:14:20.538754 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-7fbf94ddb4-jrjql" podStartSLOduration=2.95532556 podStartE2EDuration="12.538734271s" podCreationTimestamp="2025-08-13 07:14:08 +0000 UTC" firstStartedPulling="2025-08-13 07:14:10.271620245 +0000 UTC m=+46.066419428" lastFinishedPulling="2025-08-13 07:14:19.855028971 +0000 UTC m=+55.649828139" observedRunningTime="2025-08-13 07:14:20.427199222 +0000 UTC m=+56.221998414" watchObservedRunningTime="2025-08-13 07:14:20.538734271 +0000 UTC m=+56.333533453" Aug 13 07:14:20.731045 systemd[1]: Started sshd@8-172.31.16.138:22-147.75.109.163:50268.service - OpenSSH per-connection server daemon (147.75.109.163:50268). Aug 13 07:14:20.977826 sshd[5798]: Accepted publickey for core from 147.75.109.163 port 50268 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:20.981010 sshd[5798]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:20.989069 systemd-logind[1961]: New session 9 of user core. Aug 13 07:14:20.996902 systemd[1]: Started session-9.scope - Session 9 of User core. Aug 13 07:14:22.466379 sshd[5798]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:22.476987 systemd-logind[1961]: Session 9 logged out. Waiting for processes to exit. Aug 13 07:14:22.477429 systemd[1]: sshd@8-172.31.16.138:22-147.75.109.163:50268.service: Deactivated successfully. Aug 13 07:14:22.482447 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 07:14:22.485594 systemd-logind[1961]: Removed session 9. Aug 13 07:14:23.724174 containerd[1991]: time="2025-08-13T07:14:23.724126779Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:23.732916 containerd[1991]: time="2025-08-13T07:14:23.732792197Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.2: active requests=0, bytes read=51276688" Aug 13 07:14:23.737872 containerd[1991]: time="2025-08-13T07:14:23.737826292Z" level=info msg="ImageCreate event name:\"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:23.796762 containerd[1991]: time="2025-08-13T07:14:23.795999820Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:23.797184 containerd[1991]: time="2025-08-13T07:14:23.797145747Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" with image id \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\", size \"52769359\" in 3.940894026s" Aug 13 07:14:23.797306 containerd[1991]: time="2025-08-13T07:14:23.797285383Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\"" Aug 13 07:14:23.925847 containerd[1991]: time="2025-08-13T07:14:23.925803965Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 07:14:24.193148 containerd[1991]: time="2025-08-13T07:14:24.193103579Z" level=info msg="CreateContainer within sandbox \"3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Aug 13 07:14:24.234914 containerd[1991]: time="2025-08-13T07:14:24.234793095Z" level=info msg="CreateContainer within sandbox \"3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"33c9b44cbeca6552a899016ebceff443d21ce66aa9e8ddfa5a720640f678e497\"" Aug 13 07:14:24.259119 containerd[1991]: time="2025-08-13T07:14:24.257621339Z" level=info msg="StartContainer for \"33c9b44cbeca6552a899016ebceff443d21ce66aa9e8ddfa5a720640f678e497\"" Aug 13 07:14:24.265052 containerd[1991]: time="2025-08-13T07:14:24.264101961Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:24.267003 containerd[1991]: time="2025-08-13T07:14:24.266939587Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=77" Aug 13 07:14:24.271024 containerd[1991]: time="2025-08-13T07:14:24.270942771Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 345.093487ms" Aug 13 07:14:24.271024 containerd[1991]: time="2025-08-13T07:14:24.271025228Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Aug 13 07:14:24.273105 containerd[1991]: time="2025-08-13T07:14:24.273063758Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Aug 13 07:14:24.378815 containerd[1991]: time="2025-08-13T07:14:24.378655912Z" level=info msg="CreateContainer within sandbox \"687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 07:14:24.425774 systemd[1]: Started cri-containerd-33c9b44cbeca6552a899016ebceff443d21ce66aa9e8ddfa5a720640f678e497.scope - libcontainer container 33c9b44cbeca6552a899016ebceff443d21ce66aa9e8ddfa5a720640f678e497. Aug 13 07:14:24.433873 containerd[1991]: time="2025-08-13T07:14:24.433784143Z" level=info msg="CreateContainer within sandbox \"687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"053451b150fb8f5f3378144372718310a79deb643485be438aece0c4f4fee0c7\"" Aug 13 07:14:24.568607 containerd[1991]: time="2025-08-13T07:14:24.568544114Z" level=info msg="StartContainer for \"33c9b44cbeca6552a899016ebceff443d21ce66aa9e8ddfa5a720640f678e497\" returns successfully" Aug 13 07:14:24.837779 containerd[1991]: time="2025-08-13T07:14:24.837236973Z" level=info msg="StartContainer for \"053451b150fb8f5f3378144372718310a79deb643485be438aece0c4f4fee0c7\"" Aug 13 07:14:24.959426 systemd[1]: Started cri-containerd-053451b150fb8f5f3378144372718310a79deb643485be438aece0c4f4fee0c7.scope - libcontainer container 053451b150fb8f5f3378144372718310a79deb643485be438aece0c4f4fee0c7. Aug 13 07:14:25.071612 containerd[1991]: time="2025-08-13T07:14:25.071565016Z" level=info msg="StartContainer for \"053451b150fb8f5f3378144372718310a79deb643485be438aece0c4f4fee0c7\" returns successfully" Aug 13 07:14:25.076809 containerd[1991]: time="2025-08-13T07:14:25.076629592Z" level=info msg="StopPodSandbox for \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\"" Aug 13 07:14:25.440233 kubelet[3184]: I0813 07:14:25.436633 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-645bc4db8-wmhdx" podStartSLOduration=34.276619007 podStartE2EDuration="44.402960085s" podCreationTimestamp="2025-08-13 07:13:41 +0000 UTC" firstStartedPulling="2025-08-13 07:14:14.146143692 +0000 UTC m=+49.940942879" lastFinishedPulling="2025-08-13 07:14:24.272484789 +0000 UTC m=+60.067283957" observedRunningTime="2025-08-13 07:14:25.38461585 +0000 UTC m=+61.179415041" watchObservedRunningTime="2025-08-13 07:14:25.402960085 +0000 UTC m=+61.197759276" Aug 13 07:14:26.300210 kubelet[3184]: I0813 07:14:26.300091 3184 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:25.859 [WARNING][5916] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0", GenerateName:"calico-kube-controllers-5bf57cfb9b-", Namespace:"calico-system", SelfLink:"", UID:"af80a1b7-708d-43e5-8d99-c39cc31d2f2e", ResourceVersion:"1108", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bf57cfb9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530", Pod:"calico-kube-controllers-5bf57cfb9b-z778g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie2f9c874d52", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:25.865 [INFO][5916] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:25.866 [INFO][5916] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" iface="eth0" netns="" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:25.866 [INFO][5916] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:25.866 [INFO][5916] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:26.386 [INFO][5936] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:26.391 [INFO][5936] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:26.393 [INFO][5936] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:26.414 [WARNING][5936] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:26.414 [INFO][5936] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:26.427 [INFO][5936] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:26.464629 containerd[1991]: 2025-08-13 07:14:26.455 [INFO][5916] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.467929 containerd[1991]: time="2025-08-13T07:14:26.464817387Z" level=info msg="TearDown network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\" successfully" Aug 13 07:14:26.467929 containerd[1991]: time="2025-08-13T07:14:26.464897662Z" level=info msg="StopPodSandbox for \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\" returns successfully" Aug 13 07:14:26.788265 containerd[1991]: time="2025-08-13T07:14:26.788215034Z" level=info msg="RemovePodSandbox for \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\"" Aug 13 07:14:26.794270 containerd[1991]: time="2025-08-13T07:14:26.794201055Z" level=info msg="Forcibly stopping sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\"" Aug 13 07:14:26.840163 kubelet[3184]: I0813 07:14:26.839677 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5bf57cfb9b-z778g" podStartSLOduration=31.014659952 podStartE2EDuration="40.839652252s" podCreationTimestamp="2025-08-13 07:13:46 +0000 UTC" firstStartedPulling="2025-08-13 07:14:14.080483674 +0000 UTC m=+49.875282842" lastFinishedPulling="2025-08-13 07:14:23.90547594 +0000 UTC m=+59.700275142" observedRunningTime="2025-08-13 07:14:25.466674286 +0000 UTC m=+61.261473477" watchObservedRunningTime="2025-08-13 07:14:26.839652252 +0000 UTC m=+62.634451434" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.877 [WARNING][5969] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0", GenerateName:"calico-kube-controllers-5bf57cfb9b-", Namespace:"calico-system", SelfLink:"", UID:"af80a1b7-708d-43e5-8d99-c39cc31d2f2e", ResourceVersion:"1113", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bf57cfb9b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"3aac4d4fd9ebeb7358e207fb0932e3598ddd9432e827499660e6f3a2be885530", Pod:"calico-kube-controllers-5bf57cfb9b-z778g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.68.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calie2f9c874d52", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.878 [INFO][5969] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.878 [INFO][5969] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" iface="eth0" netns="" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.878 [INFO][5969] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.880 [INFO][5969] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.923 [INFO][5976] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.923 [INFO][5976] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.923 [INFO][5976] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.940 [WARNING][5976] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.940 [INFO][5976] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" HandleID="k8s-pod-network.ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Workload="ip--172--31--16--138-k8s-calico--kube--controllers--5bf57cfb9b--z778g-eth0" Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.946 [INFO][5976] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:26.954061 containerd[1991]: 2025-08-13 07:14:26.950 [INFO][5969] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea" Aug 13 07:14:26.955973 containerd[1991]: time="2025-08-13T07:14:26.954111122Z" level=info msg="TearDown network for sandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\" successfully" Aug 13 07:14:26.991123 containerd[1991]: time="2025-08-13T07:14:26.991057800Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:27.022363 containerd[1991]: time="2025-08-13T07:14:27.022300807Z" level=info msg="RemovePodSandbox \"ee5d79cae2b14af62ce1f41235297c1674721a07fd727f12fc3bc1798fc60aea\" returns successfully" Aug 13 07:14:27.030589 containerd[1991]: time="2025-08-13T07:14:27.023661188Z" level=info msg="StopPodSandbox for \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\"" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.101 [WARNING][5991] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b381d26d-9591-43bd-8a36-a31d1d40c8e1", ResourceVersion:"1007", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f", Pod:"goldmane-768f4c5c69-lw9wb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"caliad0836b7584", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.101 [INFO][5991] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.103 [INFO][5991] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" iface="eth0" netns="" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.103 [INFO][5991] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.103 [INFO][5991] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.146 [INFO][5998] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.147 [INFO][5998] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.147 [INFO][5998] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.156 [WARNING][5998] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.156 [INFO][5998] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.158 [INFO][5998] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:27.170917 containerd[1991]: 2025-08-13 07:14:27.164 [INFO][5991] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.170917 containerd[1991]: time="2025-08-13T07:14:27.170323100Z" level=info msg="TearDown network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\" successfully" Aug 13 07:14:27.170917 containerd[1991]: time="2025-08-13T07:14:27.170355821Z" level=info msg="StopPodSandbox for \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\" returns successfully" Aug 13 07:14:27.173968 containerd[1991]: time="2025-08-13T07:14:27.172121624Z" level=info msg="RemovePodSandbox for \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\"" Aug 13 07:14:27.173968 containerd[1991]: time="2025-08-13T07:14:27.172165536Z" level=info msg="Forcibly stopping sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\"" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.244 [WARNING][6012] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b381d26d-9591-43bd-8a36-a31d1d40c8e1", ResourceVersion:"1007", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f", Pod:"goldmane-768f4c5c69-lw9wb", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.68.7/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"caliad0836b7584", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.244 [INFO][6012] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.244 [INFO][6012] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" iface="eth0" netns="" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.245 [INFO][6012] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.245 [INFO][6012] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.295 [INFO][6019] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.295 [INFO][6019] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.295 [INFO][6019] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.309 [WARNING][6019] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.309 [INFO][6019] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" HandleID="k8s-pod-network.9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Workload="ip--172--31--16--138-k8s-goldmane--768f4c5c69--lw9wb-eth0" Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.311 [INFO][6019] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:27.320491 containerd[1991]: 2025-08-13 07:14:27.316 [INFO][6012] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7" Aug 13 07:14:27.324818 containerd[1991]: time="2025-08-13T07:14:27.320549972Z" level=info msg="TearDown network for sandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\" successfully" Aug 13 07:14:27.330573 containerd[1991]: time="2025-08-13T07:14:27.330368494Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:27.330573 containerd[1991]: time="2025-08-13T07:14:27.330467627Z" level=info msg="RemovePodSandbox \"9aded406c87a86b6a078d7f0cb9455e19c9a86de07315727105650e4e24b13f7\" returns successfully" Aug 13 07:14:27.331970 containerd[1991]: time="2025-08-13T07:14:27.331323367Z" level=info msg="StopPodSandbox for \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\"" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.403 [WARNING][6033] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"9f23824c-ed26-4c1b-bcd8-13d31489a3f2", ResourceVersion:"994", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90", Pod:"coredns-674b8bbfcf-gbtfq", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7cded0b467d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.404 [INFO][6033] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.404 [INFO][6033] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" iface="eth0" netns="" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.404 [INFO][6033] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.404 [INFO][6033] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.455 [INFO][6040] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.455 [INFO][6040] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.455 [INFO][6040] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.464 [WARNING][6040] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.466 [INFO][6040] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.469 [INFO][6040] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:27.476940 containerd[1991]: 2025-08-13 07:14:27.472 [INFO][6033] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.476940 containerd[1991]: time="2025-08-13T07:14:27.475537286Z" level=info msg="TearDown network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\" successfully" Aug 13 07:14:27.476940 containerd[1991]: time="2025-08-13T07:14:27.475566302Z" level=info msg="StopPodSandbox for \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\" returns successfully" Aug 13 07:14:27.476940 containerd[1991]: time="2025-08-13T07:14:27.476383602Z" level=info msg="RemovePodSandbox for \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\"" Aug 13 07:14:27.476940 containerd[1991]: time="2025-08-13T07:14:27.476419214Z" level=info msg="Forcibly stopping sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\"" Aug 13 07:14:27.558222 systemd[1]: Started sshd@9-172.31.16.138:22-147.75.109.163:50284.service - OpenSSH per-connection server daemon (147.75.109.163:50284). Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.648 [WARNING][6055] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"9f23824c-ed26-4c1b-bcd8-13d31489a3f2", ResourceVersion:"994", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"9e8b7ab678e6ce003e9c6d2118f2bf308420d030b6b419c0b78e4066acba6a90", Pod:"coredns-674b8bbfcf-gbtfq", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali7cded0b467d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.650 [INFO][6055] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.650 [INFO][6055] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" iface="eth0" netns="" Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.650 [INFO][6055] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.650 [INFO][6055] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.752 [INFO][6065] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.752 [INFO][6065] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.752 [INFO][6065] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.764 [WARNING][6065] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.764 [INFO][6065] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" HandleID="k8s-pod-network.b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--gbtfq-eth0" Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.766 [INFO][6065] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:27.773831 containerd[1991]: 2025-08-13 07:14:27.769 [INFO][6055] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69" Aug 13 07:14:27.776454 containerd[1991]: time="2025-08-13T07:14:27.775811483Z" level=info msg="TearDown network for sandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\" successfully" Aug 13 07:14:27.794876 containerd[1991]: time="2025-08-13T07:14:27.794685865Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:27.799396 containerd[1991]: time="2025-08-13T07:14:27.797078865Z" level=info msg="RemovePodSandbox \"b96f7983dd8263f77ba848b6003dc6b5f8cb61715eed3e89649d48b491e94e69\" returns successfully" Aug 13 07:14:27.800302 containerd[1991]: time="2025-08-13T07:14:27.800271813Z" level=info msg="StopPodSandbox for \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\"" Aug 13 07:14:27.881449 sshd[6062]: Accepted publickey for core from 147.75.109.163 port 50284 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:27.887461 sshd[6062]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:27.902279 systemd-logind[1961]: New session 10 of user core. Aug 13 07:14:27.904250 systemd[1]: Started session-10.scope - Session 10 of User core. Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.909 [WARNING][6080] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a57f8240-352f-45e2-8291-fe7645c155a3", ResourceVersion:"1029", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482", Pod:"csi-node-driver-782zn", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic82a766ceec", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.910 [INFO][6080] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.910 [INFO][6080] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" iface="eth0" netns="" Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.910 [INFO][6080] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.910 [INFO][6080] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.943 [INFO][6088] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.943 [INFO][6088] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.943 [INFO][6088] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.952 [WARNING][6088] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.952 [INFO][6088] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.954 [INFO][6088] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:27.959574 containerd[1991]: 2025-08-13 07:14:27.956 [INFO][6080] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:27.959574 containerd[1991]: time="2025-08-13T07:14:27.958740117Z" level=info msg="TearDown network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\" successfully" Aug 13 07:14:27.959574 containerd[1991]: time="2025-08-13T07:14:27.958767565Z" level=info msg="StopPodSandbox for \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\" returns successfully" Aug 13 07:14:27.959574 containerd[1991]: time="2025-08-13T07:14:27.959225014Z" level=info msg="RemovePodSandbox for \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\"" Aug 13 07:14:27.959574 containerd[1991]: time="2025-08-13T07:14:27.959252696Z" level=info msg="Forcibly stopping sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\"" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.044 [WARNING][6103] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"a57f8240-352f-45e2-8291-fe7645c155a3", ResourceVersion:"1029", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 46, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482", Pod:"csi-node-driver-782zn", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.68.8/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic82a766ceec", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.044 [INFO][6103] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.044 [INFO][6103] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" iface="eth0" netns="" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.044 [INFO][6103] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.044 [INFO][6103] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.109 [INFO][6115] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.110 [INFO][6115] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.110 [INFO][6115] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.120 [WARNING][6115] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.120 [INFO][6115] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" HandleID="k8s-pod-network.b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Workload="ip--172--31--16--138-k8s-csi--node--driver--782zn-eth0" Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.128 [INFO][6115] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:28.141257 containerd[1991]: 2025-08-13 07:14:28.134 [INFO][6103] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9" Aug 13 07:14:28.141257 containerd[1991]: time="2025-08-13T07:14:28.141208639Z" level=info msg="TearDown network for sandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\" successfully" Aug 13 07:14:28.166916 containerd[1991]: time="2025-08-13T07:14:28.166527485Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:28.166916 containerd[1991]: time="2025-08-13T07:14:28.166671880Z" level=info msg="RemovePodSandbox \"b0cc5415a5f6e008896d1b9eacd682e0e35ccc24cbaeabb6c48da3ab9bb9dfd9\" returns successfully" Aug 13 07:14:28.167394 containerd[1991]: time="2025-08-13T07:14:28.167309453Z" level=info msg="StopPodSandbox for \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\"" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.255 [WARNING][6129] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"0550883e-aad9-45ab-a0d8-997147ab1f77", ResourceVersion:"1041", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db", Pod:"calico-apiserver-645bc4db8-2qtgq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali84771e8c5e6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.255 [INFO][6129] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.255 [INFO][6129] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" iface="eth0" netns="" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.255 [INFO][6129] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.255 [INFO][6129] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.317 [INFO][6139] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.318 [INFO][6139] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.318 [INFO][6139] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.329 [WARNING][6139] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.329 [INFO][6139] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.336 [INFO][6139] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:28.358449 containerd[1991]: 2025-08-13 07:14:28.353 [INFO][6129] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.358449 containerd[1991]: time="2025-08-13T07:14:28.358009457Z" level=info msg="TearDown network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\" successfully" Aug 13 07:14:28.358449 containerd[1991]: time="2025-08-13T07:14:28.358063665Z" level=info msg="StopPodSandbox for \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\" returns successfully" Aug 13 07:14:28.360385 containerd[1991]: time="2025-08-13T07:14:28.360084212Z" level=info msg="RemovePodSandbox for \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\"" Aug 13 07:14:28.360385 containerd[1991]: time="2025-08-13T07:14:28.360126573Z" level=info msg="Forcibly stopping sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\"" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.429 [WARNING][6154] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"0550883e-aad9-45ab-a0d8-997147ab1f77", ResourceVersion:"1041", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"ddb95fd955b23231e5f44853fbfd41b883245081eb229bff335f51febe0b61db", Pod:"calico-apiserver-645bc4db8-2qtgq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali84771e8c5e6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.430 [INFO][6154] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.430 [INFO][6154] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" iface="eth0" netns="" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.430 [INFO][6154] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.430 [INFO][6154] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.481 [INFO][6162] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.481 [INFO][6162] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.482 [INFO][6162] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.500 [WARNING][6162] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.501 [INFO][6162] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" HandleID="k8s-pod-network.baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--2qtgq-eth0" Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.507 [INFO][6162] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:28.517407 containerd[1991]: 2025-08-13 07:14:28.513 [INFO][6154] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873" Aug 13 07:14:28.521587 containerd[1991]: time="2025-08-13T07:14:28.517442217Z" level=info msg="TearDown network for sandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\" successfully" Aug 13 07:14:28.560849 containerd[1991]: time="2025-08-13T07:14:28.560712390Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:28.561015 containerd[1991]: time="2025-08-13T07:14:28.560877383Z" level=info msg="RemovePodSandbox \"baaa55870d48173cace5e4470ee7bc2389cd2a7892533edb939d511127823873\" returns successfully" Aug 13 07:14:28.562324 containerd[1991]: time="2025-08-13T07:14:28.561438061Z" level=info msg="StopPodSandbox for \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\"" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.658 [WARNING][6176] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"e3e699d1-02d9-4ab9-afc8-37fa42f38432", ResourceVersion:"1105", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876", Pod:"calico-apiserver-645bc4db8-wmhdx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali12081f8f9f2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.659 [INFO][6176] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.659 [INFO][6176] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" iface="eth0" netns="" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.659 [INFO][6176] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.659 [INFO][6176] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.706 [INFO][6183] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.708 [INFO][6183] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.708 [INFO][6183] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.718 [WARNING][6183] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.718 [INFO][6183] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.722 [INFO][6183] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:28.737116 containerd[1991]: 2025-08-13 07:14:28.730 [INFO][6176] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.737116 containerd[1991]: time="2025-08-13T07:14:28.736931246Z" level=info msg="TearDown network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\" successfully" Aug 13 07:14:28.737116 containerd[1991]: time="2025-08-13T07:14:28.736964475Z" level=info msg="StopPodSandbox for \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\" returns successfully" Aug 13 07:14:28.752846 containerd[1991]: time="2025-08-13T07:14:28.752437367Z" level=info msg="RemovePodSandbox for \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\"" Aug 13 07:14:28.752846 containerd[1991]: time="2025-08-13T07:14:28.752486630Z" level=info msg="Forcibly stopping sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\"" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.821 [WARNING][6197] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0", GenerateName:"calico-apiserver-645bc4db8-", Namespace:"calico-apiserver", SelfLink:"", UID:"e3e699d1-02d9-4ab9-afc8-37fa42f38432", ResourceVersion:"1105", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"645bc4db8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"687f009d607ef3883f285b8f01a0788feadfe56ea066ba273106a10f34824876", Pod:"calico-apiserver-645bc4db8-wmhdx", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.68.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali12081f8f9f2", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.822 [INFO][6197] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.822 [INFO][6197] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" iface="eth0" netns="" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.822 [INFO][6197] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.822 [INFO][6197] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.861 [INFO][6205] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.861 [INFO][6205] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.861 [INFO][6205] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.879 [WARNING][6205] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.879 [INFO][6205] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" HandleID="k8s-pod-network.b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Workload="ip--172--31--16--138-k8s-calico--apiserver--645bc4db8--wmhdx-eth0" Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.886 [INFO][6205] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:28.896574 containerd[1991]: 2025-08-13 07:14:28.892 [INFO][6197] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce" Aug 13 07:14:28.899200 containerd[1991]: time="2025-08-13T07:14:28.896585198Z" level=info msg="TearDown network for sandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\" successfully" Aug 13 07:14:28.918787 containerd[1991]: time="2025-08-13T07:14:28.917348378Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:28.918787 containerd[1991]: time="2025-08-13T07:14:28.917529709Z" level=info msg="RemovePodSandbox \"b089485a2d565029b18687b44a656894281567b6b17607c2332f0161572275ce\" returns successfully" Aug 13 07:14:28.921578 containerd[1991]: time="2025-08-13T07:14:28.919864613Z" level=info msg="StopPodSandbox for \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\"" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:28.997 [WARNING][6220] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:28.998 [INFO][6220] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:28.998 [INFO][6220] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" iface="eth0" netns="" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:28.998 [INFO][6220] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:28.998 [INFO][6220] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:29.042 [INFO][6227] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:29.043 [INFO][6227] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:29.043 [INFO][6227] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:29.053 [WARNING][6227] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:29.053 [INFO][6227] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:29.055 [INFO][6227] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:29.059955 containerd[1991]: 2025-08-13 07:14:29.057 [INFO][6220] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.062035 containerd[1991]: time="2025-08-13T07:14:29.059995239Z" level=info msg="TearDown network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\" successfully" Aug 13 07:14:29.062035 containerd[1991]: time="2025-08-13T07:14:29.060019834Z" level=info msg="StopPodSandbox for \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\" returns successfully" Aug 13 07:14:29.062035 containerd[1991]: time="2025-08-13T07:14:29.060714955Z" level=info msg="RemovePodSandbox for \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\"" Aug 13 07:14:29.062035 containerd[1991]: time="2025-08-13T07:14:29.060818964Z" level=info msg="Forcibly stopping sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\"" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.134 [WARNING][6241] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" WorkloadEndpoint="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.134 [INFO][6241] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.134 [INFO][6241] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" iface="eth0" netns="" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.135 [INFO][6241] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.135 [INFO][6241] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.303 [INFO][6250] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.307 [INFO][6250] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.307 [INFO][6250] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.321 [WARNING][6250] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.321 [INFO][6250] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" HandleID="k8s-pod-network.641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Workload="ip--172--31--16--138-k8s-whisker--768487784d--kgpj7-eth0" Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.324 [INFO][6250] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:29.426358 containerd[1991]: 2025-08-13 07:14:29.331 [INFO][6241] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1" Aug 13 07:14:29.426358 containerd[1991]: time="2025-08-13T07:14:29.425979304Z" level=info msg="TearDown network for sandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\" successfully" Aug 13 07:14:29.446976 sshd[6062]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:29.457977 systemd[1]: sshd@9-172.31.16.138:22-147.75.109.163:50284.service: Deactivated successfully. Aug 13 07:14:29.460733 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 07:14:29.468300 systemd-logind[1961]: Session 10 logged out. Waiting for processes to exit. Aug 13 07:14:29.474825 containerd[1991]: time="2025-08-13T07:14:29.472787631Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:29.474825 containerd[1991]: time="2025-08-13T07:14:29.473365751Z" level=info msg="RemovePodSandbox \"641a1bce75660e63c9adff7a627d5d450b2d950d4db8a967824e2da0bf37c4d1\" returns successfully" Aug 13 07:14:29.512650 systemd[1]: Started sshd@10-172.31.16.138:22-147.75.109.163:38570.service - OpenSSH per-connection server daemon (147.75.109.163:38570). Aug 13 07:14:29.514414 systemd-logind[1961]: Removed session 10. Aug 13 07:14:29.596028 containerd[1991]: time="2025-08-13T07:14:29.595984593Z" level=info msg="StopPodSandbox for \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\"" Aug 13 07:14:29.757576 sshd[6260]: Accepted publickey for core from 147.75.109.163 port 38570 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:29.758300 sshd[6260]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:29.770577 systemd-logind[1961]: New session 11 of user core. Aug 13 07:14:29.774506 systemd[1]: Started session-11.scope - Session 11 of User core. Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.725 [WARNING][6274] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"6b61b836-afb1-4539-8d8f-b832203fca86", ResourceVersion:"998", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886", Pod:"coredns-674b8bbfcf-jkxj9", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia07661b75a4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.726 [INFO][6274] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.726 [INFO][6274] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" iface="eth0" netns="" Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.726 [INFO][6274] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.726 [INFO][6274] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.794 [INFO][6281] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.794 [INFO][6281] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.794 [INFO][6281] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.813 [WARNING][6281] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.813 [INFO][6281] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.818 [INFO][6281] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:29.825811 containerd[1991]: 2025-08-13 07:14:29.822 [INFO][6274] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:29.825811 containerd[1991]: time="2025-08-13T07:14:29.825618479Z" level=info msg="TearDown network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\" successfully" Aug 13 07:14:29.825811 containerd[1991]: time="2025-08-13T07:14:29.825668927Z" level=info msg="StopPodSandbox for \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\" returns successfully" Aug 13 07:14:29.828880 containerd[1991]: time="2025-08-13T07:14:29.827238492Z" level=info msg="RemovePodSandbox for \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\"" Aug 13 07:14:29.828880 containerd[1991]: time="2025-08-13T07:14:29.827276763Z" level=info msg="Forcibly stopping sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\"" Aug 13 07:14:29.950484 kubelet[3184]: I0813 07:14:29.949671 3184 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.927 [WARNING][6297] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"6b61b836-afb1-4539-8d8f-b832203fca86", ResourceVersion:"998", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 7, 13, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-16-138", ContainerID:"3d3bd6959ab351c1e0b4b356dae7dee2ccdaeadf9866353c67777434d60e8886", Pod:"coredns-674b8bbfcf-jkxj9", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.68.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia07661b75a4", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.929 [INFO][6297] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.929 [INFO][6297] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" iface="eth0" netns="" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.929 [INFO][6297] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.929 [INFO][6297] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.989 [INFO][6308] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.990 [INFO][6308] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:29.990 [INFO][6308] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:30.022 [WARNING][6308] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:30.022 [INFO][6308] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" HandleID="k8s-pod-network.1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Workload="ip--172--31--16--138-k8s-coredns--674b8bbfcf--jkxj9-eth0" Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:30.034 [INFO][6308] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 07:14:30.051060 containerd[1991]: 2025-08-13 07:14:30.041 [INFO][6297] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8" Aug 13 07:14:30.051060 containerd[1991]: time="2025-08-13T07:14:30.050814577Z" level=info msg="TearDown network for sandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\" successfully" Aug 13 07:14:30.069300 containerd[1991]: time="2025-08-13T07:14:30.069029468Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 07:14:30.069300 containerd[1991]: time="2025-08-13T07:14:30.069123698Z" level=info msg="RemovePodSandbox \"1fc0ffc9f6bb49e7392450f81f90cda7a5d92f7b7284616aa00a872d270bd9a8\" returns successfully" Aug 13 07:14:30.753997 sshd[6260]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:30.783325 systemd[1]: sshd@10-172.31.16.138:22-147.75.109.163:38570.service: Deactivated successfully. Aug 13 07:14:30.793008 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 07:14:30.803273 systemd-logind[1961]: Session 11 logged out. Waiting for processes to exit. Aug 13 07:14:30.817309 systemd[1]: Started sshd@11-172.31.16.138:22-147.75.109.163:38586.service - OpenSSH per-connection server daemon (147.75.109.163:38586). Aug 13 07:14:30.827806 systemd-logind[1961]: Removed session 11. Aug 13 07:14:31.119792 sshd[6321]: Accepted publickey for core from 147.75.109.163 port 38586 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:31.124678 sshd[6321]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:31.152296 systemd-logind[1961]: New session 12 of user core. Aug 13 07:14:31.158885 systemd[1]: Started session-12.scope - Session 12 of User core. Aug 13 07:14:31.682632 sshd[6321]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:31.694402 systemd[1]: sshd@11-172.31.16.138:22-147.75.109.163:38586.service: Deactivated successfully. Aug 13 07:14:31.708176 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 07:14:31.712653 systemd-logind[1961]: Session 12 logged out. Waiting for processes to exit. Aug 13 07:14:31.717101 systemd-logind[1961]: Removed session 12. Aug 13 07:14:32.527576 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2864645978.mount: Deactivated successfully. Aug 13 07:14:33.782522 containerd[1991]: time="2025-08-13T07:14:33.782335300Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.2: active requests=0, bytes read=66352308" Aug 13 07:14:33.830651 containerd[1991]: time="2025-08-13T07:14:33.830560645Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:33.929129 containerd[1991]: time="2025-08-13T07:14:33.929064404Z" level=info msg="ImageCreate event name:\"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:33.940487 containerd[1991]: time="2025-08-13T07:14:33.940441783Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:33.952565 containerd[1991]: time="2025-08-13T07:14:33.951938320Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" with image id \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\", size \"66352154\" in 9.669887236s" Aug 13 07:14:33.963928 containerd[1991]: time="2025-08-13T07:14:33.963861825Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\"" Aug 13 07:14:34.066363 containerd[1991]: time="2025-08-13T07:14:34.065714627Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Aug 13 07:14:34.331355 containerd[1991]: time="2025-08-13T07:14:34.331197492Z" level=info msg="CreateContainer within sandbox \"1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Aug 13 07:14:34.635453 containerd[1991]: time="2025-08-13T07:14:34.635310824Z" level=info msg="CreateContainer within sandbox \"1c2803276fb054a0dbae5f89e7651d8eff6bd7df8d176975512e4c7a12fe0e6f\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"21a80bf3c0679250d7cddf5f9796e4c7c8b1278b4ad7ebf43621744d6e34c7f8\"" Aug 13 07:14:34.661625 containerd[1991]: time="2025-08-13T07:14:34.661558261Z" level=info msg="StartContainer for \"21a80bf3c0679250d7cddf5f9796e4c7c8b1278b4ad7ebf43621744d6e34c7f8\"" Aug 13 07:14:34.894592 systemd[1]: run-containerd-runc-k8s.io-21a80bf3c0679250d7cddf5f9796e4c7c8b1278b4ad7ebf43621744d6e34c7f8-runc.wnROzg.mount: Deactivated successfully. Aug 13 07:14:34.906899 systemd[1]: Started cri-containerd-21a80bf3c0679250d7cddf5f9796e4c7c8b1278b4ad7ebf43621744d6e34c7f8.scope - libcontainer container 21a80bf3c0679250d7cddf5f9796e4c7c8b1278b4ad7ebf43621744d6e34c7f8. Aug 13 07:14:35.019782 containerd[1991]: time="2025-08-13T07:14:35.019692821Z" level=info msg="StartContainer for \"21a80bf3c0679250d7cddf5f9796e4c7c8b1278b4ad7ebf43621744d6e34c7f8\" returns successfully" Aug 13 07:14:35.649388 kubelet[3184]: I0813 07:14:35.542951 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-768f4c5c69-lw9wb" podStartSLOduration=30.726874389 podStartE2EDuration="50.515410206s" podCreationTimestamp="2025-08-13 07:13:45 +0000 UTC" firstStartedPulling="2025-08-13 07:14:14.266711462 +0000 UTC m=+50.061510646" lastFinishedPulling="2025-08-13 07:14:34.055247281 +0000 UTC m=+69.850046463" observedRunningTime="2025-08-13 07:14:35.404437548 +0000 UTC m=+71.199236738" watchObservedRunningTime="2025-08-13 07:14:35.515410206 +0000 UTC m=+71.310209395" Aug 13 07:14:35.917640 containerd[1991]: time="2025-08-13T07:14:35.917322216Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:35.919580 containerd[1991]: time="2025-08-13T07:14:35.919501205Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.2: active requests=0, bytes read=8759190" Aug 13 07:14:35.921614 containerd[1991]: time="2025-08-13T07:14:35.921555506Z" level=info msg="ImageCreate event name:\"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:35.924764 containerd[1991]: time="2025-08-13T07:14:35.924691799Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:35.925369 containerd[1991]: time="2025-08-13T07:14:35.925221398Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.2\" with image id \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\", size \"10251893\" in 1.858700831s" Aug 13 07:14:35.925369 containerd[1991]: time="2025-08-13T07:14:35.925254369Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\"" Aug 13 07:14:35.943748 containerd[1991]: time="2025-08-13T07:14:35.943670680Z" level=info msg="CreateContainer within sandbox \"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Aug 13 07:14:35.977563 containerd[1991]: time="2025-08-13T07:14:35.977510186Z" level=info msg="CreateContainer within sandbox \"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"c6c1f7565915504eeefe35e9cd8e5d504cee375d431c2bda5d8303b53596a17f\"" Aug 13 07:14:35.978762 containerd[1991]: time="2025-08-13T07:14:35.978626845Z" level=info msg="StartContainer for \"c6c1f7565915504eeefe35e9cd8e5d504cee375d431c2bda5d8303b53596a17f\"" Aug 13 07:14:36.038115 systemd[1]: Started cri-containerd-c6c1f7565915504eeefe35e9cd8e5d504cee375d431c2bda5d8303b53596a17f.scope - libcontainer container c6c1f7565915504eeefe35e9cd8e5d504cee375d431c2bda5d8303b53596a17f. Aug 13 07:14:36.083078 containerd[1991]: time="2025-08-13T07:14:36.083022140Z" level=info msg="StartContainer for \"c6c1f7565915504eeefe35e9cd8e5d504cee375d431c2bda5d8303b53596a17f\" returns successfully" Aug 13 07:14:36.118486 containerd[1991]: time="2025-08-13T07:14:36.111425556Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Aug 13 07:14:36.724668 systemd[1]: Started sshd@12-172.31.16.138:22-147.75.109.163:38596.service - OpenSSH per-connection server daemon (147.75.109.163:38596). Aug 13 07:14:37.009421 sshd[6457]: Accepted publickey for core from 147.75.109.163 port 38596 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:37.013455 sshd[6457]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:37.020766 systemd-logind[1961]: New session 13 of user core. Aug 13 07:14:37.026990 systemd[1]: Started session-13.scope - Session 13 of User core. Aug 13 07:14:37.444773 systemd[1]: run-containerd-runc-k8s.io-21a80bf3c0679250d7cddf5f9796e4c7c8b1278b4ad7ebf43621744d6e34c7f8-runc.Nnr1lW.mount: Deactivated successfully. Aug 13 07:14:38.045781 containerd[1991]: time="2025-08-13T07:14:38.045139879Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:38.047426 containerd[1991]: time="2025-08-13T07:14:38.047358791Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2: active requests=0, bytes read=14703784" Aug 13 07:14:38.051619 containerd[1991]: time="2025-08-13T07:14:38.050587852Z" level=info msg="ImageCreate event name:\"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:38.053741 containerd[1991]: time="2025-08-13T07:14:38.053662560Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:14:38.054880 containerd[1991]: time="2025-08-13T07:14:38.054843187Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" with image id \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\", size \"16196439\" in 1.936626582s" Aug 13 07:14:38.054985 containerd[1991]: time="2025-08-13T07:14:38.054969279Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\"" Aug 13 07:14:38.066150 containerd[1991]: time="2025-08-13T07:14:38.066092405Z" level=info msg="CreateContainer within sandbox \"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Aug 13 07:14:38.087442 containerd[1991]: time="2025-08-13T07:14:38.087401142Z" level=info msg="CreateContainer within sandbox \"89e304191e8196bc9dd2b7c68bbba02b8145d9ae4060b2b1814fb3ccda447482\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"558a45b30210d3d561f041874157507edcfba0ba75fb513786e616dcbb89decb\"" Aug 13 07:14:38.088424 containerd[1991]: time="2025-08-13T07:14:38.088398257Z" level=info msg="StartContainer for \"558a45b30210d3d561f041874157507edcfba0ba75fb513786e616dcbb89decb\"" Aug 13 07:14:38.175706 systemd[1]: Started cri-containerd-558a45b30210d3d561f041874157507edcfba0ba75fb513786e616dcbb89decb.scope - libcontainer container 558a45b30210d3d561f041874157507edcfba0ba75fb513786e616dcbb89decb. Aug 13 07:14:38.243526 containerd[1991]: time="2025-08-13T07:14:38.243477693Z" level=info msg="StartContainer for \"558a45b30210d3d561f041874157507edcfba0ba75fb513786e616dcbb89decb\" returns successfully" Aug 13 07:14:38.594035 kubelet[3184]: I0813 07:14:38.593451 3184 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-782zn" podStartSLOduration=30.37179096 podStartE2EDuration="52.57486662s" podCreationTimestamp="2025-08-13 07:13:46 +0000 UTC" firstStartedPulling="2025-08-13 07:14:15.852571456 +0000 UTC m=+51.647370641" lastFinishedPulling="2025-08-13 07:14:38.055647129 +0000 UTC m=+73.850446301" observedRunningTime="2025-08-13 07:14:38.537257832 +0000 UTC m=+74.332057012" watchObservedRunningTime="2025-08-13 07:14:38.57486662 +0000 UTC m=+74.369665809" Aug 13 07:14:38.807994 sshd[6457]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:38.816543 systemd[1]: sshd@12-172.31.16.138:22-147.75.109.163:38596.service: Deactivated successfully. Aug 13 07:14:38.821459 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 07:14:38.823768 systemd-logind[1961]: Session 13 logged out. Waiting for processes to exit. Aug 13 07:14:38.826145 systemd-logind[1961]: Removed session 13. Aug 13 07:14:38.876476 systemd[1]: Started sshd@13-172.31.16.138:22-147.75.109.163:44634.service - OpenSSH per-connection server daemon (147.75.109.163:44634). Aug 13 07:14:39.091980 sshd[6532]: Accepted publickey for core from 147.75.109.163 port 44634 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:39.094522 sshd[6532]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:39.101970 systemd-logind[1961]: New session 14 of user core. Aug 13 07:14:39.106968 systemd[1]: Started session-14.scope - Session 14 of User core. Aug 13 07:14:39.149160 kubelet[3184]: I0813 07:14:39.146970 3184 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Aug 13 07:14:39.149160 kubelet[3184]: I0813 07:14:39.147090 3184 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Aug 13 07:14:39.777506 sshd[6532]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:39.787074 systemd[1]: sshd@13-172.31.16.138:22-147.75.109.163:44634.service: Deactivated successfully. Aug 13 07:14:39.790104 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 07:14:39.791276 systemd-logind[1961]: Session 14 logged out. Waiting for processes to exit. Aug 13 07:14:39.793161 systemd-logind[1961]: Removed session 14. Aug 13 07:14:39.811101 systemd[1]: Started sshd@14-172.31.16.138:22-147.75.109.163:44646.service - OpenSSH per-connection server daemon (147.75.109.163:44646). Aug 13 07:14:40.055118 sshd[6547]: Accepted publickey for core from 147.75.109.163 port 44646 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:40.059840 sshd[6547]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:40.073142 systemd-logind[1961]: New session 15 of user core. Aug 13 07:14:40.083234 systemd[1]: Started session-15.scope - Session 15 of User core. Aug 13 07:14:41.189194 sshd[6547]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:41.197027 systemd[1]: sshd@14-172.31.16.138:22-147.75.109.163:44646.service: Deactivated successfully. Aug 13 07:14:41.202120 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 07:14:41.205732 systemd-logind[1961]: Session 15 logged out. Waiting for processes to exit. Aug 13 07:14:41.207786 systemd-logind[1961]: Removed session 15. Aug 13 07:14:41.229676 systemd[1]: Started sshd@15-172.31.16.138:22-147.75.109.163:44654.service - OpenSSH per-connection server daemon (147.75.109.163:44654). Aug 13 07:14:41.469777 sshd[6585]: Accepted publickey for core from 147.75.109.163 port 44654 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:41.473113 sshd[6585]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:41.480301 systemd-logind[1961]: New session 16 of user core. Aug 13 07:14:41.486099 systemd[1]: Started session-16.scope - Session 16 of User core. Aug 13 07:14:42.608023 sshd[6585]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:42.614951 systemd[1]: sshd@15-172.31.16.138:22-147.75.109.163:44654.service: Deactivated successfully. Aug 13 07:14:42.617742 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 07:14:42.618995 systemd-logind[1961]: Session 16 logged out. Waiting for processes to exit. Aug 13 07:14:42.621362 systemd-logind[1961]: Removed session 16. Aug 13 07:14:42.647067 systemd[1]: Started sshd@16-172.31.16.138:22-147.75.109.163:44662.service - OpenSSH per-connection server daemon (147.75.109.163:44662). Aug 13 07:14:42.888869 sshd[6598]: Accepted publickey for core from 147.75.109.163 port 44662 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:42.891325 sshd[6598]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:42.897033 systemd-logind[1961]: New session 17 of user core. Aug 13 07:14:42.902956 systemd[1]: Started session-17.scope - Session 17 of User core. Aug 13 07:14:43.290858 sshd[6598]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:43.295413 systemd-logind[1961]: Session 17 logged out. Waiting for processes to exit. Aug 13 07:14:43.295917 systemd[1]: sshd@16-172.31.16.138:22-147.75.109.163:44662.service: Deactivated successfully. Aug 13 07:14:43.311129 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 07:14:43.312494 systemd-logind[1961]: Removed session 17. Aug 13 07:14:48.346186 systemd[1]: Started sshd@17-172.31.16.138:22-147.75.109.163:33326.service - OpenSSH per-connection server daemon (147.75.109.163:33326). Aug 13 07:14:48.673978 sshd[6613]: Accepted publickey for core from 147.75.109.163 port 33326 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:48.678080 sshd[6613]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:48.690681 systemd-logind[1961]: New session 18 of user core. Aug 13 07:14:48.698046 systemd[1]: Started session-18.scope - Session 18 of User core. Aug 13 07:14:49.455281 sshd[6613]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:49.463556 systemd[1]: sshd@17-172.31.16.138:22-147.75.109.163:33326.service: Deactivated successfully. Aug 13 07:14:49.465851 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 07:14:49.466857 systemd-logind[1961]: Session 18 logged out. Waiting for processes to exit. Aug 13 07:14:49.468849 systemd-logind[1961]: Removed session 18. Aug 13 07:14:54.492162 systemd[1]: Started sshd@18-172.31.16.138:22-147.75.109.163:33336.service - OpenSSH per-connection server daemon (147.75.109.163:33336). Aug 13 07:14:54.699122 sshd[6632]: Accepted publickey for core from 147.75.109.163 port 33336 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:14:54.700660 sshd[6632]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:14:54.706229 systemd-logind[1961]: New session 19 of user core. Aug 13 07:14:54.710232 systemd[1]: Started session-19.scope - Session 19 of User core. Aug 13 07:14:54.996534 kubelet[3184]: I0813 07:14:54.996471 3184 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:14:55.759749 sshd[6632]: pam_unix(sshd:session): session closed for user core Aug 13 07:14:55.767778 systemd[1]: sshd@18-172.31.16.138:22-147.75.109.163:33336.service: Deactivated successfully. Aug 13 07:14:55.774682 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 07:14:55.778336 systemd-logind[1961]: Session 19 logged out. Waiting for processes to exit. Aug 13 07:14:55.780862 systemd-logind[1961]: Removed session 19. Aug 13 07:15:00.822314 systemd[1]: Started sshd@19-172.31.16.138:22-147.75.109.163:60838.service - OpenSSH per-connection server daemon (147.75.109.163:60838). Aug 13 07:15:01.104844 sshd[6668]: Accepted publickey for core from 147.75.109.163 port 60838 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:15:01.120605 sshd[6668]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:15:01.136115 systemd-logind[1961]: New session 20 of user core. Aug 13 07:15:01.141394 systemd[1]: Started session-20.scope - Session 20 of User core. Aug 13 07:15:04.028776 sshd[6668]: pam_unix(sshd:session): session closed for user core Aug 13 07:15:04.037607 systemd[1]: sshd@19-172.31.16.138:22-147.75.109.163:60838.service: Deactivated successfully. Aug 13 07:15:04.041705 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 07:15:04.047009 systemd-logind[1961]: Session 20 logged out. Waiting for processes to exit. Aug 13 07:15:04.060673 systemd-logind[1961]: Removed session 20. Aug 13 07:15:09.075517 systemd[1]: Started sshd@20-172.31.16.138:22-147.75.109.163:55758.service - OpenSSH per-connection server daemon (147.75.109.163:55758). Aug 13 07:15:09.369669 sshd[6706]: Accepted publickey for core from 147.75.109.163 port 55758 ssh2: RSA SHA256:EC/ch/rv0K2dityu9tU4pjM1BVuNBVPshjnRCNB2kiI Aug 13 07:15:09.374766 sshd[6706]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:15:09.384788 systemd-logind[1961]: New session 21 of user core. Aug 13 07:15:09.389965 systemd[1]: Started session-21.scope - Session 21 of User core. Aug 13 07:15:10.911707 sshd[6706]: pam_unix(sshd:session): session closed for user core Aug 13 07:15:10.916887 systemd[1]: sshd@20-172.31.16.138:22-147.75.109.163:55758.service: Deactivated successfully. Aug 13 07:15:10.924259 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 07:15:10.926326 systemd-logind[1961]: Session 21 logged out. Waiting for processes to exit. Aug 13 07:15:10.930265 systemd-logind[1961]: Removed session 21.