Apr 30 03:33:43.906692 kernel: Linux version 6.6.88-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Tue Apr 29 23:03:20 -00 2025 Apr 30 03:33:43.906717 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=c687c1f8aad1bd5ea19c342ca6f52efb69b4807a131e3bd7f3f07b950e1ec39d Apr 30 03:33:43.906730 kernel: BIOS-provided physical RAM map: Apr 30 03:33:43.906737 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Apr 30 03:33:43.906743 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000786cdfff] usable Apr 30 03:33:43.906750 kernel: BIOS-e820: [mem 0x00000000786ce000-0x00000000787cdfff] type 20 Apr 30 03:33:43.906758 kernel: BIOS-e820: [mem 0x00000000787ce000-0x000000007894dfff] reserved Apr 30 03:33:43.906765 kernel: BIOS-e820: [mem 0x000000007894e000-0x000000007895dfff] ACPI data Apr 30 03:33:43.906772 kernel: BIOS-e820: [mem 0x000000007895e000-0x00000000789ddfff] ACPI NVS Apr 30 03:33:43.906781 kernel: BIOS-e820: [mem 0x00000000789de000-0x000000007c97bfff] usable Apr 30 03:33:43.906788 kernel: BIOS-e820: [mem 0x000000007c97c000-0x000000007c9fffff] reserved Apr 30 03:33:43.906795 kernel: NX (Execute Disable) protection: active Apr 30 03:33:43.906801 kernel: APIC: Static calls initialized Apr 30 03:33:43.906808 kernel: efi: EFI v2.7 by EDK II Apr 30 03:33:43.906817 kernel: efi: SMBIOS=0x7886a000 ACPI=0x7895d000 ACPI 2.0=0x7895d014 MEMATTR=0x77003518 Apr 30 03:33:43.906827 kernel: SMBIOS 2.7 present. Apr 30 03:33:43.906835 kernel: DMI: Amazon EC2 t3.small/, BIOS 1.0 10/16/2017 Apr 30 03:33:43.906843 kernel: Hypervisor detected: KVM Apr 30 03:33:43.906850 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Apr 30 03:33:43.906858 kernel: kvm-clock: using sched offset of 3815370530 cycles Apr 30 03:33:43.906866 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Apr 30 03:33:43.906874 kernel: tsc: Detected 2499.996 MHz processor Apr 30 03:33:43.906882 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Apr 30 03:33:43.906890 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Apr 30 03:33:43.906898 kernel: last_pfn = 0x7c97c max_arch_pfn = 0x400000000 Apr 30 03:33:43.906908 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Apr 30 03:33:43.906916 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Apr 30 03:33:43.906924 kernel: Using GB pages for direct mapping Apr 30 03:33:43.906931 kernel: Secure boot disabled Apr 30 03:33:43.906939 kernel: ACPI: Early table checksum verification disabled Apr 30 03:33:43.906947 kernel: ACPI: RSDP 0x000000007895D014 000024 (v02 AMAZON) Apr 30 03:33:43.906954 kernel: ACPI: XSDT 0x000000007895C0E8 00006C (v01 AMAZON AMZNFACP 00000001 01000013) Apr 30 03:33:43.906962 kernel: ACPI: FACP 0x0000000078955000 000114 (v01 AMAZON AMZNFACP 00000001 AMZN 00000001) Apr 30 03:33:43.906970 kernel: ACPI: DSDT 0x0000000078956000 00115A (v01 AMAZON AMZNDSDT 00000001 AMZN 00000001) Apr 30 03:33:43.906980 kernel: ACPI: FACS 0x00000000789D0000 000040 Apr 30 03:33:43.906987 kernel: ACPI: WAET 0x000000007895B000 000028 (v01 AMAZON AMZNWAET 00000001 AMZN 00000001) Apr 30 03:33:43.906995 kernel: ACPI: SLIT 0x000000007895A000 00006C (v01 AMAZON AMZNSLIT 00000001 AMZN 00000001) Apr 30 03:33:43.907003 kernel: ACPI: APIC 0x0000000078959000 000076 (v01 AMAZON AMZNAPIC 00000001 AMZN 00000001) Apr 30 03:33:43.907010 kernel: ACPI: SRAT 0x0000000078958000 0000A0 (v01 AMAZON AMZNSRAT 00000001 AMZN 00000001) Apr 30 03:33:43.907018 kernel: ACPI: HPET 0x0000000078954000 000038 (v01 AMAZON AMZNHPET 00000001 AMZN 00000001) Apr 30 03:33:43.907030 kernel: ACPI: SSDT 0x0000000078953000 000759 (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Apr 30 03:33:43.907040 kernel: ACPI: SSDT 0x0000000078952000 00007F (v01 AMAZON AMZNSSDT 00000001 AMZN 00000001) Apr 30 03:33:43.907048 kernel: ACPI: BGRT 0x0000000078951000 000038 (v01 AMAZON AMAZON 00000002 01000013) Apr 30 03:33:43.907057 kernel: ACPI: Reserving FACP table memory at [mem 0x78955000-0x78955113] Apr 30 03:33:43.907065 kernel: ACPI: Reserving DSDT table memory at [mem 0x78956000-0x78957159] Apr 30 03:33:43.907073 kernel: ACPI: Reserving FACS table memory at [mem 0x789d0000-0x789d003f] Apr 30 03:33:43.907082 kernel: ACPI: Reserving WAET table memory at [mem 0x7895b000-0x7895b027] Apr 30 03:33:43.907092 kernel: ACPI: Reserving SLIT table memory at [mem 0x7895a000-0x7895a06b] Apr 30 03:33:43.907100 kernel: ACPI: Reserving APIC table memory at [mem 0x78959000-0x78959075] Apr 30 03:33:43.907108 kernel: ACPI: Reserving SRAT table memory at [mem 0x78958000-0x7895809f] Apr 30 03:33:43.907116 kernel: ACPI: Reserving HPET table memory at [mem 0x78954000-0x78954037] Apr 30 03:33:43.907124 kernel: ACPI: Reserving SSDT table memory at [mem 0x78953000-0x78953758] Apr 30 03:33:43.907132 kernel: ACPI: Reserving SSDT table memory at [mem 0x78952000-0x7895207e] Apr 30 03:33:43.907140 kernel: ACPI: Reserving BGRT table memory at [mem 0x78951000-0x78951037] Apr 30 03:33:43.907148 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Apr 30 03:33:43.907157 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Apr 30 03:33:43.907165 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x7fffffff] Apr 30 03:33:43.907175 kernel: NUMA: Initialized distance table, cnt=1 Apr 30 03:33:43.907183 kernel: NODE_DATA(0) allocated [mem 0x7a8ef000-0x7a8f4fff] Apr 30 03:33:43.907191 kernel: Zone ranges: Apr 30 03:33:43.907200 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Apr 30 03:33:43.907208 kernel: DMA32 [mem 0x0000000001000000-0x000000007c97bfff] Apr 30 03:33:43.907216 kernel: Normal empty Apr 30 03:33:43.907224 kernel: Movable zone start for each node Apr 30 03:33:43.907232 kernel: Early memory node ranges Apr 30 03:33:43.907240 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Apr 30 03:33:43.907251 kernel: node 0: [mem 0x0000000000100000-0x00000000786cdfff] Apr 30 03:33:43.907259 kernel: node 0: [mem 0x00000000789de000-0x000000007c97bfff] Apr 30 03:33:43.907267 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007c97bfff] Apr 30 03:33:43.907275 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Apr 30 03:33:43.907284 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Apr 30 03:33:43.907292 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Apr 30 03:33:43.907300 kernel: On node 0, zone DMA32: 13956 pages in unavailable ranges Apr 30 03:33:43.907308 kernel: ACPI: PM-Timer IO Port: 0xb008 Apr 30 03:33:43.907317 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Apr 30 03:33:43.907327 kernel: IOAPIC[0]: apic_id 0, version 32, address 0xfec00000, GSI 0-23 Apr 30 03:33:43.907335 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Apr 30 03:33:43.907343 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Apr 30 03:33:43.907351 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Apr 30 03:33:43.907360 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Apr 30 03:33:43.907368 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Apr 30 03:33:43.907376 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Apr 30 03:33:43.907384 kernel: TSC deadline timer available Apr 30 03:33:43.907392 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Apr 30 03:33:43.907413 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Apr 30 03:33:43.907425 kernel: [mem 0x7ca00000-0xffffffff] available for PCI devices Apr 30 03:33:43.907433 kernel: Booting paravirtualized kernel on KVM Apr 30 03:33:43.907441 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Apr 30 03:33:43.907450 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:2 nr_cpu_ids:2 nr_node_ids:1 Apr 30 03:33:43.907458 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u1048576 Apr 30 03:33:43.907466 kernel: pcpu-alloc: s197096 r8192 d32280 u1048576 alloc=1*2097152 Apr 30 03:33:43.907474 kernel: pcpu-alloc: [0] 0 1 Apr 30 03:33:43.907482 kernel: kvm-guest: PV spinlocks enabled Apr 30 03:33:43.907490 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Apr 30 03:33:43.907503 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=c687c1f8aad1bd5ea19c342ca6f52efb69b4807a131e3bd7f3f07b950e1ec39d Apr 30 03:33:43.907512 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Apr 30 03:33:43.907520 kernel: random: crng init done Apr 30 03:33:43.907528 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 30 03:33:43.907536 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Apr 30 03:33:43.907544 kernel: Fallback order for Node 0: 0 Apr 30 03:33:43.907552 kernel: Built 1 zonelists, mobility grouping on. Total pages: 501318 Apr 30 03:33:43.907561 kernel: Policy zone: DMA32 Apr 30 03:33:43.907571 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 30 03:33:43.907580 kernel: Memory: 1874608K/2037804K available (12288K kernel code, 2295K rwdata, 22740K rodata, 42864K init, 2328K bss, 162936K reserved, 0K cma-reserved) Apr 30 03:33:43.907588 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 30 03:33:43.907596 kernel: Kernel/User page tables isolation: enabled Apr 30 03:33:43.907604 kernel: ftrace: allocating 37944 entries in 149 pages Apr 30 03:33:43.907613 kernel: ftrace: allocated 149 pages with 4 groups Apr 30 03:33:43.907621 kernel: Dynamic Preempt: voluntary Apr 30 03:33:43.907629 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 30 03:33:43.907638 kernel: rcu: RCU event tracing is enabled. Apr 30 03:33:43.907649 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 30 03:33:43.907658 kernel: Trampoline variant of Tasks RCU enabled. Apr 30 03:33:43.907666 kernel: Rude variant of Tasks RCU enabled. Apr 30 03:33:43.907674 kernel: Tracing variant of Tasks RCU enabled. Apr 30 03:33:43.907682 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 30 03:33:43.907691 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 30 03:33:43.907699 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Apr 30 03:33:43.907718 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 30 03:33:43.907727 kernel: Console: colour dummy device 80x25 Apr 30 03:33:43.907736 kernel: printk: console [tty0] enabled Apr 30 03:33:43.907744 kernel: printk: console [ttyS0] enabled Apr 30 03:33:43.907756 kernel: ACPI: Core revision 20230628 Apr 30 03:33:43.907765 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 30580167144 ns Apr 30 03:33:43.907774 kernel: APIC: Switch to symmetric I/O mode setup Apr 30 03:33:43.907783 kernel: x2apic enabled Apr 30 03:33:43.907791 kernel: APIC: Switched APIC routing to: physical x2apic Apr 30 03:33:43.907800 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Apr 30 03:33:43.907812 kernel: Calibrating delay loop (skipped) preset value.. 4999.99 BogoMIPS (lpj=2499996) Apr 30 03:33:43.907821 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Apr 30 03:33:43.907830 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Apr 30 03:33:43.907839 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Apr 30 03:33:43.907847 kernel: Spectre V2 : Mitigation: Retpolines Apr 30 03:33:43.907856 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Apr 30 03:33:43.907864 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Apr 30 03:33:43.907873 kernel: RETBleed: WARNING: Spectre v2 mitigation leaves CPU vulnerable to RETBleed attacks, data leaks possible! Apr 30 03:33:43.907882 kernel: RETBleed: Vulnerable Apr 30 03:33:43.907893 kernel: Speculative Store Bypass: Vulnerable Apr 30 03:33:43.907902 kernel: MDS: Vulnerable: Clear CPU buffers attempted, no microcode Apr 30 03:33:43.907910 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Apr 30 03:33:43.907919 kernel: GDS: Unknown: Dependent on hypervisor status Apr 30 03:33:43.907928 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Apr 30 03:33:43.907937 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Apr 30 03:33:43.907945 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Apr 30 03:33:43.907954 kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers' Apr 30 03:33:43.907963 kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR' Apr 30 03:33:43.907972 kernel: x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' Apr 30 03:33:43.907980 kernel: x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' Apr 30 03:33:43.907991 kernel: x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' Apr 30 03:33:43.908000 kernel: x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' Apr 30 03:33:43.908009 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Apr 30 03:33:43.908017 kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64 Apr 30 03:33:43.908026 kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64 Apr 30 03:33:43.908035 kernel: x86/fpu: xstate_offset[5]: 960, xstate_sizes[5]: 64 Apr 30 03:33:43.908044 kernel: x86/fpu: xstate_offset[6]: 1024, xstate_sizes[6]: 512 Apr 30 03:33:43.908052 kernel: x86/fpu: xstate_offset[7]: 1536, xstate_sizes[7]: 1024 Apr 30 03:33:43.908061 kernel: x86/fpu: xstate_offset[9]: 2560, xstate_sizes[9]: 8 Apr 30 03:33:43.908070 kernel: x86/fpu: Enabled xstate features 0x2ff, context size is 2568 bytes, using 'compacted' format. Apr 30 03:33:43.908079 kernel: Freeing SMP alternatives memory: 32K Apr 30 03:33:43.908090 kernel: pid_max: default: 32768 minimum: 301 Apr 30 03:33:43.908099 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 30 03:33:43.908108 kernel: landlock: Up and running. Apr 30 03:33:43.908116 kernel: SELinux: Initializing. Apr 30 03:33:43.908125 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Apr 30 03:33:43.908134 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Apr 30 03:33:43.908142 kernel: smpboot: CPU0: Intel(R) Xeon(R) Platinum 8259CL CPU @ 2.50GHz (family: 0x6, model: 0x55, stepping: 0x7) Apr 30 03:33:43.908151 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 30 03:33:43.908160 kernel: RCU Tasks Rude: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 30 03:33:43.908169 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Apr 30 03:33:43.908178 kernel: Performance Events: unsupported p6 CPU model 85 no PMU driver, software events only. Apr 30 03:33:43.908190 kernel: signal: max sigframe size: 3632 Apr 30 03:33:43.908199 kernel: rcu: Hierarchical SRCU implementation. Apr 30 03:33:43.908208 kernel: rcu: Max phase no-delay instances is 400. Apr 30 03:33:43.908217 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Apr 30 03:33:43.908226 kernel: smp: Bringing up secondary CPUs ... Apr 30 03:33:43.908234 kernel: smpboot: x86: Booting SMP configuration: Apr 30 03:33:43.908243 kernel: .... node #0, CPUs: #1 Apr 30 03:33:43.908253 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Apr 30 03:33:43.908262 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Apr 30 03:33:43.908273 kernel: smp: Brought up 1 node, 2 CPUs Apr 30 03:33:43.908282 kernel: smpboot: Max logical packages: 1 Apr 30 03:33:43.908291 kernel: smpboot: Total of 2 processors activated (9999.98 BogoMIPS) Apr 30 03:33:43.908300 kernel: devtmpfs: initialized Apr 30 03:33:43.908309 kernel: x86/mm: Memory block size: 128MB Apr 30 03:33:43.908318 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7895e000-0x789ddfff] (524288 bytes) Apr 30 03:33:43.908327 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 30 03:33:43.908336 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 30 03:33:43.908347 kernel: pinctrl core: initialized pinctrl subsystem Apr 30 03:33:43.908356 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 30 03:33:43.908365 kernel: audit: initializing netlink subsys (disabled) Apr 30 03:33:43.908374 kernel: audit: type=2000 audit(1745984023.281:1): state=initialized audit_enabled=0 res=1 Apr 30 03:33:43.908382 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 30 03:33:43.908391 kernel: thermal_sys: Registered thermal governor 'user_space' Apr 30 03:33:43.908410 kernel: cpuidle: using governor menu Apr 30 03:33:43.908419 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 30 03:33:43.908428 kernel: dca service started, version 1.12.1 Apr 30 03:33:43.908440 kernel: PCI: Using configuration type 1 for base access Apr 30 03:33:43.908449 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Apr 30 03:33:43.908458 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 30 03:33:43.908467 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Apr 30 03:33:43.908476 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 30 03:33:43.908484 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Apr 30 03:33:43.908493 kernel: ACPI: Added _OSI(Module Device) Apr 30 03:33:43.908503 kernel: ACPI: Added _OSI(Processor Device) Apr 30 03:33:43.908511 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Apr 30 03:33:43.908524 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 30 03:33:43.908533 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Apr 30 03:33:43.908542 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Apr 30 03:33:43.908551 kernel: ACPI: Interpreter enabled Apr 30 03:33:43.908559 kernel: ACPI: PM: (supports S0 S5) Apr 30 03:33:43.908568 kernel: ACPI: Using IOAPIC for interrupt routing Apr 30 03:33:43.908577 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Apr 30 03:33:43.908586 kernel: PCI: Using E820 reservations for host bridge windows Apr 30 03:33:43.908594 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Apr 30 03:33:43.908606 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 30 03:33:43.908773 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Apr 30 03:33:43.908877 kernel: acpi PNP0A03:00: _OSC: not requesting OS control; OS requires [ExtendedConfig ASPM ClockPM MSI] Apr 30 03:33:43.908971 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended configuration space under this bridge Apr 30 03:33:43.908982 kernel: acpiphp: Slot [3] registered Apr 30 03:33:43.908991 kernel: acpiphp: Slot [4] registered Apr 30 03:33:43.909000 kernel: acpiphp: Slot [5] registered Apr 30 03:33:43.909009 kernel: acpiphp: Slot [6] registered Apr 30 03:33:43.909022 kernel: acpiphp: Slot [7] registered Apr 30 03:33:43.909030 kernel: acpiphp: Slot [8] registered Apr 30 03:33:43.909039 kernel: acpiphp: Slot [9] registered Apr 30 03:33:43.909048 kernel: acpiphp: Slot [10] registered Apr 30 03:33:43.909057 kernel: acpiphp: Slot [11] registered Apr 30 03:33:43.909065 kernel: acpiphp: Slot [12] registered Apr 30 03:33:43.909074 kernel: acpiphp: Slot [13] registered Apr 30 03:33:43.909083 kernel: acpiphp: Slot [14] registered Apr 30 03:33:43.909092 kernel: acpiphp: Slot [15] registered Apr 30 03:33:43.909104 kernel: acpiphp: Slot [16] registered Apr 30 03:33:43.909113 kernel: acpiphp: Slot [17] registered Apr 30 03:33:43.909122 kernel: acpiphp: Slot [18] registered Apr 30 03:33:43.909130 kernel: acpiphp: Slot [19] registered Apr 30 03:33:43.909139 kernel: acpiphp: Slot [20] registered Apr 30 03:33:43.909148 kernel: acpiphp: Slot [21] registered Apr 30 03:33:43.909157 kernel: acpiphp: Slot [22] registered Apr 30 03:33:43.909165 kernel: acpiphp: Slot [23] registered Apr 30 03:33:43.909174 kernel: acpiphp: Slot [24] registered Apr 30 03:33:43.909185 kernel: acpiphp: Slot [25] registered Apr 30 03:33:43.909194 kernel: acpiphp: Slot [26] registered Apr 30 03:33:43.909202 kernel: acpiphp: Slot [27] registered Apr 30 03:33:43.909211 kernel: acpiphp: Slot [28] registered Apr 30 03:33:43.909220 kernel: acpiphp: Slot [29] registered Apr 30 03:33:43.909228 kernel: acpiphp: Slot [30] registered Apr 30 03:33:43.909237 kernel: acpiphp: Slot [31] registered Apr 30 03:33:43.909246 kernel: PCI host bridge to bus 0000:00 Apr 30 03:33:43.909345 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Apr 30 03:33:43.909461 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Apr 30 03:33:43.909547 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Apr 30 03:33:43.909631 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Apr 30 03:33:43.909715 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x2000ffffffff window] Apr 30 03:33:43.909797 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 30 03:33:43.909909 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Apr 30 03:33:43.910013 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Apr 30 03:33:43.910122 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x000000 Apr 30 03:33:43.910215 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Apr 30 03:33:43.910309 kernel: pci 0000:00:01.3: PIIX4 devres E PIO at fff0-ffff Apr 30 03:33:43.910528 kernel: pci 0000:00:01.3: PIIX4 devres F MMIO at ffc00000-ffffffff Apr 30 03:33:43.910636 kernel: pci 0000:00:01.3: PIIX4 devres G PIO at fff0-ffff Apr 30 03:33:43.910729 kernel: pci 0000:00:01.3: PIIX4 devres H MMIO at ffc00000-ffffffff Apr 30 03:33:43.910828 kernel: pci 0000:00:01.3: PIIX4 devres I PIO at fff0-ffff Apr 30 03:33:43.910920 kernel: pci 0000:00:01.3: PIIX4 devres J PIO at fff0-ffff Apr 30 03:33:43.911019 kernel: pci 0000:00:03.0: [1d0f:1111] type 00 class 0x030000 Apr 30 03:33:43.911113 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x80000000-0x803fffff pref] Apr 30 03:33:43.911207 kernel: pci 0000:00:03.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Apr 30 03:33:43.911298 kernel: pci 0000:00:03.0: BAR 0: assigned to efifb Apr 30 03:33:43.911392 kernel: pci 0000:00:03.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Apr 30 03:33:43.911531 kernel: pci 0000:00:04.0: [1d0f:8061] type 00 class 0x010802 Apr 30 03:33:43.911622 kernel: pci 0000:00:04.0: reg 0x10: [mem 0x80404000-0x80407fff] Apr 30 03:33:43.911718 kernel: pci 0000:00:05.0: [1d0f:ec20] type 00 class 0x020000 Apr 30 03:33:43.911809 kernel: pci 0000:00:05.0: reg 0x10: [mem 0x80400000-0x80403fff] Apr 30 03:33:43.911821 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Apr 30 03:33:43.911831 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Apr 30 03:33:43.911840 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Apr 30 03:33:43.911852 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Apr 30 03:33:43.911861 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Apr 30 03:33:43.911870 kernel: iommu: Default domain type: Translated Apr 30 03:33:43.911879 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Apr 30 03:33:43.911887 kernel: efivars: Registered efivars operations Apr 30 03:33:43.911897 kernel: PCI: Using ACPI for IRQ routing Apr 30 03:33:43.911906 kernel: PCI: pci_cache_line_size set to 64 bytes Apr 30 03:33:43.911914 kernel: e820: reserve RAM buffer [mem 0x786ce000-0x7bffffff] Apr 30 03:33:43.911923 kernel: e820: reserve RAM buffer [mem 0x7c97c000-0x7fffffff] Apr 30 03:33:43.912019 kernel: pci 0000:00:03.0: vgaarb: setting as boot VGA device Apr 30 03:33:43.912112 kernel: pci 0000:00:03.0: vgaarb: bridge control possible Apr 30 03:33:43.912204 kernel: pci 0000:00:03.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Apr 30 03:33:43.912216 kernel: vgaarb: loaded Apr 30 03:33:43.912226 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0 Apr 30 03:33:43.912235 kernel: hpet0: 8 comparators, 32-bit 62.500000 MHz counter Apr 30 03:33:43.912244 kernel: clocksource: Switched to clocksource kvm-clock Apr 30 03:33:43.912253 kernel: VFS: Disk quotas dquot_6.6.0 Apr 30 03:33:43.912262 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 30 03:33:43.912275 kernel: pnp: PnP ACPI init Apr 30 03:33:43.912284 kernel: pnp: PnP ACPI: found 5 devices Apr 30 03:33:43.912293 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Apr 30 03:33:43.912302 kernel: NET: Registered PF_INET protocol family Apr 30 03:33:43.912311 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 30 03:33:43.912321 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Apr 30 03:33:43.912330 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 30 03:33:43.912339 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Apr 30 03:33:43.912350 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Apr 30 03:33:43.912359 kernel: TCP: Hash tables configured (established 16384 bind 16384) Apr 30 03:33:43.912368 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Apr 30 03:33:43.912377 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Apr 30 03:33:43.912386 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 30 03:33:43.912395 kernel: NET: Registered PF_XDP protocol family Apr 30 03:33:43.912511 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Apr 30 03:33:43.912596 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Apr 30 03:33:43.912679 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Apr 30 03:33:43.912768 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Apr 30 03:33:43.912850 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x2000ffffffff window] Apr 30 03:33:43.912947 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Apr 30 03:33:43.912960 kernel: PCI: CLS 0 bytes, default 64 Apr 30 03:33:43.912969 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Apr 30 03:33:43.912978 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x24093623c91, max_idle_ns: 440795291220 ns Apr 30 03:33:43.912987 kernel: clocksource: Switched to clocksource tsc Apr 30 03:33:43.912997 kernel: Initialise system trusted keyrings Apr 30 03:33:43.913010 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Apr 30 03:33:43.913019 kernel: Key type asymmetric registered Apr 30 03:33:43.913027 kernel: Asymmetric key parser 'x509' registered Apr 30 03:33:43.913043 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Apr 30 03:33:43.913058 kernel: io scheduler mq-deadline registered Apr 30 03:33:43.913080 kernel: io scheduler kyber registered Apr 30 03:33:43.913093 kernel: io scheduler bfq registered Apr 30 03:33:43.913102 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Apr 30 03:33:43.913111 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 30 03:33:43.913121 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Apr 30 03:33:43.913133 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Apr 30 03:33:43.913142 kernel: i8042: Warning: Keylock active Apr 30 03:33:43.913151 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Apr 30 03:33:43.913159 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Apr 30 03:33:43.913267 kernel: rtc_cmos 00:00: RTC can wake from S4 Apr 30 03:33:43.913354 kernel: rtc_cmos 00:00: registered as rtc0 Apr 30 03:33:43.913497 kernel: rtc_cmos 00:00: setting system clock to 2025-04-30T03:33:43 UTC (1745984023) Apr 30 03:33:43.913587 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Apr 30 03:33:43.913598 kernel: intel_pstate: CPU model not supported Apr 30 03:33:43.913607 kernel: efifb: probing for efifb Apr 30 03:33:43.913616 kernel: efifb: framebuffer at 0x80000000, using 1920k, total 1920k Apr 30 03:33:43.913625 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Apr 30 03:33:43.913634 kernel: efifb: scrolling: redraw Apr 30 03:33:43.913643 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 30 03:33:43.913652 kernel: Console: switching to colour frame buffer device 100x37 Apr 30 03:33:43.913661 kernel: fb0: EFI VGA frame buffer device Apr 30 03:33:43.913673 kernel: pstore: Using crash dump compression: deflate Apr 30 03:33:43.913682 kernel: pstore: Registered efi_pstore as persistent store backend Apr 30 03:33:43.913691 kernel: NET: Registered PF_INET6 protocol family Apr 30 03:33:43.913700 kernel: Segment Routing with IPv6 Apr 30 03:33:43.913709 kernel: In-situ OAM (IOAM) with IPv6 Apr 30 03:33:43.913717 kernel: NET: Registered PF_PACKET protocol family Apr 30 03:33:43.913727 kernel: Key type dns_resolver registered Apr 30 03:33:43.913754 kernel: IPI shorthand broadcast: enabled Apr 30 03:33:43.913765 kernel: sched_clock: Marking stable (465002626, 127426547)->(661325203, -68896030) Apr 30 03:33:43.913778 kernel: registered taskstats version 1 Apr 30 03:33:43.913787 kernel: Loading compiled-in X.509 certificates Apr 30 03:33:43.913797 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.88-flatcar: 4a2605119c3649b55d5796c3fe312b2581bff37b' Apr 30 03:33:43.913806 kernel: Key type .fscrypt registered Apr 30 03:33:43.913815 kernel: Key type fscrypt-provisioning registered Apr 30 03:33:43.913824 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 30 03:33:43.913834 kernel: ima: Allocated hash algorithm: sha1 Apr 30 03:33:43.913843 kernel: ima: No architecture policies found Apr 30 03:33:43.913852 kernel: clk: Disabling unused clocks Apr 30 03:33:43.913864 kernel: Freeing unused kernel image (initmem) memory: 42864K Apr 30 03:33:43.913873 kernel: Write protecting the kernel read-only data: 36864k Apr 30 03:33:43.913883 kernel: Freeing unused kernel image (rodata/data gap) memory: 1836K Apr 30 03:33:43.913892 kernel: Run /init as init process Apr 30 03:33:43.913901 kernel: with arguments: Apr 30 03:33:43.913910 kernel: /init Apr 30 03:33:43.913920 kernel: with environment: Apr 30 03:33:43.913929 kernel: HOME=/ Apr 30 03:33:43.913938 kernel: TERM=linux Apr 30 03:33:43.913950 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Apr 30 03:33:43.913962 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 30 03:33:43.913974 systemd[1]: Detected virtualization amazon. Apr 30 03:33:43.913984 systemd[1]: Detected architecture x86-64. Apr 30 03:33:43.913993 systemd[1]: Running in initrd. Apr 30 03:33:43.914003 systemd[1]: No hostname configured, using default hostname. Apr 30 03:33:43.914012 systemd[1]: Hostname set to . Apr 30 03:33:43.914025 systemd[1]: Initializing machine ID from VM UUID. Apr 30 03:33:43.914034 systemd[1]: Queued start job for default target initrd.target. Apr 30 03:33:43.914044 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 03:33:43.914053 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 03:33:43.914064 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 30 03:33:43.914074 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 30 03:33:43.914083 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 30 03:33:43.914096 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 30 03:33:43.914106 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 30 03:33:43.914116 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 30 03:33:43.914126 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 03:33:43.914136 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 30 03:33:43.914148 systemd[1]: Reached target paths.target - Path Units. Apr 30 03:33:43.914158 systemd[1]: Reached target slices.target - Slice Units. Apr 30 03:33:43.914167 systemd[1]: Reached target swap.target - Swaps. Apr 30 03:33:43.914177 systemd[1]: Reached target timers.target - Timer Units. Apr 30 03:33:43.914187 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 03:33:43.914196 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 03:33:43.914206 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 30 03:33:43.914216 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Apr 30 03:33:43.914228 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 30 03:33:43.914238 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 30 03:33:43.914247 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 03:33:43.914258 systemd[1]: Reached target sockets.target - Socket Units. Apr 30 03:33:43.914267 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 30 03:33:43.914277 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 30 03:33:43.914286 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 30 03:33:43.914296 systemd[1]: Starting systemd-fsck-usr.service... Apr 30 03:33:43.914306 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 30 03:33:43.914318 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 30 03:33:43.914328 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 03:33:43.914338 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 30 03:33:43.914347 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 03:33:43.914357 systemd[1]: Finished systemd-fsck-usr.service. Apr 30 03:33:43.914390 systemd-journald[178]: Collecting audit messages is disabled. Apr 30 03:33:43.917143 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 30 03:33:43.917159 systemd-journald[178]: Journal started Apr 30 03:33:43.917190 systemd-journald[178]: Runtime Journal (/run/log/journal/ec2604ebf32ac5bac3fe163d0660277f) is 4.7M, max 38.2M, 33.4M free. Apr 30 03:33:43.917445 systemd-modules-load[179]: Inserted module 'overlay' Apr 30 03:33:43.920451 systemd[1]: Started systemd-journald.service - Journal Service. Apr 30 03:33:43.920935 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 03:33:43.928702 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 03:33:43.931559 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 30 03:33:43.935649 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 30 03:33:43.949625 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 30 03:33:43.952147 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 03:33:43.954542 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 30 03:33:43.959438 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 30 03:33:43.969426 kernel: Bridge firewalling registered Apr 30 03:33:43.970509 systemd-modules-load[179]: Inserted module 'br_netfilter' Apr 30 03:33:43.974533 dracut-cmdline[207]: dracut-dracut-053 Apr 30 03:33:43.974533 dracut-cmdline[207]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=ec2 modprobe.blacklist=xen_fbfront net.ifnames=0 nvme_core.io_timeout=4294967295 verity.usrhash=c687c1f8aad1bd5ea19c342ca6f52efb69b4807a131e3bd7f3f07b950e1ec39d Apr 30 03:33:43.975130 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 30 03:33:43.975997 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 03:33:43.977647 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 03:33:43.986701 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 30 03:33:43.994623 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 30 03:33:44.003467 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Apr 30 03:33:44.007674 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 30 03:33:44.028646 systemd-resolved[251]: Positive Trust Anchors: Apr 30 03:33:44.029263 systemd-resolved[251]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 30 03:33:44.029303 systemd-resolved[251]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 30 03:33:44.034875 systemd-resolved[251]: Defaulting to hostname 'linux'. Apr 30 03:33:44.035846 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 30 03:33:44.036279 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 30 03:33:44.048431 kernel: SCSI subsystem initialized Apr 30 03:33:44.058424 kernel: Loading iSCSI transport class v2.0-870. Apr 30 03:33:44.069430 kernel: iscsi: registered transport (tcp) Apr 30 03:33:44.091945 kernel: iscsi: registered transport (qla4xxx) Apr 30 03:33:44.092015 kernel: QLogic iSCSI HBA Driver Apr 30 03:33:44.134054 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 30 03:33:44.141599 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 30 03:33:44.167700 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 30 03:33:44.167785 kernel: device-mapper: uevent: version 1.0.3 Apr 30 03:33:44.167809 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 30 03:33:44.210436 kernel: raid6: avx512x4 gen() 18109 MB/s Apr 30 03:33:44.228424 kernel: raid6: avx512x2 gen() 18241 MB/s Apr 30 03:33:44.246426 kernel: raid6: avx512x1 gen() 16995 MB/s Apr 30 03:33:44.264424 kernel: raid6: avx2x4 gen() 18133 MB/s Apr 30 03:33:44.281426 kernel: raid6: avx2x2 gen() 18131 MB/s Apr 30 03:33:44.299643 kernel: raid6: avx2x1 gen() 13741 MB/s Apr 30 03:33:44.299693 kernel: raid6: using algorithm avx512x2 gen() 18241 MB/s Apr 30 03:33:44.318544 kernel: raid6: .... xor() 24543 MB/s, rmw enabled Apr 30 03:33:44.318615 kernel: raid6: using avx512x2 recovery algorithm Apr 30 03:33:44.340435 kernel: xor: automatically using best checksumming function avx Apr 30 03:33:44.504434 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 30 03:33:44.515165 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 30 03:33:44.524636 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 03:33:44.537633 systemd-udevd[399]: Using default interface naming scheme 'v255'. Apr 30 03:33:44.542837 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 03:33:44.552612 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 30 03:33:44.571301 dracut-pre-trigger[405]: rd.md=0: removing MD RAID activation Apr 30 03:33:44.602646 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 03:33:44.609606 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 30 03:33:44.660398 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 03:33:44.670641 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 30 03:33:44.686792 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 30 03:33:44.688291 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 03:33:44.688830 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 03:33:44.689306 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 30 03:33:44.696705 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 30 03:33:44.728149 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 30 03:33:44.764446 kernel: cryptd: max_cpu_qlen set to 1000 Apr 30 03:33:44.773506 kernel: ena 0000:00:05.0: ENA device version: 0.10 Apr 30 03:33:44.811297 kernel: ena 0000:00:05.0: ENA controller version: 0.0.1 implementation version 1 Apr 30 03:33:44.811569 kernel: ena 0000:00:05.0: LLQ is not supported Fallback to host mode policy. Apr 30 03:33:44.811728 kernel: ena 0000:00:05.0: Elastic Network Adapter (ENA) found at mem 80400000, mac addr 06:28:24:02:ac:29 Apr 30 03:33:44.811881 kernel: AVX2 version of gcm_enc/dec engaged. Apr 30 03:33:44.811905 kernel: AES CTR mode by8 optimization enabled Apr 30 03:33:44.779390 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 03:33:44.779570 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 03:33:44.782260 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 03:33:44.782903 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 03:33:44.783115 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 03:33:44.783810 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 03:33:44.794864 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 03:33:44.815587 (udev-worker)[459]: Network interface NamePolicy= disabled on kernel command line. Apr 30 03:33:44.817483 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 03:33:44.842078 kernel: nvme nvme0: pci function 0000:00:04.0 Apr 30 03:33:44.842337 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Apr 30 03:33:44.817636 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 03:33:44.830142 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 03:33:44.855426 kernel: nvme nvme0: 2/0/0 default/read/poll queues Apr 30 03:33:44.863609 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 30 03:33:44.863681 kernel: GPT:9289727 != 16777215 Apr 30 03:33:44.863702 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 30 03:33:44.863721 kernel: GPT:9289727 != 16777215 Apr 30 03:33:44.863739 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 30 03:33:44.863759 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 03:33:44.871094 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 03:33:44.880611 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 03:33:44.910793 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 03:33:44.938303 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/nvme0n1p6 scanned by (udev-worker) (447) Apr 30 03:33:44.962425 kernel: BTRFS: device fsid 24af5149-14c0-4f50-b6d3-2f5c9259df26 devid 1 transid 38 /dev/nvme0n1p3 scanned by (udev-worker) (462) Apr 30 03:33:45.023119 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Apr 30 03:33:45.029803 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Amazon Elastic Block Store ROOT. Apr 30 03:33:45.045117 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Amazon Elastic Block Store USR-A. Apr 30 03:33:45.045660 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Amazon Elastic Block Store USR-A. Apr 30 03:33:45.053108 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Amazon Elastic Block Store EFI-SYSTEM. Apr 30 03:33:45.065692 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 30 03:33:45.072732 disk-uuid[634]: Primary Header is updated. Apr 30 03:33:45.072732 disk-uuid[634]: Secondary Entries is updated. Apr 30 03:33:45.072732 disk-uuid[634]: Secondary Header is updated. Apr 30 03:33:45.078423 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 03:33:45.084168 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 03:33:46.093827 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 03:33:46.093902 disk-uuid[635]: The operation has completed successfully. Apr 30 03:33:46.227485 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 30 03:33:46.227603 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 30 03:33:46.249598 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 30 03:33:46.254324 sh[978]: Success Apr 30 03:33:46.275616 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Apr 30 03:33:46.365386 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 30 03:33:46.382470 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 30 03:33:46.383464 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 30 03:33:46.413470 kernel: BTRFS info (device dm-0): first mount of filesystem 24af5149-14c0-4f50-b6d3-2f5c9259df26 Apr 30 03:33:46.413536 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Apr 30 03:33:46.413550 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 30 03:33:46.417215 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 30 03:33:46.417294 kernel: BTRFS info (device dm-0): using free space tree Apr 30 03:33:46.503444 kernel: BTRFS info (device dm-0): enabling ssd optimizations Apr 30 03:33:46.516425 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 30 03:33:46.517451 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 30 03:33:46.521559 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 30 03:33:46.523551 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 30 03:33:46.547778 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem dea0d870-fd31-489b-84db-7261ba2c88d5 Apr 30 03:33:46.547839 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Apr 30 03:33:46.547853 kernel: BTRFS info (device nvme0n1p6): using free space tree Apr 30 03:33:46.555462 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Apr 30 03:33:46.567287 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 30 03:33:46.567802 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem dea0d870-fd31-489b-84db-7261ba2c88d5 Apr 30 03:33:46.574645 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 30 03:33:46.578732 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 30 03:33:46.616541 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 03:33:46.621594 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 30 03:33:46.651594 systemd-networkd[1170]: lo: Link UP Apr 30 03:33:46.651605 systemd-networkd[1170]: lo: Gained carrier Apr 30 03:33:46.653508 systemd-networkd[1170]: Enumeration completed Apr 30 03:33:46.653958 systemd-networkd[1170]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 03:33:46.653963 systemd-networkd[1170]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 03:33:46.655239 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 30 03:33:46.656898 systemd[1]: Reached target network.target - Network. Apr 30 03:33:46.656948 systemd-networkd[1170]: eth0: Link UP Apr 30 03:33:46.656953 systemd-networkd[1170]: eth0: Gained carrier Apr 30 03:33:46.656965 systemd-networkd[1170]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 03:33:46.677512 systemd-networkd[1170]: eth0: DHCPv4 address 172.31.20.13/20, gateway 172.31.16.1 acquired from 172.31.16.1 Apr 30 03:33:46.931049 ignition[1121]: Ignition 2.19.0 Apr 30 03:33:46.931063 ignition[1121]: Stage: fetch-offline Apr 30 03:33:46.931326 ignition[1121]: no configs at "/usr/lib/ignition/base.d" Apr 30 03:33:46.931341 ignition[1121]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Apr 30 03:33:46.933200 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 03:33:46.931709 ignition[1121]: Ignition finished successfully Apr 30 03:33:46.938834 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Apr 30 03:33:46.953905 ignition[1178]: Ignition 2.19.0 Apr 30 03:33:46.953918 ignition[1178]: Stage: fetch Apr 30 03:33:46.954366 ignition[1178]: no configs at "/usr/lib/ignition/base.d" Apr 30 03:33:46.954381 ignition[1178]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Apr 30 03:33:46.954651 ignition[1178]: PUT http://169.254.169.254/latest/api/token: attempt #1 Apr 30 03:33:46.996390 ignition[1178]: PUT result: OK Apr 30 03:33:47.015823 ignition[1178]: parsed url from cmdline: "" Apr 30 03:33:47.015835 ignition[1178]: no config URL provided Apr 30 03:33:47.015846 ignition[1178]: reading system config file "/usr/lib/ignition/user.ign" Apr 30 03:33:47.015863 ignition[1178]: no config at "/usr/lib/ignition/user.ign" Apr 30 03:33:47.015892 ignition[1178]: PUT http://169.254.169.254/latest/api/token: attempt #1 Apr 30 03:33:47.016891 ignition[1178]: PUT result: OK Apr 30 03:33:47.016948 ignition[1178]: GET http://169.254.169.254/2019-10-01/user-data: attempt #1 Apr 30 03:33:47.022521 ignition[1178]: GET result: OK Apr 30 03:33:47.022715 ignition[1178]: parsing config with SHA512: 041878bb1c12c4b6b7766bcf22f60b6c21bfd190fc4b119f57a0e4d52c4e0bed9ade1cc6026ab198d4234abf1c59b482d61aedfa664954c4e63f927a08ae630c Apr 30 03:33:47.027343 unknown[1178]: fetched base config from "system" Apr 30 03:33:47.027364 unknown[1178]: fetched base config from "system" Apr 30 03:33:47.028593 ignition[1178]: fetch: fetch complete Apr 30 03:33:47.027373 unknown[1178]: fetched user config from "aws" Apr 30 03:33:47.028602 ignition[1178]: fetch: fetch passed Apr 30 03:33:47.030832 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Apr 30 03:33:47.028660 ignition[1178]: Ignition finished successfully Apr 30 03:33:47.036617 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 30 03:33:47.052311 ignition[1184]: Ignition 2.19.0 Apr 30 03:33:47.052378 ignition[1184]: Stage: kargs Apr 30 03:33:47.052861 ignition[1184]: no configs at "/usr/lib/ignition/base.d" Apr 30 03:33:47.052875 ignition[1184]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Apr 30 03:33:47.052998 ignition[1184]: PUT http://169.254.169.254/latest/api/token: attempt #1 Apr 30 03:33:47.054246 ignition[1184]: PUT result: OK Apr 30 03:33:47.057451 ignition[1184]: kargs: kargs passed Apr 30 03:33:47.057510 ignition[1184]: Ignition finished successfully Apr 30 03:33:47.059093 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 30 03:33:47.063640 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 30 03:33:47.078333 ignition[1191]: Ignition 2.19.0 Apr 30 03:33:47.078345 ignition[1191]: Stage: disks Apr 30 03:33:47.078795 ignition[1191]: no configs at "/usr/lib/ignition/base.d" Apr 30 03:33:47.078804 ignition[1191]: no config dir at "/usr/lib/ignition/base.platform.d/aws" Apr 30 03:33:47.078890 ignition[1191]: PUT http://169.254.169.254/latest/api/token: attempt #1 Apr 30 03:33:47.080936 ignition[1191]: PUT result: OK Apr 30 03:33:47.084707 ignition[1191]: disks: disks passed Apr 30 03:33:47.084766 ignition[1191]: Ignition finished successfully Apr 30 03:33:47.085693 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 30 03:33:47.086668 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 30 03:33:47.086991 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 30 03:33:47.087471 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 30 03:33:47.087806 systemd[1]: Reached target sysinit.target - System Initialization. Apr 30 03:33:47.088292 systemd[1]: Reached target basic.target - Basic System. Apr 30 03:33:47.100669 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 30 03:33:47.132454 systemd-fsck[1199]: ROOT: clean, 14/553520 files, 52654/553472 blocks Apr 30 03:33:47.135785 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 30 03:33:47.141578 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 30 03:33:47.241432 kernel: EXT4-fs (nvme0n1p9): mounted filesystem c246962b-d3a7-4703-a2cb-a633fbca1b76 r/w with ordered data mode. Quota mode: none. Apr 30 03:33:47.242684 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 30 03:33:47.243824 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 30 03:33:47.250591 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 03:33:47.253838 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 30 03:33:47.255807 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Apr 30 03:33:47.256518 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 30 03:33:47.256550 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 03:33:47.263084 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 30 03:33:47.269609 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 30 03:33:47.274423 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/nvme0n1p6 scanned by mount (1218) Apr 30 03:33:47.278238 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem dea0d870-fd31-489b-84db-7261ba2c88d5 Apr 30 03:33:47.278301 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Apr 30 03:33:47.278322 kernel: BTRFS info (device nvme0n1p6): using free space tree Apr 30 03:33:47.286431 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Apr 30 03:33:47.288736 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 03:33:47.495309 initrd-setup-root[1243]: cut: /sysroot/etc/passwd: No such file or directory Apr 30 03:33:47.500090 initrd-setup-root[1250]: cut: /sysroot/etc/group: No such file or directory Apr 30 03:33:47.504511 initrd-setup-root[1257]: cut: /sysroot/etc/shadow: No such file or directory Apr 30 03:33:47.508786 initrd-setup-root[1264]: cut: /sysroot/etc/gshadow: No such file or directory Apr 30 03:33:47.693545 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 30 03:33:47.699547 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 30 03:33:47.701599 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 30 03:33:47.708410 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 30 03:33:47.710437 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem dea0d870-fd31-489b-84db-7261ba2c88d5 Apr 30 03:33:47.738662 ignition[1336]: INFO : Ignition 2.19.0 Apr 30 03:33:47.740483 ignition[1336]: INFO : Stage: mount Apr 30 03:33:47.740483 ignition[1336]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 03:33:47.740483 ignition[1336]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Apr 30 03:33:47.743446 ignition[1336]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Apr 30 03:33:47.743446 ignition[1336]: INFO : PUT result: OK Apr 30 03:33:47.747084 ignition[1336]: INFO : mount: mount passed Apr 30 03:33:47.748871 ignition[1336]: INFO : Ignition finished successfully Apr 30 03:33:47.750263 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 30 03:33:47.754642 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 30 03:33:47.766811 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 30 03:33:47.782676 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 03:33:47.802426 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme0n1p6 scanned by mount (1348) Apr 30 03:33:47.802609 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem dea0d870-fd31-489b-84db-7261ba2c88d5 Apr 30 03:33:47.805627 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-intel) checksum algorithm Apr 30 03:33:47.805683 kernel: BTRFS info (device nvme0n1p6): using free space tree Apr 30 03:33:47.812433 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Apr 30 03:33:47.814281 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 03:33:47.850412 ignition[1365]: INFO : Ignition 2.19.0 Apr 30 03:33:47.850412 ignition[1365]: INFO : Stage: files Apr 30 03:33:47.851997 ignition[1365]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 03:33:47.851997 ignition[1365]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Apr 30 03:33:47.851997 ignition[1365]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Apr 30 03:33:47.853303 ignition[1365]: INFO : PUT result: OK Apr 30 03:33:47.855217 ignition[1365]: DEBUG : files: compiled without relabeling support, skipping Apr 30 03:33:47.869651 ignition[1365]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 30 03:33:47.869651 ignition[1365]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 30 03:33:47.898246 ignition[1365]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 30 03:33:47.899174 ignition[1365]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 30 03:33:47.899174 ignition[1365]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 30 03:33:47.898796 unknown[1365]: wrote ssh authorized keys file for user: core Apr 30 03:33:47.901105 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Apr 30 03:33:47.901859 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Apr 30 03:33:47.976095 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 30 03:33:48.047542 systemd-networkd[1170]: eth0: Gained IPv6LL Apr 30 03:33:48.165421 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Apr 30 03:33:48.165421 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Apr 30 03:33:48.166982 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 Apr 30 03:33:48.603316 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 30 03:33:48.933281 ignition[1365]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Apr 30 03:33:48.933281 ignition[1365]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 30 03:33:48.935370 ignition[1365]: INFO : files: files passed Apr 30 03:33:48.935370 ignition[1365]: INFO : Ignition finished successfully Apr 30 03:33:48.936588 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 30 03:33:48.945666 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 30 03:33:48.948967 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 30 03:33:48.952173 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 30 03:33:48.952308 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 30 03:33:48.976122 initrd-setup-root-after-ignition[1393]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 03:33:48.976122 initrd-setup-root-after-ignition[1393]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 30 03:33:48.980180 initrd-setup-root-after-ignition[1397]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 03:33:48.980635 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 03:33:48.982304 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 30 03:33:48.986734 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 30 03:33:49.013258 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 30 03:33:49.013419 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 30 03:33:49.015045 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 30 03:33:49.015916 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 30 03:33:49.016792 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 30 03:33:49.028652 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 30 03:33:49.041705 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 03:33:49.047627 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 30 03:33:49.059539 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 30 03:33:49.060834 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 03:33:49.061635 systemd[1]: Stopped target timers.target - Timer Units. Apr 30 03:33:49.062384 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 30 03:33:49.062758 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 03:33:49.063802 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 30 03:33:49.064648 systemd[1]: Stopped target basic.target - Basic System. Apr 30 03:33:49.065397 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 30 03:33:49.066162 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 03:33:49.066999 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 30 03:33:49.067759 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 30 03:33:49.068517 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 03:33:49.069274 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 30 03:33:49.070384 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 30 03:33:49.071232 systemd[1]: Stopped target swap.target - Swaps. Apr 30 03:33:49.071934 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 30 03:33:49.072112 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 30 03:33:49.073188 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 30 03:33:49.073968 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 03:33:49.074752 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 30 03:33:49.074899 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 03:33:49.075540 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 30 03:33:49.075712 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 30 03:33:49.077040 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 30 03:33:49.077223 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 03:33:49.077929 systemd[1]: ignition-files.service: Deactivated successfully. Apr 30 03:33:49.078079 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 30 03:33:49.084764 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 30 03:33:49.089729 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 30 03:33:49.091166 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 30 03:33:49.091394 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 03:33:49.093427 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 30 03:33:49.093610 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 03:33:49.102287 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 30 03:33:49.102431 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 30 03:33:49.108547 ignition[1417]: INFO : Ignition 2.19.0 Apr 30 03:33:49.108547 ignition[1417]: INFO : Stage: umount Apr 30 03:33:49.111503 ignition[1417]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 03:33:49.111503 ignition[1417]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/aws" Apr 30 03:33:49.111503 ignition[1417]: INFO : PUT http://169.254.169.254/latest/api/token: attempt #1 Apr 30 03:33:49.111503 ignition[1417]: INFO : PUT result: OK Apr 30 03:33:49.113925 ignition[1417]: INFO : umount: umount passed Apr 30 03:33:49.113925 ignition[1417]: INFO : Ignition finished successfully Apr 30 03:33:49.115898 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 30 03:33:49.116042 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 30 03:33:49.117175 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 30 03:33:49.117299 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 30 03:33:49.118666 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 30 03:33:49.118738 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 30 03:33:49.119275 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 30 03:33:49.119352 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Apr 30 03:33:49.119860 systemd[1]: Stopped target network.target - Network. Apr 30 03:33:49.120501 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 30 03:33:49.120562 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 03:33:49.120992 systemd[1]: Stopped target paths.target - Path Units. Apr 30 03:33:49.123825 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 30 03:33:49.127476 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 03:33:49.127930 systemd[1]: Stopped target slices.target - Slice Units. Apr 30 03:33:49.128391 systemd[1]: Stopped target sockets.target - Socket Units. Apr 30 03:33:49.130732 systemd[1]: iscsid.socket: Deactivated successfully. Apr 30 03:33:49.130794 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 03:33:49.131248 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 30 03:33:49.131293 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 03:33:49.131731 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 30 03:33:49.131793 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 30 03:33:49.132240 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 30 03:33:49.132337 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 30 03:33:49.133204 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 30 03:33:49.133845 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 30 03:33:49.136062 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 30 03:33:49.137634 systemd-networkd[1170]: eth0: DHCPv6 lease lost Apr 30 03:33:49.139879 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 30 03:33:49.141669 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 30 03:33:49.143039 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 30 03:33:49.143177 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 30 03:33:49.145655 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 30 03:33:49.146152 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 30 03:33:49.147690 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 30 03:33:49.147750 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 30 03:33:49.148511 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 30 03:33:49.148572 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 30 03:33:49.156537 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 30 03:33:49.157735 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 30 03:33:49.157822 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 03:33:49.158359 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 30 03:33:49.158434 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 30 03:33:49.158921 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 30 03:33:49.158976 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 30 03:33:49.159594 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 30 03:33:49.159649 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 03:33:49.160747 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 03:33:49.174822 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 30 03:33:49.174939 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 30 03:33:49.176076 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 30 03:33:49.176220 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 03:33:49.177734 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 30 03:33:49.177803 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 30 03:33:49.178614 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 30 03:33:49.178652 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 03:33:49.179349 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 30 03:33:49.179443 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 30 03:33:49.180688 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 30 03:33:49.180754 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 30 03:33:49.181914 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 03:33:49.181973 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 03:33:49.188741 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 30 03:33:49.189431 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 30 03:33:49.189522 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 03:33:49.190176 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Apr 30 03:33:49.190239 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 30 03:33:49.190896 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 30 03:33:49.190956 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 03:33:49.191564 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 03:33:49.191617 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 03:33:49.198597 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 30 03:33:49.198731 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 30 03:33:49.200058 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 30 03:33:49.205642 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 30 03:33:49.215614 systemd[1]: Switching root. Apr 30 03:33:49.240959 systemd-journald[178]: Journal stopped Apr 30 03:33:50.970351 systemd-journald[178]: Received SIGTERM from PID 1 (systemd). Apr 30 03:33:50.984638 kernel: SELinux: policy capability network_peer_controls=1 Apr 30 03:33:50.984673 kernel: SELinux: policy capability open_perms=1 Apr 30 03:33:50.984700 kernel: SELinux: policy capability extended_socket_class=1 Apr 30 03:33:50.984720 kernel: SELinux: policy capability always_check_network=0 Apr 30 03:33:50.984739 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 30 03:33:50.984767 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 30 03:33:50.984787 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 30 03:33:50.984807 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 30 03:33:50.984833 kernel: audit: type=1403 audit(1745984029.658:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 30 03:33:50.984860 systemd[1]: Successfully loaded SELinux policy in 51.594ms. Apr 30 03:33:50.984895 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.647ms. Apr 30 03:33:50.984918 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Apr 30 03:33:50.984939 systemd[1]: Detected virtualization amazon. Apr 30 03:33:50.984961 systemd[1]: Detected architecture x86-64. Apr 30 03:33:50.984982 systemd[1]: Detected first boot. Apr 30 03:33:50.985004 systemd[1]: Initializing machine ID from VM UUID. Apr 30 03:33:50.985032 zram_generator::config[1460]: No configuration found. Apr 30 03:33:50.985060 systemd[1]: Populated /etc with preset unit settings. Apr 30 03:33:50.985081 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 30 03:33:50.985105 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 30 03:33:50.985128 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 30 03:33:50.985151 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 30 03:33:50.985173 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 30 03:33:50.985195 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 30 03:33:50.985216 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 30 03:33:50.985243 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 30 03:33:50.985264 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 30 03:33:50.985290 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 30 03:33:50.985310 systemd[1]: Created slice user.slice - User and Session Slice. Apr 30 03:33:50.985332 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 03:33:50.985353 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 03:33:50.985375 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 30 03:33:50.985396 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 30 03:33:50.985437 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 30 03:33:50.985456 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 30 03:33:50.985475 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Apr 30 03:33:50.985499 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 03:33:50.985520 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 30 03:33:50.985541 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 30 03:33:50.985562 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 30 03:33:50.985584 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 30 03:33:50.985605 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 03:33:50.985626 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 30 03:33:50.985647 systemd[1]: Reached target slices.target - Slice Units. Apr 30 03:33:50.985672 systemd[1]: Reached target swap.target - Swaps. Apr 30 03:33:50.985693 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 30 03:33:50.985713 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 30 03:33:50.985735 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 30 03:33:50.985756 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 30 03:33:50.985778 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 03:33:50.985799 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 30 03:33:50.985820 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 30 03:33:50.985841 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 30 03:33:50.985866 systemd[1]: Mounting media.mount - External Media Directory... Apr 30 03:33:50.985888 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:50.985909 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 30 03:33:50.985931 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 30 03:33:50.985953 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 30 03:33:50.985972 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 30 03:33:50.985992 systemd[1]: Reached target machines.target - Containers. Apr 30 03:33:50.986013 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 30 03:33:50.986038 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 03:33:50.986060 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 30 03:33:50.986083 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 30 03:33:50.986104 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 03:33:50.986126 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 30 03:33:50.986147 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 30 03:33:50.986168 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 30 03:33:50.986188 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 30 03:33:50.986210 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 30 03:33:50.986234 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 30 03:33:50.986255 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 30 03:33:50.986276 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 30 03:33:50.986297 systemd[1]: Stopped systemd-fsck-usr.service. Apr 30 03:33:50.986318 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 30 03:33:50.986339 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 30 03:33:50.986361 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 30 03:33:50.986382 kernel: fuse: init (API version 7.39) Apr 30 03:33:50.995108 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 30 03:33:50.995164 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 30 03:33:50.995189 systemd[1]: verity-setup.service: Deactivated successfully. Apr 30 03:33:50.995212 systemd[1]: Stopped verity-setup.service. Apr 30 03:33:50.995236 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:50.995257 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 30 03:33:50.995280 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 30 03:33:50.995302 systemd[1]: Mounted media.mount - External Media Directory. Apr 30 03:33:50.995324 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 30 03:33:50.995349 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 30 03:33:50.995371 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 30 03:33:50.995393 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 03:33:50.995424 kernel: loop: module loaded Apr 30 03:33:50.995444 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 30 03:33:50.995469 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 30 03:33:50.995491 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 03:33:50.995513 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 03:33:50.995535 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 30 03:33:50.995561 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 30 03:33:50.995583 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 30 03:33:50.995609 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 30 03:33:50.995631 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 30 03:33:50.995653 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 30 03:33:50.995674 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 30 03:33:50.995696 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 30 03:33:50.995717 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 30 03:33:50.995739 kernel: ACPI: bus type drm_connector registered Apr 30 03:33:50.995794 systemd-journald[1538]: Collecting audit messages is disabled. Apr 30 03:33:50.995840 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 30 03:33:50.995862 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 30 03:33:50.995884 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 30 03:33:50.995906 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 30 03:33:50.995927 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Apr 30 03:33:50.995948 systemd-journald[1538]: Journal started Apr 30 03:33:50.995993 systemd-journald[1538]: Runtime Journal (/run/log/journal/ec2604ebf32ac5bac3fe163d0660277f) is 4.7M, max 38.2M, 33.4M free. Apr 30 03:33:50.560142 systemd[1]: Queued start job for default target multi-user.target. Apr 30 03:33:50.603890 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Apr 30 03:33:50.604375 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 30 03:33:51.013573 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 30 03:33:51.025358 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 30 03:33:51.025482 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 03:33:51.035427 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 30 03:33:51.040562 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 30 03:33:51.053430 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 30 03:33:51.058464 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 30 03:33:51.069046 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 30 03:33:51.088442 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 30 03:33:51.106323 systemd[1]: Started systemd-journald.service - Journal Service. Apr 30 03:33:51.119691 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 30 03:33:51.122718 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 30 03:33:51.122903 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 30 03:33:51.124012 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 30 03:33:51.125910 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 03:33:51.127723 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 30 03:33:51.129673 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 30 03:33:51.131369 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 30 03:33:51.133105 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 30 03:33:51.162250 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 30 03:33:51.169638 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 30 03:33:51.174000 systemd-tmpfiles[1571]: ACLs are not supported, ignoring. Apr 30 03:33:51.174033 systemd-tmpfiles[1571]: ACLs are not supported, ignoring. Apr 30 03:33:51.178656 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Apr 30 03:33:51.180906 kernel: loop0: detected capacity change from 0 to 142488 Apr 30 03:33:51.183611 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 30 03:33:51.187091 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 30 03:33:51.208574 systemd-journald[1538]: Time spent on flushing to /var/log/journal/ec2604ebf32ac5bac3fe163d0660277f is 96.049ms for 993 entries. Apr 30 03:33:51.208574 systemd-journald[1538]: System Journal (/var/log/journal/ec2604ebf32ac5bac3fe163d0660277f) is 8.0M, max 195.6M, 187.6M free. Apr 30 03:33:51.313393 systemd-journald[1538]: Received client request to flush runtime journal. Apr 30 03:33:51.313985 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 30 03:33:51.207172 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 30 03:33:51.234794 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 30 03:33:51.247834 udevadm[1599]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Apr 30 03:33:51.317969 kernel: loop1: detected capacity change from 0 to 218376 Apr 30 03:33:51.306125 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 30 03:33:51.318119 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 30 03:33:51.331127 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 30 03:33:51.339602 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 30 03:33:51.341539 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 30 03:33:51.342459 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Apr 30 03:33:51.372119 systemd-tmpfiles[1611]: ACLs are not supported, ignoring. Apr 30 03:33:51.372558 systemd-tmpfiles[1611]: ACLs are not supported, ignoring. Apr 30 03:33:51.378937 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 03:33:51.473832 kernel: loop2: detected capacity change from 0 to 140768 Apr 30 03:33:51.568424 kernel: loop3: detected capacity change from 0 to 61336 Apr 30 03:33:51.622522 kernel: loop4: detected capacity change from 0 to 142488 Apr 30 03:33:51.652812 kernel: loop5: detected capacity change from 0 to 218376 Apr 30 03:33:51.689908 kernel: loop6: detected capacity change from 0 to 140768 Apr 30 03:33:51.722574 kernel: loop7: detected capacity change from 0 to 61336 Apr 30 03:33:51.749653 (sd-merge)[1618]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-ami'. Apr 30 03:33:51.751156 (sd-merge)[1618]: Merged extensions into '/usr'. Apr 30 03:33:51.759919 systemd[1]: Reloading requested from client PID 1570 ('systemd-sysext') (unit systemd-sysext.service)... Apr 30 03:33:51.760092 systemd[1]: Reloading... Apr 30 03:33:51.890864 zram_generator::config[1645]: No configuration found. Apr 30 03:33:52.097465 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 03:33:52.199350 systemd[1]: Reloading finished in 438 ms. Apr 30 03:33:52.230413 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 30 03:33:52.242693 systemd[1]: Starting ensure-sysext.service... Apr 30 03:33:52.246029 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 30 03:33:52.263014 ldconfig[1567]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 30 03:33:52.270272 systemd[1]: Reloading requested from client PID 1696 ('systemctl') (unit ensure-sysext.service)... Apr 30 03:33:52.270393 systemd[1]: Reloading... Apr 30 03:33:52.293002 systemd-tmpfiles[1697]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 30 03:33:52.293924 systemd-tmpfiles[1697]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 30 03:33:52.294931 systemd-tmpfiles[1697]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 30 03:33:52.295249 systemd-tmpfiles[1697]: ACLs are not supported, ignoring. Apr 30 03:33:52.295344 systemd-tmpfiles[1697]: ACLs are not supported, ignoring. Apr 30 03:33:52.299624 systemd-tmpfiles[1697]: Detected autofs mount point /boot during canonicalization of boot. Apr 30 03:33:52.299638 systemd-tmpfiles[1697]: Skipping /boot Apr 30 03:33:52.309450 systemd-tmpfiles[1697]: Detected autofs mount point /boot during canonicalization of boot. Apr 30 03:33:52.309462 systemd-tmpfiles[1697]: Skipping /boot Apr 30 03:33:52.351123 zram_generator::config[1724]: No configuration found. Apr 30 03:33:52.473230 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 03:33:52.528580 systemd[1]: Reloading finished in 257 ms. Apr 30 03:33:52.545638 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 30 03:33:52.546513 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 30 03:33:52.552873 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 03:33:52.565629 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 30 03:33:52.569624 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 30 03:33:52.574684 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 30 03:33:52.582696 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 30 03:33:52.591608 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 03:33:52.595697 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 30 03:33:52.621766 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 30 03:33:52.625762 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:52.626951 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 03:33:52.633899 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 03:33:52.645959 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 30 03:33:52.652786 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 30 03:33:52.653533 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 03:33:52.653713 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:52.662865 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:52.664909 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 03:33:52.665242 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 03:33:52.665494 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:52.673985 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:52.675007 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 03:33:52.682762 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 30 03:33:52.683566 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 03:33:52.683851 systemd[1]: Reached target time-set.target - System Time Set. Apr 30 03:33:52.685375 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Apr 30 03:33:52.687618 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 30 03:33:52.689348 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 30 03:33:52.694124 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 30 03:33:52.694355 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 30 03:33:52.700747 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 30 03:33:52.700982 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 30 03:33:52.702247 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 03:33:52.703759 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 03:33:52.714532 systemd[1]: Finished ensure-sysext.service. Apr 30 03:33:52.722651 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 30 03:33:52.724592 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 30 03:33:52.731684 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 30 03:33:52.743031 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 30 03:33:52.743240 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 30 03:33:52.746541 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 30 03:33:52.746805 systemd-udevd[1785]: Using default interface naming scheme 'v255'. Apr 30 03:33:52.766666 augenrules[1817]: No rules Apr 30 03:33:52.764755 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 30 03:33:52.784509 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 30 03:33:52.787761 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 30 03:33:52.789669 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 30 03:33:52.807678 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 03:33:52.818721 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 30 03:33:52.869017 systemd-resolved[1784]: Positive Trust Anchors: Apr 30 03:33:52.869465 systemd-resolved[1784]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 30 03:33:52.869532 systemd-resolved[1784]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 30 03:33:52.883508 systemd-resolved[1784]: Defaulting to hostname 'linux'. Apr 30 03:33:52.887942 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 30 03:33:52.888680 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 30 03:33:52.921185 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Apr 30 03:33:52.934520 (udev-worker)[1828]: Network interface NamePolicy= disabled on kernel command line. Apr 30 03:33:52.948092 systemd-networkd[1829]: lo: Link UP Apr 30 03:33:52.948585 systemd-networkd[1829]: lo: Gained carrier Apr 30 03:33:52.952504 systemd-networkd[1829]: Enumeration completed Apr 30 03:33:52.952651 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 30 03:33:52.953003 systemd-networkd[1829]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 03:33:52.953009 systemd-networkd[1829]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 03:33:52.955609 systemd[1]: Reached target network.target - Network. Apr 30 03:33:52.962234 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 30 03:33:52.967558 systemd-networkd[1829]: eth0: Link UP Apr 30 03:33:52.967802 systemd-networkd[1829]: eth0: Gained carrier Apr 30 03:33:52.967836 systemd-networkd[1829]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 03:33:52.977195 systemd-networkd[1829]: eth0: DHCPv4 address 172.31.20.13/20, gateway 172.31.16.1 acquired from 172.31.16.1 Apr 30 03:33:53.035711 systemd-networkd[1829]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Apr 30 03:33:53.045431 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input4 Apr 30 03:33:53.051469 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input3 Apr 30 03:33:53.053428 kernel: ACPI: button: Power Button [PWRF] Apr 30 03:33:53.062493 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Apr 30 03:33:53.062879 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input5 Apr 30 03:33:53.071442 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (1827) Apr 30 03:33:53.071531 kernel: ACPI: button: Sleep Button [SLPF] Apr 30 03:33:53.124876 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 03:33:53.164544 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 03:33:53.164805 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 03:33:53.172496 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 03:33:53.232426 kernel: mousedev: PS/2 mouse device common for all mice Apr 30 03:33:53.293564 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 03:33:53.300968 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Amazon Elastic Block Store OEM. Apr 30 03:33:53.301801 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 30 03:33:53.311679 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 30 03:33:53.315556 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 30 03:33:53.330456 lvm[1948]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 30 03:33:53.339891 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 30 03:33:53.357725 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 30 03:33:53.358998 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 30 03:33:53.359488 systemd[1]: Reached target sysinit.target - System Initialization. Apr 30 03:33:53.359967 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 30 03:33:53.360369 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 30 03:33:53.360881 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 30 03:33:53.361328 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 30 03:33:53.361675 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 30 03:33:53.361992 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 30 03:33:53.362023 systemd[1]: Reached target paths.target - Path Units. Apr 30 03:33:53.362325 systemd[1]: Reached target timers.target - Timer Units. Apr 30 03:33:53.363921 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 30 03:33:53.365650 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 30 03:33:53.371504 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 30 03:33:53.374244 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 30 03:33:53.375423 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 30 03:33:53.375931 systemd[1]: Reached target sockets.target - Socket Units. Apr 30 03:33:53.376281 systemd[1]: Reached target basic.target - Basic System. Apr 30 03:33:53.376940 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 30 03:33:53.377053 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 30 03:33:53.379259 systemd[1]: Starting containerd.service - containerd container runtime... Apr 30 03:33:53.382925 lvm[1955]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 30 03:33:53.391002 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 30 03:33:53.394602 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 30 03:33:53.404620 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 30 03:33:53.415754 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 30 03:33:53.416165 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 30 03:33:53.424604 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 30 03:33:53.427612 jq[1959]: false Apr 30 03:33:53.428115 systemd[1]: Started ntpd.service - Network Time Service. Apr 30 03:33:53.433303 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 30 03:33:53.438544 systemd[1]: Starting setup-oem.service - Setup OEM... Apr 30 03:33:53.445650 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 30 03:33:53.447766 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 30 03:33:53.457580 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 30 03:33:53.458392 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 30 03:33:53.458885 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 30 03:33:53.463074 systemd[1]: Starting update-engine.service - Update Engine... Apr 30 03:33:53.471385 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 30 03:33:53.474466 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 30 03:33:53.478284 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 30 03:33:53.478496 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 30 03:33:53.484009 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 30 03:33:53.484493 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 30 03:33:53.507996 jq[1974]: true Apr 30 03:33:53.532670 update_engine[1972]: I20250430 03:33:53.526083 1972 main.cc:92] Flatcar Update Engine starting Apr 30 03:33:53.534004 coreos-metadata[1957]: Apr 30 03:33:53.533 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Apr 30 03:33:53.539161 coreos-metadata[1957]: Apr 30 03:33:53.539 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-id: Attempt #1 Apr 30 03:33:53.544212 systemd-logind[1970]: Watching system buttons on /dev/input/event1 (Power Button) Apr 30 03:33:53.544541 systemd-logind[1970]: Watching system buttons on /dev/input/event3 (Sleep Button) Apr 30 03:33:53.544639 systemd-logind[1970]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 30 03:33:53.544993 systemd-logind[1970]: New seat seat0. Apr 30 03:33:53.545110 coreos-metadata[1957]: Apr 30 03:33:53.545 INFO Fetch successful Apr 30 03:33:53.545179 coreos-metadata[1957]: Apr 30 03:33:53.545 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/instance-type: Attempt #1 Apr 30 03:33:53.546944 systemd[1]: Started systemd-logind.service - User Login Management. Apr 30 03:33:53.547536 coreos-metadata[1957]: Apr 30 03:33:53.547 INFO Fetch successful Apr 30 03:33:53.547605 coreos-metadata[1957]: Apr 30 03:33:53.547 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/local-ipv4: Attempt #1 Apr 30 03:33:53.553120 coreos-metadata[1957]: Apr 30 03:33:53.551 INFO Fetch successful Apr 30 03:33:53.553120 coreos-metadata[1957]: Apr 30 03:33:53.551 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-ipv4: Attempt #1 Apr 30 03:33:53.552877 dbus-daemon[1958]: [system] SELinux support is enabled Apr 30 03:33:53.555899 extend-filesystems[1960]: Found loop4 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found loop5 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found loop6 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found loop7 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found nvme0n1 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found nvme0n1p1 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found nvme0n1p2 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found nvme0n1p3 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found usr Apr 30 03:33:53.555899 extend-filesystems[1960]: Found nvme0n1p4 Apr 30 03:33:53.555899 extend-filesystems[1960]: Found nvme0n1p6 Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.553 INFO Fetch successful Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.553 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/ipv6: Attempt #1 Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.565 INFO Fetch failed with 404: resource not found Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.565 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone: Attempt #1 Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.568 INFO Fetch successful Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.568 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/placement/availability-zone-id: Attempt #1 Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.571 INFO Fetch successful Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.571 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/hostname: Attempt #1 Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.572 INFO Fetch successful Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.572 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-hostname: Attempt #1 Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.573 INFO Fetch successful Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.573 INFO Fetching http://169.254.169.254/2021-01-03/dynamic/instance-identity/document: Attempt #1 Apr 30 03:33:53.577497 coreos-metadata[1957]: Apr 30 03:33:53.575 INFO Fetch successful Apr 30 03:33:53.553646 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 30 03:33:53.577882 update_engine[1972]: I20250430 03:33:53.563066 1972 update_check_scheduler.cc:74] Next update check in 7m53s Apr 30 03:33:53.577909 jq[1991]: true Apr 30 03:33:53.577981 extend-filesystems[1960]: Found nvme0n1p7 Apr 30 03:33:53.577981 extend-filesystems[1960]: Found nvme0n1p9 Apr 30 03:33:53.577981 extend-filesystems[1960]: Checking size of /dev/nvme0n1p9 Apr 30 03:33:53.562113 dbus-daemon[1958]: [system] Successfully activated service 'org.freedesktop.systemd1' Apr 30 03:33:53.557472 systemd[1]: motdgen.service: Deactivated successfully. Apr 30 03:33:53.562727 dbus-daemon[1958]: [system] Activating via systemd: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.2' (uid=244 pid=1829 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Apr 30 03:33:53.557639 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 30 03:33:53.560145 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 30 03:33:53.560192 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 30 03:33:53.562185 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 30 03:33:53.562206 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 30 03:33:53.564167 (ntainerd)[1992]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 30 03:33:53.575857 systemd[1]: Started update-engine.service - Update Engine. Apr 30 03:33:53.592087 systemd[1]: Starting systemd-hostnamed.service - Hostname Service... Apr 30 03:33:53.595923 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 30 03:33:53.598861 systemd[1]: Finished setup-oem.service - Setup OEM. Apr 30 03:33:53.614914 tar[1977]: linux-amd64/LICENSE Apr 30 03:33:53.617835 ntpd[1964]: ntpd 4.2.8p17@1.4004-o Tue Apr 29 22:12:23 UTC 2025 (1): Starting Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: ntpd 4.2.8p17@1.4004-o Tue Apr 29 22:12:23 UTC 2025 (1): Starting Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: ---------------------------------------------------- Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: ntp-4 is maintained by Network Time Foundation, Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: corporation. Support and training for ntp-4 are Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: available at https://www.nwtime.org/support Apr 30 03:33:53.620578 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: ---------------------------------------------------- Apr 30 03:33:53.617859 ntpd[1964]: Command line: /usr/sbin/ntpd -g -n -u ntp:ntp Apr 30 03:33:53.617867 ntpd[1964]: ---------------------------------------------------- Apr 30 03:33:53.617873 ntpd[1964]: ntp-4 is maintained by Network Time Foundation, Apr 30 03:33:53.617880 ntpd[1964]: Inc. (NTF), a non-profit 501(c)(3) public-benefit Apr 30 03:33:53.617887 ntpd[1964]: corporation. Support and training for ntp-4 are Apr 30 03:33:53.617893 ntpd[1964]: available at https://www.nwtime.org/support Apr 30 03:33:53.617900 ntpd[1964]: ---------------------------------------------------- Apr 30 03:33:53.622332 tar[1977]: linux-amd64/helm Apr 30 03:33:53.623249 ntpd[1964]: proto: precision = 0.054 usec (-24) Apr 30 03:33:53.623331 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: proto: precision = 0.054 usec (-24) Apr 30 03:33:53.624144 ntpd[1964]: basedate set to 2025-04-17 Apr 30 03:33:53.626427 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: basedate set to 2025-04-17 Apr 30 03:33:53.626427 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: gps base set to 2025-04-20 (week 2363) Apr 30 03:33:53.624162 ntpd[1964]: gps base set to 2025-04-20 (week 2363) Apr 30 03:33:53.636711 ntpd[1964]: Listen and drop on 0 v6wildcard [::]:123 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Listen and drop on 0 v6wildcard [::]:123 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Listen normally on 2 lo 127.0.0.1:123 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Listen normally on 3 eth0 172.31.20.13:123 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Listen normally on 4 lo [::1]:123 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: bind(21) AF_INET6 fe80::428:24ff:fe02:ac29%2#123 flags 0x11 failed: Cannot assign requested address Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: unable to create socket on eth0 (5) for fe80::428:24ff:fe02:ac29%2#123 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: failed to init interface for address fe80::428:24ff:fe02:ac29%2 Apr 30 03:33:53.638581 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: Listening on routing socket on fd #21 for interface updates Apr 30 03:33:53.636773 ntpd[1964]: Listen and drop on 1 v4wildcard 0.0.0.0:123 Apr 30 03:33:53.636931 ntpd[1964]: Listen normally on 2 lo 127.0.0.1:123 Apr 30 03:33:53.636958 ntpd[1964]: Listen normally on 3 eth0 172.31.20.13:123 Apr 30 03:33:53.636988 ntpd[1964]: Listen normally on 4 lo [::1]:123 Apr 30 03:33:53.637020 ntpd[1964]: bind(21) AF_INET6 fe80::428:24ff:fe02:ac29%2#123 flags 0x11 failed: Cannot assign requested address Apr 30 03:33:53.637037 ntpd[1964]: unable to create socket on eth0 (5) for fe80::428:24ff:fe02:ac29%2#123 Apr 30 03:33:53.637049 ntpd[1964]: failed to init interface for address fe80::428:24ff:fe02:ac29%2 Apr 30 03:33:53.637071 ntpd[1964]: Listening on routing socket on fd #21 for interface updates Apr 30 03:33:53.644476 extend-filesystems[1960]: Resized partition /dev/nvme0n1p9 Apr 30 03:33:53.661488 extend-filesystems[2030]: resize2fs 1.47.1 (20-May-2024) Apr 30 03:33:53.662540 ntpd[1964]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Apr 30 03:33:53.663560 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Apr 30 03:33:53.663560 ntpd[1964]: 30 Apr 03:33:53 ntpd[1964]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Apr 30 03:33:53.662579 ntpd[1964]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized Apr 30 03:33:53.673594 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 1489915 blocks Apr 30 03:33:53.687206 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 30 03:33:53.688435 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Apr 30 03:33:53.725042 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (1833) Apr 30 03:33:53.752887 dbus-daemon[1958]: [system] Successfully activated service 'org.freedesktop.hostname1' Apr 30 03:33:53.753053 systemd[1]: Started systemd-hostnamed.service - Hostname Service. Apr 30 03:33:53.756197 dbus-daemon[1958]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=2004 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Apr 30 03:33:53.768543 systemd[1]: Starting polkit.service - Authorization Manager... Apr 30 03:33:53.798375 sshd_keygen[2005]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 30 03:33:53.798775 polkitd[2057]: Started polkitd version 121 Apr 30 03:33:53.804742 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 1489915 Apr 30 03:33:53.812107 polkitd[2057]: Loading rules from directory /etc/polkit-1/rules.d Apr 30 03:33:53.821847 systemd[1]: Started polkit.service - Authorization Manager. Apr 30 03:33:53.812164 polkitd[2057]: Loading rules from directory /usr/share/polkit-1/rules.d Apr 30 03:33:53.817544 polkitd[2057]: Finished loading, compiling and executing 2 rules Apr 30 03:33:53.821557 dbus-daemon[1958]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Apr 30 03:33:53.822262 polkitd[2057]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Apr 30 03:33:53.828532 extend-filesystems[2030]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Apr 30 03:33:53.828532 extend-filesystems[2030]: old_desc_blocks = 1, new_desc_blocks = 1 Apr 30 03:33:53.828532 extend-filesystems[2030]: The filesystem on /dev/nvme0n1p9 is now 1489915 (4k) blocks long. Apr 30 03:33:53.827754 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 30 03:33:53.840563 bash[2032]: Updated "/home/core/.ssh/authorized_keys" Apr 30 03:33:53.840649 extend-filesystems[1960]: Resized filesystem in /dev/nvme0n1p9 Apr 30 03:33:53.828163 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 30 03:33:53.831923 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 30 03:33:53.842319 systemd[1]: Starting sshkeys.service... Apr 30 03:33:53.846570 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 30 03:33:53.849938 locksmithd[2006]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 30 03:33:53.860715 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 30 03:33:53.894659 systemd-hostnamed[2004]: Hostname set to (transient) Apr 30 03:33:53.895039 systemd-resolved[1784]: System hostname changed to 'ip-172-31-20-13'. Apr 30 03:33:53.899746 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Apr 30 03:33:53.906800 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Apr 30 03:33:53.908570 systemd[1]: issuegen.service: Deactivated successfully. Apr 30 03:33:53.908738 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 30 03:33:53.932856 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 30 03:33:53.975251 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 30 03:33:53.989828 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 30 03:33:54.000079 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Apr 30 03:33:54.001083 systemd[1]: Reached target getty.target - Login Prompts. Apr 30 03:33:54.007231 coreos-metadata[2135]: Apr 30 03:33:54.007 INFO Putting http://169.254.169.254/latest/api/token: Attempt #1 Apr 30 03:33:54.022148 coreos-metadata[2135]: Apr 30 03:33:54.021 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys: Attempt #1 Apr 30 03:33:54.022751 coreos-metadata[2135]: Apr 30 03:33:54.022 INFO Fetch successful Apr 30 03:33:54.022751 coreos-metadata[2135]: Apr 30 03:33:54.022 INFO Fetching http://169.254.169.254/2021-01-03/meta-data/public-keys/0/openssh-key: Attempt #1 Apr 30 03:33:54.029930 coreos-metadata[2135]: Apr 30 03:33:54.029 INFO Fetch successful Apr 30 03:33:54.030017 containerd[1992]: time="2025-04-30T03:33:54.029940807Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Apr 30 03:33:54.037248 unknown[2135]: wrote ssh authorized keys file for user: core Apr 30 03:33:54.062524 containerd[1992]: time="2025-04-30T03:33:54.062397870Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.063988154Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.88-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.064019541Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.064035862Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.064166351Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.064181717Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.064229964Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.064240752Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064429 containerd[1992]: time="2025-04-30T03:33:54.064387941Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064663 containerd[1992]: time="2025-04-30T03:33:54.064416106Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064709 containerd[1992]: time="2025-04-30T03:33:54.064698184Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064754 containerd[1992]: time="2025-04-30T03:33:54.064744845Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 30 03:33:54.064861 containerd[1992]: time="2025-04-30T03:33:54.064850709Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 30 03:33:54.065085 containerd[1992]: time="2025-04-30T03:33:54.065069478Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 30 03:33:54.065254 containerd[1992]: time="2025-04-30T03:33:54.065239547Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 03:33:54.065301 containerd[1992]: time="2025-04-30T03:33:54.065291604Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 30 03:33:54.065426 containerd[1992]: time="2025-04-30T03:33:54.065413940Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 30 03:33:54.069247 containerd[1992]: time="2025-04-30T03:33:54.068393869Z" level=info msg="metadata content store policy set" policy=shared Apr 30 03:33:54.079448 containerd[1992]: time="2025-04-30T03:33:54.079335456Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 30 03:33:54.079792 containerd[1992]: time="2025-04-30T03:33:54.079583172Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 30 03:33:54.079792 containerd[1992]: time="2025-04-30T03:33:54.079607701Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 30 03:33:54.079792 containerd[1992]: time="2025-04-30T03:33:54.079660213Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 30 03:33:54.079792 containerd[1992]: time="2025-04-30T03:33:54.079676049Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 30 03:33:54.080233 update-ssh-keys[2173]: Updated "/home/core/.ssh/authorized_keys" Apr 30 03:33:54.080477 containerd[1992]: time="2025-04-30T03:33:54.080175658Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.080830933Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.080940294Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.080955782Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.080967939Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.080980941Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.080997395Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081010676Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081025072Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081040684Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081054074Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081066647Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081077988Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081096291Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081359 containerd[1992]: time="2025-04-30T03:33:54.081108768Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.080948 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081140053Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081154776Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081166093Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081178603Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081189788Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081201918Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081214091Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081227343Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081237912Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081249664Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081261441Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081275834Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081294994Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081306118Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.081745 containerd[1992]: time="2025-04-30T03:33:54.081316699Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082061796Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082087265Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082157772Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082171125Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082183001Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082200379Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082210335Z" level=info msg="NRI interface is disabled by configuration." Apr 30 03:33:54.082502 containerd[1992]: time="2025-04-30T03:33:54.082220060Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 30 03:33:54.083238 containerd[1992]: time="2025-04-30T03:33:54.082768970Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 30 03:33:54.083238 containerd[1992]: time="2025-04-30T03:33:54.082826503Z" level=info msg="Connect containerd service" Apr 30 03:33:54.083238 containerd[1992]: time="2025-04-30T03:33:54.082865474Z" level=info msg="using legacy CRI server" Apr 30 03:33:54.083238 containerd[1992]: time="2025-04-30T03:33:54.082872617Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 30 03:33:54.083238 containerd[1992]: time="2025-04-30T03:33:54.082968700Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 30 03:33:54.083871 containerd[1992]: time="2025-04-30T03:33:54.083849992Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 30 03:33:54.084044 containerd[1992]: time="2025-04-30T03:33:54.084018773Z" level=info msg="Start subscribing containerd event" Apr 30 03:33:54.084207 containerd[1992]: time="2025-04-30T03:33:54.084195802Z" level=info msg="Start recovering state" Apr 30 03:33:54.084444 containerd[1992]: time="2025-04-30T03:33:54.084296393Z" level=info msg="Start event monitor" Apr 30 03:33:54.084444 containerd[1992]: time="2025-04-30T03:33:54.084316562Z" level=info msg="Start snapshots syncer" Apr 30 03:33:54.084444 containerd[1992]: time="2025-04-30T03:33:54.084324705Z" level=info msg="Start cni network conf syncer for default" Apr 30 03:33:54.084444 containerd[1992]: time="2025-04-30T03:33:54.084331943Z" level=info msg="Start streaming server" Apr 30 03:33:54.084698 systemd[1]: Finished sshkeys.service. Apr 30 03:33:54.085036 containerd[1992]: time="2025-04-30T03:33:54.084955531Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 30 03:33:54.085036 containerd[1992]: time="2025-04-30T03:33:54.085009647Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 30 03:33:54.086448 containerd[1992]: time="2025-04-30T03:33:54.085468303Z" level=info msg="containerd successfully booted in 0.062484s" Apr 30 03:33:54.085541 systemd[1]: Started containerd.service - containerd container runtime. Apr 30 03:33:54.447551 systemd-networkd[1829]: eth0: Gained IPv6LL Apr 30 03:33:54.452373 tar[1977]: linux-amd64/README.md Apr 30 03:33:54.454103 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 30 03:33:54.468369 systemd[1]: Reached target network-online.target - Network is Online. Apr 30 03:33:54.472615 systemd[1]: Started amazon-ssm-agent.service - amazon-ssm-agent. Apr 30 03:33:54.480786 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:33:54.485016 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 30 03:33:54.488505 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 30 03:33:54.525431 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 30 03:33:54.539436 amazon-ssm-agent[2182]: Initializing new seelog logger Apr 30 03:33:54.539436 amazon-ssm-agent[2182]: New Seelog Logger Creation Complete Apr 30 03:33:54.539436 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.539436 amazon-ssm-agent[2182]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.539788 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 processing appconfig overrides Apr 30 03:33:54.539788 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.539788 amazon-ssm-agent[2182]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.539865 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 processing appconfig overrides Apr 30 03:33:54.540099 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO Proxy environment variables: Apr 30 03:33:54.540140 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.540219 amazon-ssm-agent[2182]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.540343 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 processing appconfig overrides Apr 30 03:33:54.542698 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 Found config file at /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.542775 amazon-ssm-agent[2182]: Applying config override from /etc/amazon/ssm/amazon-ssm-agent.json. Apr 30 03:33:54.542890 amazon-ssm-agent[2182]: 2025/04/30 03:33:54 processing appconfig overrides Apr 30 03:33:54.641551 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO https_proxy: Apr 30 03:33:54.739134 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO http_proxy: Apr 30 03:33:54.790160 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO no_proxy: Apr 30 03:33:54.790160 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO Checking if agent identity type OnPrem can be assumed Apr 30 03:33:54.790160 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO Checking if agent identity type EC2 can be assumed Apr 30 03:33:54.790160 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO Agent will take identity from EC2 Apr 30 03:33:54.790160 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [amazon-ssm-agent] using named pipe channel for IPC Apr 30 03:33:54.790160 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [amazon-ssm-agent] using named pipe channel for IPC Apr 30 03:33:54.790160 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [amazon-ssm-agent] using named pipe channel for IPC Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [amazon-ssm-agent] amazon-ssm-agent - v3.2.0.0 Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [amazon-ssm-agent] OS: linux, Arch: amd64 Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [amazon-ssm-agent] Starting Core Agent Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [amazon-ssm-agent] registrar detected. Attempting registration Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [Registrar] Starting registrar module Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [EC2Identity] no registration info found for ec2 instance, attempting registration Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [EC2Identity] EC2 registration was successful. Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [CredentialRefresher] credentialRefresher has started Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [CredentialRefresher] Starting credentials refresher loop Apr 30 03:33:54.790589 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO EC2RoleProvider Successfully connected with instance profile role credentials Apr 30 03:33:54.837843 amazon-ssm-agent[2182]: 2025-04-30 03:33:54 INFO [CredentialRefresher] Next credential rotation will be in 32.283327431016666 minutes Apr 30 03:33:55.807606 amazon-ssm-agent[2182]: 2025-04-30 03:33:55 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker is not running, starting worker process Apr 30 03:33:55.908036 amazon-ssm-agent[2182]: 2025-04-30 03:33:55 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] [WorkerProvider] Worker ssm-agent-worker (pid:2202) started Apr 30 03:33:56.008211 amazon-ssm-agent[2182]: 2025-04-30 03:33:55 INFO [amazon-ssm-agent] [LongRunningWorkerContainer] Monitor long running worker health every 60 seconds Apr 30 03:33:56.535286 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:33:56.536652 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 30 03:33:56.537460 systemd[1]: Startup finished in 593ms (kernel) + 5.954s (initrd) + 6.928s (userspace) = 13.476s. Apr 30 03:33:56.544429 (kubelet)[2216]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 03:33:56.618275 ntpd[1964]: Listen normally on 6 eth0 [fe80::428:24ff:fe02:ac29%2]:123 Apr 30 03:33:56.618727 ntpd[1964]: 30 Apr 03:33:56 ntpd[1964]: Listen normally on 6 eth0 [fe80::428:24ff:fe02:ac29%2]:123 Apr 30 03:33:57.619608 kubelet[2216]: E0430 03:33:57.619553 2216 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 03:33:57.621818 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 03:33:57.622037 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 03:33:57.622725 systemd[1]: kubelet.service: Consumed 1.040s CPU time. Apr 30 03:33:57.882542 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 30 03:33:57.894908 systemd[1]: Started sshd@0-172.31.20.13:22-147.75.109.163:53530.service - OpenSSH per-connection server daemon (147.75.109.163:53530). Apr 30 03:33:58.143074 sshd[2228]: Accepted publickey for core from 147.75.109.163 port 53530 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:33:58.145054 sshd[2228]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:33:58.153277 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 30 03:33:58.158854 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 30 03:33:58.160857 systemd-logind[1970]: New session 1 of user core. Apr 30 03:33:58.171390 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 30 03:33:58.177826 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 30 03:33:58.192352 (systemd)[2233]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 30 03:33:58.308334 systemd[2233]: Queued start job for default target default.target. Apr 30 03:33:58.316609 systemd[2233]: Created slice app.slice - User Application Slice. Apr 30 03:33:58.316648 systemd[2233]: Reached target paths.target - Paths. Apr 30 03:33:58.316663 systemd[2233]: Reached target timers.target - Timers. Apr 30 03:33:58.317997 systemd[2233]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 30 03:33:58.329743 systemd[2233]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 30 03:33:58.329865 systemd[2233]: Reached target sockets.target - Sockets. Apr 30 03:33:58.329879 systemd[2233]: Reached target basic.target - Basic System. Apr 30 03:33:58.329918 systemd[2233]: Reached target default.target - Main User Target. Apr 30 03:33:58.329947 systemd[2233]: Startup finished in 130ms. Apr 30 03:33:58.330081 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 30 03:33:58.340652 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 30 03:33:58.547753 systemd[1]: Started sshd@1-172.31.20.13:22-147.75.109.163:53534.service - OpenSSH per-connection server daemon (147.75.109.163:53534). Apr 30 03:33:58.790903 sshd[2244]: Accepted publickey for core from 147.75.109.163 port 53534 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:33:58.792335 sshd[2244]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:33:58.797693 systemd-logind[1970]: New session 2 of user core. Apr 30 03:33:58.803713 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 30 03:33:58.983881 sshd[2244]: pam_unix(sshd:session): session closed for user core Apr 30 03:33:58.987217 systemd[1]: sshd@1-172.31.20.13:22-147.75.109.163:53534.service: Deactivated successfully. Apr 30 03:33:58.988860 systemd[1]: session-2.scope: Deactivated successfully. Apr 30 03:33:58.989427 systemd-logind[1970]: Session 2 logged out. Waiting for processes to exit. Apr 30 03:33:58.990691 systemd-logind[1970]: Removed session 2. Apr 30 03:33:59.037746 systemd[1]: Started sshd@2-172.31.20.13:22-147.75.109.163:53540.service - OpenSSH per-connection server daemon (147.75.109.163:53540). Apr 30 03:33:59.280661 sshd[2251]: Accepted publickey for core from 147.75.109.163 port 53540 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:33:59.281929 sshd[2251]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:33:59.286183 systemd-logind[1970]: New session 3 of user core. Apr 30 03:33:59.295660 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 30 03:33:59.469638 sshd[2251]: pam_unix(sshd:session): session closed for user core Apr 30 03:33:59.472345 systemd[1]: sshd@2-172.31.20.13:22-147.75.109.163:53540.service: Deactivated successfully. Apr 30 03:33:59.473911 systemd[1]: session-3.scope: Deactivated successfully. Apr 30 03:33:59.475058 systemd-logind[1970]: Session 3 logged out. Waiting for processes to exit. Apr 30 03:33:59.476077 systemd-logind[1970]: Removed session 3. Apr 30 03:33:59.514145 systemd[1]: Started sshd@3-172.31.20.13:22-147.75.109.163:53550.service - OpenSSH per-connection server daemon (147.75.109.163:53550). Apr 30 03:33:59.757451 sshd[2258]: Accepted publickey for core from 147.75.109.163 port 53550 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:33:59.759134 sshd[2258]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:33:59.770771 systemd-logind[1970]: New session 4 of user core. Apr 30 03:33:59.777654 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 30 03:33:59.948121 sshd[2258]: pam_unix(sshd:session): session closed for user core Apr 30 03:33:59.950895 systemd[1]: sshd@3-172.31.20.13:22-147.75.109.163:53550.service: Deactivated successfully. Apr 30 03:33:59.952481 systemd[1]: session-4.scope: Deactivated successfully. Apr 30 03:33:59.954008 systemd-logind[1970]: Session 4 logged out. Waiting for processes to exit. Apr 30 03:33:59.955088 systemd-logind[1970]: Removed session 4. Apr 30 03:33:59.995282 systemd[1]: Started sshd@4-172.31.20.13:22-147.75.109.163:53566.service - OpenSSH per-connection server daemon (147.75.109.163:53566). Apr 30 03:34:00.246775 sshd[2265]: Accepted publickey for core from 147.75.109.163 port 53566 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:34:00.248142 sshd[2265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:34:00.253312 systemd-logind[1970]: New session 5 of user core. Apr 30 03:34:00.262993 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 30 03:34:00.415826 sudo[2268]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 30 03:34:00.416223 sudo[2268]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 03:34:00.428333 sudo[2268]: pam_unix(sudo:session): session closed for user root Apr 30 03:34:00.465699 sshd[2265]: pam_unix(sshd:session): session closed for user core Apr 30 03:34:00.469336 systemd[1]: sshd@4-172.31.20.13:22-147.75.109.163:53566.service: Deactivated successfully. Apr 30 03:34:00.471500 systemd[1]: session-5.scope: Deactivated successfully. Apr 30 03:34:00.472962 systemd-logind[1970]: Session 5 logged out. Waiting for processes to exit. Apr 30 03:34:00.474230 systemd-logind[1970]: Removed session 5. Apr 30 03:34:00.515781 systemd[1]: Started sshd@5-172.31.20.13:22-147.75.109.163:53568.service - OpenSSH per-connection server daemon (147.75.109.163:53568). Apr 30 03:34:00.987168 systemd-resolved[1784]: Clock change detected. Flushing caches. Apr 30 03:34:01.126075 sshd[2273]: Accepted publickey for core from 147.75.109.163 port 53568 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:34:01.127784 sshd[2273]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:34:01.133159 systemd-logind[1970]: New session 6 of user core. Apr 30 03:34:01.144243 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 30 03:34:01.281661 sudo[2277]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 30 03:34:01.281964 sudo[2277]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 03:34:01.286158 sudo[2277]: pam_unix(sudo:session): session closed for user root Apr 30 03:34:01.292266 sudo[2276]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Apr 30 03:34:01.292674 sudo[2276]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 03:34:01.305388 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Apr 30 03:34:01.317392 auditctl[2280]: No rules Apr 30 03:34:01.317859 systemd[1]: audit-rules.service: Deactivated successfully. Apr 30 03:34:01.318098 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Apr 30 03:34:01.324517 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Apr 30 03:34:01.353062 augenrules[2298]: No rules Apr 30 03:34:01.354482 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Apr 30 03:34:01.355975 sudo[2276]: pam_unix(sudo:session): session closed for user root Apr 30 03:34:01.394293 sshd[2273]: pam_unix(sshd:session): session closed for user core Apr 30 03:34:01.401367 systemd[1]: sshd@5-172.31.20.13:22-147.75.109.163:53568.service: Deactivated successfully. Apr 30 03:34:01.403614 systemd[1]: session-6.scope: Deactivated successfully. Apr 30 03:34:01.405542 systemd-logind[1970]: Session 6 logged out. Waiting for processes to exit. Apr 30 03:34:01.413609 systemd-logind[1970]: Removed session 6. Apr 30 03:34:01.442423 systemd[1]: Started sshd@6-172.31.20.13:22-147.75.109.163:53570.service - OpenSSH per-connection server daemon (147.75.109.163:53570). Apr 30 03:34:01.684954 sshd[2306]: Accepted publickey for core from 147.75.109.163 port 53570 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:34:01.692597 sshd[2306]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:34:01.704497 systemd-logind[1970]: New session 7 of user core. Apr 30 03:34:01.716327 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 30 03:34:01.839581 sudo[2309]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 30 03:34:01.839980 sudo[2309]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 03:34:02.372398 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 30 03:34:02.373690 (dockerd)[2325]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 30 03:34:02.794210 dockerd[2325]: time="2025-04-30T03:34:02.794147110Z" level=info msg="Starting up" Apr 30 03:34:02.893550 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport907156743-merged.mount: Deactivated successfully. Apr 30 03:34:02.985090 dockerd[2325]: time="2025-04-30T03:34:02.985045344Z" level=info msg="Loading containers: start." Apr 30 03:34:03.130151 kernel: Initializing XFRM netlink socket Apr 30 03:34:03.161587 (udev-worker)[2348]: Network interface NamePolicy= disabled on kernel command line. Apr 30 03:34:03.222190 systemd-networkd[1829]: docker0: Link UP Apr 30 03:34:03.247836 dockerd[2325]: time="2025-04-30T03:34:03.247786851Z" level=info msg="Loading containers: done." Apr 30 03:34:03.267257 dockerd[2325]: time="2025-04-30T03:34:03.267203091Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 30 03:34:03.267425 dockerd[2325]: time="2025-04-30T03:34:03.267312527Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Apr 30 03:34:03.267425 dockerd[2325]: time="2025-04-30T03:34:03.267415925Z" level=info msg="Daemon has completed initialization" Apr 30 03:34:03.301308 dockerd[2325]: time="2025-04-30T03:34:03.301126144Z" level=info msg="API listen on /run/docker.sock" Apr 30 03:34:03.301595 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 30 03:34:04.588843 containerd[1992]: time="2025-04-30T03:34:04.588804629Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\"" Apr 30 03:34:05.174060 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3761007038.mount: Deactivated successfully. Apr 30 03:34:07.057549 containerd[1992]: time="2025-04-30T03:34:07.057479905Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:07.058601 containerd[1992]: time="2025-04-30T03:34:07.058550221Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.4: active requests=0, bytes read=28682879" Apr 30 03:34:07.060204 containerd[1992]: time="2025-04-30T03:34:07.060152439Z" level=info msg="ImageCreate event name:\"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:07.063758 containerd[1992]: time="2025-04-30T03:34:07.063706464Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:07.065080 containerd[1992]: time="2025-04-30T03:34:07.064861597Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.4\" with image id \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\", size \"28679679\" in 2.476014645s" Apr 30 03:34:07.065080 containerd[1992]: time="2025-04-30T03:34:07.064909150Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\" returns image reference \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\"" Apr 30 03:34:07.065935 containerd[1992]: time="2025-04-30T03:34:07.065752673Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\"" Apr 30 03:34:08.017811 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 30 03:34:08.025885 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:34:08.214152 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:34:08.218742 (kubelet)[2525]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 03:34:08.270167 kubelet[2525]: E0430 03:34:08.270037 2525 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 03:34:08.274117 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 03:34:08.274318 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 03:34:09.188845 containerd[1992]: time="2025-04-30T03:34:09.188790760Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:09.190150 containerd[1992]: time="2025-04-30T03:34:09.190103081Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.4: active requests=0, bytes read=24779589" Apr 30 03:34:09.191246 containerd[1992]: time="2025-04-30T03:34:09.191187360Z" level=info msg="ImageCreate event name:\"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:09.195267 containerd[1992]: time="2025-04-30T03:34:09.194881319Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:09.196043 containerd[1992]: time="2025-04-30T03:34:09.195988827Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.4\" with image id \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\", size \"26267962\" in 2.130200581s" Apr 30 03:34:09.196137 containerd[1992]: time="2025-04-30T03:34:09.196047116Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\" returns image reference \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\"" Apr 30 03:34:09.196776 containerd[1992]: time="2025-04-30T03:34:09.196752210Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\"" Apr 30 03:34:10.954155 containerd[1992]: time="2025-04-30T03:34:10.954100411Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:10.956087 containerd[1992]: time="2025-04-30T03:34:10.956028192Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.4: active requests=0, bytes read=19169938" Apr 30 03:34:10.957446 containerd[1992]: time="2025-04-30T03:34:10.957396274Z" level=info msg="ImageCreate event name:\"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:10.961116 containerd[1992]: time="2025-04-30T03:34:10.961079345Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:10.962147 containerd[1992]: time="2025-04-30T03:34:10.962117851Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.4\" with image id \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\", size \"20658329\" in 1.765250093s" Apr 30 03:34:10.962343 containerd[1992]: time="2025-04-30T03:34:10.962231497Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\" returns image reference \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\"" Apr 30 03:34:10.962737 containerd[1992]: time="2025-04-30T03:34:10.962695580Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\"" Apr 30 03:34:12.326602 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4084986796.mount: Deactivated successfully. Apr 30 03:34:12.811525 containerd[1992]: time="2025-04-30T03:34:12.811459689Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:12.812509 containerd[1992]: time="2025-04-30T03:34:12.812447718Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.4: active requests=0, bytes read=30917856" Apr 30 03:34:12.813589 containerd[1992]: time="2025-04-30T03:34:12.813536565Z" level=info msg="ImageCreate event name:\"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:12.816030 containerd[1992]: time="2025-04-30T03:34:12.815660679Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:12.816332 containerd[1992]: time="2025-04-30T03:34:12.816198006Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.4\" with image id \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\", repo tag \"registry.k8s.io/kube-proxy:v1.32.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\", size \"30916875\" in 1.853358137s" Apr 30 03:34:12.816332 containerd[1992]: time="2025-04-30T03:34:12.816230913Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\" returns image reference \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\"" Apr 30 03:34:12.816835 containerd[1992]: time="2025-04-30T03:34:12.816737713Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Apr 30 03:34:13.446361 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1553862734.mount: Deactivated successfully. Apr 30 03:34:14.403667 containerd[1992]: time="2025-04-30T03:34:14.403607984Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:14.404746 containerd[1992]: time="2025-04-30T03:34:14.404698128Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Apr 30 03:34:14.405786 containerd[1992]: time="2025-04-30T03:34:14.405720658Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:14.410573 containerd[1992]: time="2025-04-30T03:34:14.408972633Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:14.410573 containerd[1992]: time="2025-04-30T03:34:14.410147451Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.593377275s" Apr 30 03:34:14.410573 containerd[1992]: time="2025-04-30T03:34:14.410188868Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Apr 30 03:34:14.411234 containerd[1992]: time="2025-04-30T03:34:14.411212526Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 30 03:34:14.980366 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1747602404.mount: Deactivated successfully. Apr 30 03:34:14.987839 containerd[1992]: time="2025-04-30T03:34:14.987792962Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:14.989892 containerd[1992]: time="2025-04-30T03:34:14.989637587Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Apr 30 03:34:14.992716 containerd[1992]: time="2025-04-30T03:34:14.991257256Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:14.995221 containerd[1992]: time="2025-04-30T03:34:14.995168362Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:14.996216 containerd[1992]: time="2025-04-30T03:34:14.995733160Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 584.361946ms" Apr 30 03:34:14.996216 containerd[1992]: time="2025-04-30T03:34:14.995765964Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Apr 30 03:34:14.996637 containerd[1992]: time="2025-04-30T03:34:14.996610568Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Apr 30 03:34:15.597411 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2204458293.mount: Deactivated successfully. Apr 30 03:34:17.937286 containerd[1992]: time="2025-04-30T03:34:17.937225489Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:17.938435 containerd[1992]: time="2025-04-30T03:34:17.938379747Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551360" Apr 30 03:34:17.940763 containerd[1992]: time="2025-04-30T03:34:17.940701882Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:17.943823 containerd[1992]: time="2025-04-30T03:34:17.943762743Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:17.945092 containerd[1992]: time="2025-04-30T03:34:17.944930457Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.948288432s" Apr 30 03:34:17.945092 containerd[1992]: time="2025-04-30T03:34:17.944965575Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Apr 30 03:34:18.436750 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 30 03:34:18.442337 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:34:18.682160 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:34:18.686321 (kubelet)[2685]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 03:34:18.753410 kubelet[2685]: E0430 03:34:18.753295 2685 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 03:34:18.756863 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 03:34:18.757186 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 03:34:20.818885 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:34:20.825362 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:34:20.867053 systemd[1]: Reloading requested from client PID 2700 ('systemctl') (unit session-7.scope)... Apr 30 03:34:20.867227 systemd[1]: Reloading... Apr 30 03:34:20.990058 zram_generator::config[2741]: No configuration found. Apr 30 03:34:21.141528 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 03:34:21.228235 systemd[1]: Reloading finished in 360 ms. Apr 30 03:34:21.288929 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:34:21.292829 systemd[1]: kubelet.service: Deactivated successfully. Apr 30 03:34:21.293117 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:34:21.297383 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:34:21.505808 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:34:21.516485 (kubelet)[2806]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 30 03:34:21.566289 kubelet[2806]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 03:34:21.566289 kubelet[2806]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 30 03:34:21.566289 kubelet[2806]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 03:34:21.566697 kubelet[2806]: I0430 03:34:21.566347 2806 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 30 03:34:21.769334 kubelet[2806]: I0430 03:34:21.769292 2806 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Apr 30 03:34:21.769334 kubelet[2806]: I0430 03:34:21.769322 2806 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 30 03:34:21.769710 kubelet[2806]: I0430 03:34:21.769674 2806 server.go:954] "Client rotation is on, will bootstrap in background" Apr 30 03:34:21.834486 kubelet[2806]: I0430 03:34:21.834442 2806 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 30 03:34:21.835751 kubelet[2806]: E0430 03:34:21.835699 2806 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.20.13:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:21.856051 kubelet[2806]: E0430 03:34:21.855981 2806 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 30 03:34:21.856051 kubelet[2806]: I0430 03:34:21.856044 2806 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 30 03:34:21.860813 kubelet[2806]: I0430 03:34:21.860780 2806 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 30 03:34:21.866448 kubelet[2806]: I0430 03:34:21.866384 2806 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 30 03:34:21.866677 kubelet[2806]: I0430 03:34:21.866443 2806 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-20-13","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 30 03:34:21.868774 kubelet[2806]: I0430 03:34:21.868729 2806 topology_manager.go:138] "Creating topology manager with none policy" Apr 30 03:34:21.868774 kubelet[2806]: I0430 03:34:21.868766 2806 container_manager_linux.go:304] "Creating device plugin manager" Apr 30 03:34:21.868947 kubelet[2806]: I0430 03:34:21.868928 2806 state_mem.go:36] "Initialized new in-memory state store" Apr 30 03:34:21.875378 kubelet[2806]: I0430 03:34:21.875340 2806 kubelet.go:446] "Attempting to sync node with API server" Apr 30 03:34:21.875378 kubelet[2806]: I0430 03:34:21.875373 2806 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 30 03:34:21.875378 kubelet[2806]: I0430 03:34:21.875400 2806 kubelet.go:352] "Adding apiserver pod source" Apr 30 03:34:21.875743 kubelet[2806]: I0430 03:34:21.875411 2806 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 30 03:34:21.882553 kubelet[2806]: W0430 03:34:21.882328 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.20.13:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:21.882553 kubelet[2806]: E0430 03:34:21.882384 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.20.13:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:21.882711 kubelet[2806]: W0430 03:34:21.882627 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.20.13:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-13&limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:21.882711 kubelet[2806]: E0430 03:34:21.882658 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.20.13:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-13&limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:21.882764 kubelet[2806]: I0430 03:34:21.882756 2806 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 30 03:34:21.887759 kubelet[2806]: I0430 03:34:21.887642 2806 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Apr 30 03:34:21.888827 kubelet[2806]: W0430 03:34:21.888627 2806 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 30 03:34:21.891267 kubelet[2806]: I0430 03:34:21.891243 2806 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 30 03:34:21.891361 kubelet[2806]: I0430 03:34:21.891279 2806 server.go:1287] "Started kubelet" Apr 30 03:34:21.896027 kubelet[2806]: I0430 03:34:21.895975 2806 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Apr 30 03:34:21.899576 kubelet[2806]: I0430 03:34:21.898998 2806 server.go:490] "Adding debug handlers to kubelet server" Apr 30 03:34:21.903047 kubelet[2806]: I0430 03:34:21.902606 2806 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 30 03:34:21.903047 kubelet[2806]: I0430 03:34:21.902959 2806 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 30 03:34:21.907346 kubelet[2806]: I0430 03:34:21.906542 2806 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 30 03:34:21.910117 kubelet[2806]: E0430 03:34:21.904470 2806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.20.13:6443/api/v1/namespaces/default/events\": dial tcp 172.31.20.13:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-20-13.183afb40f085acce default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-20-13,UID:ip-172-31-20-13,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-20-13,},FirstTimestamp:2025-04-30 03:34:21.891259598 +0000 UTC m=+0.369872071,LastTimestamp:2025-04-30 03:34:21.891259598 +0000 UTC m=+0.369872071,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-20-13,}" Apr 30 03:34:21.913056 kubelet[2806]: I0430 03:34:21.912108 2806 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 30 03:34:21.915117 kubelet[2806]: E0430 03:34:21.915094 2806 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ip-172-31-20-13\" not found" Apr 30 03:34:21.918309 kubelet[2806]: I0430 03:34:21.918290 2806 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 30 03:34:21.923264 kubelet[2806]: I0430 03:34:21.923237 2806 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Apr 30 03:34:21.925577 kubelet[2806]: I0430 03:34:21.925271 2806 factory.go:221] Registration of the systemd container factory successfully Apr 30 03:34:21.925577 kubelet[2806]: I0430 03:34:21.925405 2806 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 30 03:34:21.926189 kubelet[2806]: I0430 03:34:21.926164 2806 reconciler.go:26] "Reconciler: start to sync state" Apr 30 03:34:21.927567 kubelet[2806]: W0430 03:34:21.927119 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.20.13:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:21.927567 kubelet[2806]: E0430 03:34:21.927188 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.20.13:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:21.927567 kubelet[2806]: E0430 03:34:21.927275 2806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-13?timeout=10s\": dial tcp 172.31.20.13:6443: connect: connection refused" interval="200ms" Apr 30 03:34:21.935021 kubelet[2806]: E0430 03:34:21.934034 2806 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 30 03:34:21.936281 kubelet[2806]: I0430 03:34:21.936186 2806 factory.go:221] Registration of the containerd container factory successfully Apr 30 03:34:21.953024 kubelet[2806]: I0430 03:34:21.951472 2806 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 30 03:34:21.953151 kubelet[2806]: I0430 03:34:21.953118 2806 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 30 03:34:21.953151 kubelet[2806]: I0430 03:34:21.953141 2806 status_manager.go:227] "Starting to sync pod status with apiserver" Apr 30 03:34:21.953247 kubelet[2806]: I0430 03:34:21.953161 2806 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 30 03:34:21.953247 kubelet[2806]: I0430 03:34:21.953171 2806 kubelet.go:2388] "Starting kubelet main sync loop" Apr 30 03:34:21.953247 kubelet[2806]: E0430 03:34:21.953223 2806 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 30 03:34:21.962367 kubelet[2806]: W0430 03:34:21.962330 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.20.13:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:21.962504 kubelet[2806]: E0430 03:34:21.962376 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.20.13:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:21.965064 kubelet[2806]: I0430 03:34:21.965031 2806 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 30 03:34:21.965064 kubelet[2806]: I0430 03:34:21.965057 2806 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 30 03:34:21.965221 kubelet[2806]: I0430 03:34:21.965081 2806 state_mem.go:36] "Initialized new in-memory state store" Apr 30 03:34:21.967552 kubelet[2806]: I0430 03:34:21.967509 2806 policy_none.go:49] "None policy: Start" Apr 30 03:34:21.967552 kubelet[2806]: I0430 03:34:21.967535 2806 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 30 03:34:21.967552 kubelet[2806]: I0430 03:34:21.967547 2806 state_mem.go:35] "Initializing new in-memory state store" Apr 30 03:34:21.973834 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 30 03:34:21.987541 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 30 03:34:21.991449 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 30 03:34:22.003371 kubelet[2806]: I0430 03:34:22.003231 2806 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 30 03:34:22.003472 kubelet[2806]: I0430 03:34:22.003414 2806 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 30 03:34:22.003472 kubelet[2806]: I0430 03:34:22.003433 2806 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 30 03:34:22.005607 kubelet[2806]: I0430 03:34:22.003825 2806 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 30 03:34:22.005607 kubelet[2806]: E0430 03:34:22.004819 2806 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 30 03:34:22.005607 kubelet[2806]: E0430 03:34:22.004850 2806 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-172-31-20-13\" not found" Apr 30 03:34:22.061840 systemd[1]: Created slice kubepods-burstable-pod739b71698845d25aff1f725b42fc217d.slice - libcontainer container kubepods-burstable-pod739b71698845d25aff1f725b42fc217d.slice. Apr 30 03:34:22.085677 kubelet[2806]: E0430 03:34:22.085649 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:22.088728 systemd[1]: Created slice kubepods-burstable-pod5daeed804a0bd8056e6c7202a23fb658.slice - libcontainer container kubepods-burstable-pod5daeed804a0bd8056e6c7202a23fb658.slice. Apr 30 03:34:22.091334 kubelet[2806]: E0430 03:34:22.091161 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:22.093628 systemd[1]: Created slice kubepods-burstable-podfb4d70b6b6bcc7a5e8cce843eaf8b6fb.slice - libcontainer container kubepods-burstable-podfb4d70b6b6bcc7a5e8cce843eaf8b6fb.slice. Apr 30 03:34:22.095466 kubelet[2806]: E0430 03:34:22.095439 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:22.105921 kubelet[2806]: I0430 03:34:22.105884 2806 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-20-13" Apr 30 03:34:22.106221 kubelet[2806]: E0430 03:34:22.106199 2806 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.20.13:6443/api/v1/nodes\": dial tcp 172.31.20.13:6443: connect: connection refused" node="ip-172-31-20-13" Apr 30 03:34:22.127415 kubelet[2806]: I0430 03:34:22.127374 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/739b71698845d25aff1f725b42fc217d-ca-certs\") pod \"kube-apiserver-ip-172-31-20-13\" (UID: \"739b71698845d25aff1f725b42fc217d\") " pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:22.127726 kubelet[2806]: I0430 03:34:22.127552 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/739b71698845d25aff1f725b42fc217d-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-13\" (UID: \"739b71698845d25aff1f725b42fc217d\") " pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:22.127726 kubelet[2806]: I0430 03:34:22.127578 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/739b71698845d25aff1f725b42fc217d-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-13\" (UID: \"739b71698845d25aff1f725b42fc217d\") " pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:22.127726 kubelet[2806]: I0430 03:34:22.127596 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fb4d70b6b6bcc7a5e8cce843eaf8b6fb-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-13\" (UID: \"fb4d70b6b6bcc7a5e8cce843eaf8b6fb\") " pod="kube-system/kube-scheduler-ip-172-31-20-13" Apr 30 03:34:22.127726 kubelet[2806]: I0430 03:34:22.127613 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:22.127726 kubelet[2806]: I0430 03:34:22.127639 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:22.127908 kubelet[2806]: E0430 03:34:22.127636 2806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-13?timeout=10s\": dial tcp 172.31.20.13:6443: connect: connection refused" interval="400ms" Apr 30 03:34:22.127908 kubelet[2806]: I0430 03:34:22.127654 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:22.127908 kubelet[2806]: I0430 03:34:22.127668 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:22.127908 kubelet[2806]: I0430 03:34:22.127683 2806 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:22.308237 kubelet[2806]: I0430 03:34:22.308156 2806 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-20-13" Apr 30 03:34:22.308456 kubelet[2806]: E0430 03:34:22.308433 2806 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.20.13:6443/api/v1/nodes\": dial tcp 172.31.20.13:6443: connect: connection refused" node="ip-172-31-20-13" Apr 30 03:34:22.387649 containerd[1992]: time="2025-04-30T03:34:22.387537722Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-13,Uid:739b71698845d25aff1f725b42fc217d,Namespace:kube-system,Attempt:0,}" Apr 30 03:34:22.399779 containerd[1992]: time="2025-04-30T03:34:22.399513712Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-13,Uid:fb4d70b6b6bcc7a5e8cce843eaf8b6fb,Namespace:kube-system,Attempt:0,}" Apr 30 03:34:22.399779 containerd[1992]: time="2025-04-30T03:34:22.399514424Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-13,Uid:5daeed804a0bd8056e6c7202a23fb658,Namespace:kube-system,Attempt:0,}" Apr 30 03:34:22.528589 kubelet[2806]: E0430 03:34:22.528549 2806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-13?timeout=10s\": dial tcp 172.31.20.13:6443: connect: connection refused" interval="800ms" Apr 30 03:34:22.710067 kubelet[2806]: I0430 03:34:22.709931 2806 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-20-13" Apr 30 03:34:22.710694 kubelet[2806]: E0430 03:34:22.710528 2806 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.20.13:6443/api/v1/nodes\": dial tcp 172.31.20.13:6443: connect: connection refused" node="ip-172-31-20-13" Apr 30 03:34:22.735892 kubelet[2806]: W0430 03:34:22.735833 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://172.31.20.13:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-13&limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:22.736071 kubelet[2806]: E0430 03:34:22.735904 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://172.31.20.13:6443/api/v1/nodes?fieldSelector=metadata.name%3Dip-172-31-20-13&limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:22.870056 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1147389087.mount: Deactivated successfully. Apr 30 03:34:22.880471 containerd[1992]: time="2025-04-30T03:34:22.880416784Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 03:34:22.881451 containerd[1992]: time="2025-04-30T03:34:22.881397605Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 03:34:22.888636 containerd[1992]: time="2025-04-30T03:34:22.886928721Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 30 03:34:22.888636 containerd[1992]: time="2025-04-30T03:34:22.886980304Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 03:34:22.888851 containerd[1992]: time="2025-04-30T03:34:22.888822318Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 03:34:22.889938 containerd[1992]: time="2025-04-30T03:34:22.889881230Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 30 03:34:22.891259 containerd[1992]: time="2025-04-30T03:34:22.891185296Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Apr 30 03:34:22.892542 containerd[1992]: time="2025-04-30T03:34:22.892468341Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 03:34:22.895216 containerd[1992]: time="2025-04-30T03:34:22.894390628Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 494.790565ms" Apr 30 03:34:22.896279 containerd[1992]: time="2025-04-30T03:34:22.896141889Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 496.339381ms" Apr 30 03:34:22.901223 containerd[1992]: time="2025-04-30T03:34:22.901154907Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 513.53013ms" Apr 30 03:34:22.953855 kubelet[2806]: W0430 03:34:22.953765 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://172.31.20.13:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:22.953855 kubelet[2806]: E0430 03:34:22.953830 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://172.31.20.13:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:23.064320 containerd[1992]: time="2025-04-30T03:34:23.062813410Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:34:23.064320 containerd[1992]: time="2025-04-30T03:34:23.062882891Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:34:23.064320 containerd[1992]: time="2025-04-30T03:34:23.062894618Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:23.064320 containerd[1992]: time="2025-04-30T03:34:23.062990612Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:23.070030 containerd[1992]: time="2025-04-30T03:34:23.069544498Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:34:23.070030 containerd[1992]: time="2025-04-30T03:34:23.069598622Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:34:23.070030 containerd[1992]: time="2025-04-30T03:34:23.069614698Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:23.070030 containerd[1992]: time="2025-04-30T03:34:23.069685143Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:23.080463 containerd[1992]: time="2025-04-30T03:34:23.080341400Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:34:23.080463 containerd[1992]: time="2025-04-30T03:34:23.080388905Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:34:23.080463 containerd[1992]: time="2025-04-30T03:34:23.080403077Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:23.080841 containerd[1992]: time="2025-04-30T03:34:23.080472422Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:23.093205 systemd[1]: Started cri-containerd-0fb05fba6baf3ddf48c1a1b263a4a233b276c8b3e1262eb16bdedcbbdd2b6eff.scope - libcontainer container 0fb05fba6baf3ddf48c1a1b263a4a233b276c8b3e1262eb16bdedcbbdd2b6eff. Apr 30 03:34:23.101275 systemd[1]: Started cri-containerd-fd706056d2aea998d870b25264e075c95063aea9b4239b51e33bd4a0a773e582.scope - libcontainer container fd706056d2aea998d870b25264e075c95063aea9b4239b51e33bd4a0a773e582. Apr 30 03:34:23.121272 systemd[1]: Started cri-containerd-d1b18b6167552f4eb7dbe629997ba1d2d6a2845cc6896e2772332d0042ada763.scope - libcontainer container d1b18b6167552f4eb7dbe629997ba1d2d6a2845cc6896e2772332d0042ada763. Apr 30 03:34:23.203585 containerd[1992]: time="2025-04-30T03:34:23.203420515Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ip-172-31-20-13,Uid:5daeed804a0bd8056e6c7202a23fb658,Namespace:kube-system,Attempt:0,} returns sandbox id \"d1b18b6167552f4eb7dbe629997ba1d2d6a2845cc6896e2772332d0042ada763\"" Apr 30 03:34:23.205420 kubelet[2806]: W0430 03:34:23.205322 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://172.31.20.13:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:23.205420 kubelet[2806]: E0430 03:34:23.205370 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://172.31.20.13:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:23.207548 containerd[1992]: time="2025-04-30T03:34:23.207511917Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ip-172-31-20-13,Uid:fb4d70b6b6bcc7a5e8cce843eaf8b6fb,Namespace:kube-system,Attempt:0,} returns sandbox id \"fd706056d2aea998d870b25264e075c95063aea9b4239b51e33bd4a0a773e582\"" Apr 30 03:34:23.217844 containerd[1992]: time="2025-04-30T03:34:23.217794906Z" level=info msg="CreateContainer within sandbox \"d1b18b6167552f4eb7dbe629997ba1d2d6a2845cc6896e2772332d0042ada763\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 30 03:34:23.218783 containerd[1992]: time="2025-04-30T03:34:23.218314713Z" level=info msg="CreateContainer within sandbox \"fd706056d2aea998d870b25264e075c95063aea9b4239b51e33bd4a0a773e582\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 30 03:34:23.240220 containerd[1992]: time="2025-04-30T03:34:23.240186007Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ip-172-31-20-13,Uid:739b71698845d25aff1f725b42fc217d,Namespace:kube-system,Attempt:0,} returns sandbox id \"0fb05fba6baf3ddf48c1a1b263a4a233b276c8b3e1262eb16bdedcbbdd2b6eff\"" Apr 30 03:34:23.245195 containerd[1992]: time="2025-04-30T03:34:23.245155656Z" level=info msg="CreateContainer within sandbox \"0fb05fba6baf3ddf48c1a1b263a4a233b276c8b3e1262eb16bdedcbbdd2b6eff\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 30 03:34:23.250516 containerd[1992]: time="2025-04-30T03:34:23.250448139Z" level=info msg="CreateContainer within sandbox \"d1b18b6167552f4eb7dbe629997ba1d2d6a2845cc6896e2772332d0042ada763\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514\"" Apr 30 03:34:23.251970 containerd[1992]: time="2025-04-30T03:34:23.251879707Z" level=info msg="StartContainer for \"9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514\"" Apr 30 03:34:23.266762 containerd[1992]: time="2025-04-30T03:34:23.266235469Z" level=info msg="CreateContainer within sandbox \"fd706056d2aea998d870b25264e075c95063aea9b4239b51e33bd4a0a773e582\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a\"" Apr 30 03:34:23.267396 containerd[1992]: time="2025-04-30T03:34:23.267266616Z" level=info msg="StartContainer for \"ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a\"" Apr 30 03:34:23.273788 containerd[1992]: time="2025-04-30T03:34:23.273320848Z" level=info msg="CreateContainer within sandbox \"0fb05fba6baf3ddf48c1a1b263a4a233b276c8b3e1262eb16bdedcbbdd2b6eff\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"b795750b21e3cc81a0d85790f8a06bf6b704cffbf2edc5a4e7d3543b02ab7301\"" Apr 30 03:34:23.276669 containerd[1992]: time="2025-04-30T03:34:23.275311488Z" level=info msg="StartContainer for \"b795750b21e3cc81a0d85790f8a06bf6b704cffbf2edc5a4e7d3543b02ab7301\"" Apr 30 03:34:23.301626 systemd[1]: Started cri-containerd-9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514.scope - libcontainer container 9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514. Apr 30 03:34:23.320259 systemd[1]: Started cri-containerd-ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a.scope - libcontainer container ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a. Apr 30 03:34:23.330295 kubelet[2806]: E0430 03:34:23.329447 2806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://172.31.20.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-13?timeout=10s\": dial tcp 172.31.20.13:6443: connect: connection refused" interval="1.6s" Apr 30 03:34:23.344233 systemd[1]: Started cri-containerd-b795750b21e3cc81a0d85790f8a06bf6b704cffbf2edc5a4e7d3543b02ab7301.scope - libcontainer container b795750b21e3cc81a0d85790f8a06bf6b704cffbf2edc5a4e7d3543b02ab7301. Apr 30 03:34:23.405216 containerd[1992]: time="2025-04-30T03:34:23.404947432Z" level=info msg="StartContainer for \"9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514\" returns successfully" Apr 30 03:34:23.422133 containerd[1992]: time="2025-04-30T03:34:23.421885934Z" level=info msg="StartContainer for \"b795750b21e3cc81a0d85790f8a06bf6b704cffbf2edc5a4e7d3543b02ab7301\" returns successfully" Apr 30 03:34:23.453312 containerd[1992]: time="2025-04-30T03:34:23.452775529Z" level=info msg="StartContainer for \"ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a\" returns successfully" Apr 30 03:34:23.476557 kubelet[2806]: W0430 03:34:23.476449 2806 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://172.31.20.13:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 172.31.20.13:6443: connect: connection refused Apr 30 03:34:23.476771 kubelet[2806]: E0430 03:34:23.476528 2806 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://172.31.20.13:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:23.513572 kubelet[2806]: I0430 03:34:23.513031 2806 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-20-13" Apr 30 03:34:23.513572 kubelet[2806]: E0430 03:34:23.513358 2806 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://172.31.20.13:6443/api/v1/nodes\": dial tcp 172.31.20.13:6443: connect: connection refused" node="ip-172-31-20-13" Apr 30 03:34:23.964460 kubelet[2806]: E0430 03:34:23.964411 2806 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://172.31.20.13:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 172.31.20.13:6443: connect: connection refused" logger="UnhandledError" Apr 30 03:34:23.978276 kubelet[2806]: E0430 03:34:23.977453 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:23.985613 kubelet[2806]: E0430 03:34:23.983684 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:23.996874 kubelet[2806]: E0430 03:34:23.996580 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:24.297734 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Apr 30 03:34:24.340588 kubelet[2806]: E0430 03:34:24.340463 2806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://172.31.20.13:6443/api/v1/namespaces/default/events\": dial tcp 172.31.20.13:6443: connect: connection refused" event="&Event{ObjectMeta:{ip-172-31-20-13.183afb40f085acce default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-172-31-20-13,UID:ip-172-31-20-13,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-172-31-20-13,},FirstTimestamp:2025-04-30 03:34:21.891259598 +0000 UTC m=+0.369872071,LastTimestamp:2025-04-30 03:34:21.891259598 +0000 UTC m=+0.369872071,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-172-31-20-13,}" Apr 30 03:34:24.994035 kubelet[2806]: E0430 03:34:24.993746 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:24.997052 kubelet[2806]: E0430 03:34:24.996705 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:24.997509 kubelet[2806]: E0430 03:34:24.996963 2806 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:25.118297 kubelet[2806]: I0430 03:34:25.117493 2806 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-20-13" Apr 30 03:34:26.509308 kubelet[2806]: E0430 03:34:26.509271 2806 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-172-31-20-13\" not found" node="ip-172-31-20-13" Apr 30 03:34:26.612349 kubelet[2806]: I0430 03:34:26.611948 2806 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-20-13" Apr 30 03:34:26.622127 kubelet[2806]: I0430 03:34:26.621861 2806 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:26.634547 kubelet[2806]: E0430 03:34:26.634511 2806 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ip-172-31-20-13\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:26.635045 kubelet[2806]: I0430 03:34:26.634765 2806 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:26.639219 kubelet[2806]: E0430 03:34:26.639193 2806 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ip-172-31-20-13\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:26.639609 kubelet[2806]: I0430 03:34:26.639431 2806 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-20-13" Apr 30 03:34:26.641560 kubelet[2806]: E0430 03:34:26.641532 2806 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-20-13\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ip-172-31-20-13" Apr 30 03:34:26.882148 kubelet[2806]: I0430 03:34:26.881847 2806 apiserver.go:52] "Watching apiserver" Apr 30 03:34:26.923636 kubelet[2806]: I0430 03:34:26.923586 2806 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Apr 30 03:34:28.492552 systemd[1]: Reloading requested from client PID 3087 ('systemctl') (unit session-7.scope)... Apr 30 03:34:28.492570 systemd[1]: Reloading... Apr 30 03:34:28.576053 zram_generator::config[3130]: No configuration found. Apr 30 03:34:28.705035 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 03:34:28.806337 systemd[1]: Reloading finished in 313 ms. Apr 30 03:34:28.845856 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:34:28.861408 systemd[1]: kubelet.service: Deactivated successfully. Apr 30 03:34:28.861662 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:34:28.868357 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 03:34:29.092187 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 03:34:29.098348 (kubelet)[3187]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 30 03:34:29.152259 kubelet[3187]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 03:34:29.154026 kubelet[3187]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 30 03:34:29.154026 kubelet[3187]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 03:34:29.154026 kubelet[3187]: I0430 03:34:29.152718 3187 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 30 03:34:29.160756 kubelet[3187]: I0430 03:34:29.160715 3187 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Apr 30 03:34:29.160756 kubelet[3187]: I0430 03:34:29.160746 3187 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 30 03:34:29.161057 kubelet[3187]: I0430 03:34:29.161018 3187 server.go:954] "Client rotation is on, will bootstrap in background" Apr 30 03:34:29.164747 kubelet[3187]: I0430 03:34:29.163853 3187 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Apr 30 03:34:29.170085 kubelet[3187]: I0430 03:34:29.169935 3187 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 30 03:34:29.174981 kubelet[3187]: E0430 03:34:29.174773 3187 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 30 03:34:29.174981 kubelet[3187]: I0430 03:34:29.174985 3187 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 30 03:34:29.178081 kubelet[3187]: I0430 03:34:29.178046 3187 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 30 03:34:29.178256 kubelet[3187]: I0430 03:34:29.178222 3187 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 30 03:34:29.178419 kubelet[3187]: I0430 03:34:29.178254 3187 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-172-31-20-13","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 30 03:34:29.178419 kubelet[3187]: I0430 03:34:29.178418 3187 topology_manager.go:138] "Creating topology manager with none policy" Apr 30 03:34:29.178579 kubelet[3187]: I0430 03:34:29.178429 3187 container_manager_linux.go:304] "Creating device plugin manager" Apr 30 03:34:29.178579 kubelet[3187]: I0430 03:34:29.178464 3187 state_mem.go:36] "Initialized new in-memory state store" Apr 30 03:34:29.181024 kubelet[3187]: I0430 03:34:29.179755 3187 kubelet.go:446] "Attempting to sync node with API server" Apr 30 03:34:29.181024 kubelet[3187]: I0430 03:34:29.179786 3187 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 30 03:34:29.181024 kubelet[3187]: I0430 03:34:29.179805 3187 kubelet.go:352] "Adding apiserver pod source" Apr 30 03:34:29.181024 kubelet[3187]: I0430 03:34:29.179816 3187 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 30 03:34:29.189109 kubelet[3187]: I0430 03:34:29.189085 3187 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Apr 30 03:34:29.193223 kubelet[3187]: I0430 03:34:29.192899 3187 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Apr 30 03:34:29.193497 kubelet[3187]: I0430 03:34:29.193480 3187 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 30 03:34:29.193556 kubelet[3187]: I0430 03:34:29.193513 3187 server.go:1287] "Started kubelet" Apr 30 03:34:29.195819 kubelet[3187]: I0430 03:34:29.195781 3187 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Apr 30 03:34:29.196132 kubelet[3187]: I0430 03:34:29.196076 3187 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 30 03:34:29.198204 kubelet[3187]: I0430 03:34:29.198187 3187 server.go:490] "Adding debug handlers to kubelet server" Apr 30 03:34:29.199339 kubelet[3187]: I0430 03:34:29.199298 3187 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 30 03:34:29.199564 kubelet[3187]: I0430 03:34:29.199553 3187 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 30 03:34:29.200640 kubelet[3187]: I0430 03:34:29.200622 3187 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 30 03:34:29.203171 kubelet[3187]: I0430 03:34:29.203158 3187 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 30 03:34:29.205068 kubelet[3187]: I0430 03:34:29.205050 3187 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Apr 30 03:34:29.206072 kubelet[3187]: I0430 03:34:29.205209 3187 reconciler.go:26] "Reconciler: start to sync state" Apr 30 03:34:29.208623 kubelet[3187]: I0430 03:34:29.207104 3187 factory.go:221] Registration of the systemd container factory successfully Apr 30 03:34:29.208623 kubelet[3187]: I0430 03:34:29.207206 3187 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 30 03:34:29.211504 kubelet[3187]: I0430 03:34:29.211485 3187 factory.go:221] Registration of the containerd container factory successfully Apr 30 03:34:29.216564 kubelet[3187]: I0430 03:34:29.216514 3187 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 30 03:34:29.217824 kubelet[3187]: I0430 03:34:29.217565 3187 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 30 03:34:29.217824 kubelet[3187]: I0430 03:34:29.217591 3187 status_manager.go:227] "Starting to sync pod status with apiserver" Apr 30 03:34:29.217824 kubelet[3187]: I0430 03:34:29.217609 3187 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 30 03:34:29.217824 kubelet[3187]: I0430 03:34:29.217615 3187 kubelet.go:2388] "Starting kubelet main sync loop" Apr 30 03:34:29.217824 kubelet[3187]: E0430 03:34:29.217656 3187 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 30 03:34:29.225864 kubelet[3187]: E0430 03:34:29.225837 3187 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 30 03:34:29.276732 kubelet[3187]: I0430 03:34:29.276703 3187 cpu_manager.go:221] "Starting CPU manager" policy="none" Apr 30 03:34:29.277037 kubelet[3187]: I0430 03:34:29.276878 3187 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Apr 30 03:34:29.277037 kubelet[3187]: I0430 03:34:29.276901 3187 state_mem.go:36] "Initialized new in-memory state store" Apr 30 03:34:29.277178 kubelet[3187]: I0430 03:34:29.277105 3187 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 30 03:34:29.277178 kubelet[3187]: I0430 03:34:29.277119 3187 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 30 03:34:29.277178 kubelet[3187]: I0430 03:34:29.277145 3187 policy_none.go:49] "None policy: Start" Apr 30 03:34:29.277178 kubelet[3187]: I0430 03:34:29.277159 3187 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 30 03:34:29.277178 kubelet[3187]: I0430 03:34:29.277172 3187 state_mem.go:35] "Initializing new in-memory state store" Apr 30 03:34:29.277367 kubelet[3187]: I0430 03:34:29.277316 3187 state_mem.go:75] "Updated machine memory state" Apr 30 03:34:29.281595 kubelet[3187]: I0430 03:34:29.281566 3187 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 30 03:34:29.281771 kubelet[3187]: I0430 03:34:29.281753 3187 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 30 03:34:29.281824 kubelet[3187]: I0430 03:34:29.281769 3187 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 30 03:34:29.285068 kubelet[3187]: I0430 03:34:29.284955 3187 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 30 03:34:29.286443 kubelet[3187]: E0430 03:34:29.285865 3187 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Apr 30 03:34:29.320090 kubelet[3187]: I0430 03:34:29.320057 3187 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-20-13" Apr 30 03:34:29.320385 kubelet[3187]: I0430 03:34:29.320371 3187 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:29.320654 kubelet[3187]: I0430 03:34:29.320397 3187 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:29.389660 kubelet[3187]: I0430 03:34:29.389252 3187 kubelet_node_status.go:76] "Attempting to register node" node="ip-172-31-20-13" Apr 30 03:34:29.396416 kubelet[3187]: I0430 03:34:29.396382 3187 kubelet_node_status.go:125] "Node was previously registered" node="ip-172-31-20-13" Apr 30 03:34:29.396548 kubelet[3187]: I0430 03:34:29.396454 3187 kubelet_node_status.go:79] "Successfully registered node" node="ip-172-31-20-13" Apr 30 03:34:29.507522 kubelet[3187]: I0430 03:34:29.507272 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-kubeconfig\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:29.507522 kubelet[3187]: I0430 03:34:29.507315 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-flexvolume-dir\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:29.507522 kubelet[3187]: I0430 03:34:29.507338 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/739b71698845d25aff1f725b42fc217d-k8s-certs\") pod \"kube-apiserver-ip-172-31-20-13\" (UID: \"739b71698845d25aff1f725b42fc217d\") " pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:29.507522 kubelet[3187]: I0430 03:34:29.507356 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/739b71698845d25aff1f725b42fc217d-usr-share-ca-certificates\") pod \"kube-apiserver-ip-172-31-20-13\" (UID: \"739b71698845d25aff1f725b42fc217d\") " pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:29.507522 kubelet[3187]: I0430 03:34:29.507374 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-ca-certs\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:29.507775 kubelet[3187]: I0430 03:34:29.507390 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-k8s-certs\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:29.507775 kubelet[3187]: I0430 03:34:29.507410 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5daeed804a0bd8056e6c7202a23fb658-usr-share-ca-certificates\") pod \"kube-controller-manager-ip-172-31-20-13\" (UID: \"5daeed804a0bd8056e6c7202a23fb658\") " pod="kube-system/kube-controller-manager-ip-172-31-20-13" Apr 30 03:34:29.507775 kubelet[3187]: I0430 03:34:29.507438 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fb4d70b6b6bcc7a5e8cce843eaf8b6fb-kubeconfig\") pod \"kube-scheduler-ip-172-31-20-13\" (UID: \"fb4d70b6b6bcc7a5e8cce843eaf8b6fb\") " pod="kube-system/kube-scheduler-ip-172-31-20-13" Apr 30 03:34:29.507775 kubelet[3187]: I0430 03:34:29.507455 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/739b71698845d25aff1f725b42fc217d-ca-certs\") pod \"kube-apiserver-ip-172-31-20-13\" (UID: \"739b71698845d25aff1f725b42fc217d\") " pod="kube-system/kube-apiserver-ip-172-31-20-13" Apr 30 03:34:30.185303 kubelet[3187]: I0430 03:34:30.185261 3187 apiserver.go:52] "Watching apiserver" Apr 30 03:34:30.205424 kubelet[3187]: I0430 03:34:30.205384 3187 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Apr 30 03:34:30.242342 kubelet[3187]: I0430 03:34:30.242315 3187 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ip-172-31-20-13" Apr 30 03:34:30.248398 kubelet[3187]: E0430 03:34:30.248329 3187 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ip-172-31-20-13\" already exists" pod="kube-system/kube-scheduler-ip-172-31-20-13" Apr 30 03:34:30.272789 kubelet[3187]: I0430 03:34:30.272430 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ip-172-31-20-13" podStartSLOduration=1.272385221 podStartE2EDuration="1.272385221s" podCreationTimestamp="2025-04-30 03:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 03:34:30.271930797 +0000 UTC m=+1.167218432" watchObservedRunningTime="2025-04-30 03:34:30.272385221 +0000 UTC m=+1.167672845" Apr 30 03:34:30.294138 kubelet[3187]: I0430 03:34:30.293942 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ip-172-31-20-13" podStartSLOduration=1.293920658 podStartE2EDuration="1.293920658s" podCreationTimestamp="2025-04-30 03:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 03:34:30.283081893 +0000 UTC m=+1.178369528" watchObservedRunningTime="2025-04-30 03:34:30.293920658 +0000 UTC m=+1.189208292" Apr 30 03:34:30.311254 kubelet[3187]: I0430 03:34:30.310602 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ip-172-31-20-13" podStartSLOduration=1.31058198 podStartE2EDuration="1.31058198s" podCreationTimestamp="2025-04-30 03:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 03:34:30.294985436 +0000 UTC m=+1.190273075" watchObservedRunningTime="2025-04-30 03:34:30.31058198 +0000 UTC m=+1.205869615" Apr 30 03:34:34.840362 sudo[2309]: pam_unix(sudo:session): session closed for user root Apr 30 03:34:34.877369 sshd[2306]: pam_unix(sshd:session): session closed for user core Apr 30 03:34:34.881115 systemd[1]: sshd@6-172.31.20.13:22-147.75.109.163:53570.service: Deactivated successfully. Apr 30 03:34:34.883871 systemd[1]: session-7.scope: Deactivated successfully. Apr 30 03:34:34.884156 systemd[1]: session-7.scope: Consumed 4.836s CPU time, 141.9M memory peak, 0B memory swap peak. Apr 30 03:34:34.884659 systemd-logind[1970]: Session 7 logged out. Waiting for processes to exit. Apr 30 03:34:34.885815 systemd-logind[1970]: Removed session 7. Apr 30 03:34:36.219963 kubelet[3187]: I0430 03:34:36.219926 3187 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 30 03:34:36.221830 containerd[1992]: time="2025-04-30T03:34:36.221781184Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 30 03:34:36.222847 kubelet[3187]: I0430 03:34:36.222275 3187 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 30 03:34:37.195374 systemd[1]: Created slice kubepods-besteffort-podccd8b844_7a81_4d0f_8402_b2f92a811b6a.slice - libcontainer container kubepods-besteffort-podccd8b844_7a81_4d0f_8402_b2f92a811b6a.slice. Apr 30 03:34:37.266721 kubelet[3187]: I0430 03:34:37.266539 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/ccd8b844-7a81-4d0f-8402-b2f92a811b6a-kube-proxy\") pod \"kube-proxy-f58f2\" (UID: \"ccd8b844-7a81-4d0f-8402-b2f92a811b6a\") " pod="kube-system/kube-proxy-f58f2" Apr 30 03:34:37.266721 kubelet[3187]: I0430 03:34:37.266587 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ccd8b844-7a81-4d0f-8402-b2f92a811b6a-lib-modules\") pod \"kube-proxy-f58f2\" (UID: \"ccd8b844-7a81-4d0f-8402-b2f92a811b6a\") " pod="kube-system/kube-proxy-f58f2" Apr 30 03:34:37.266721 kubelet[3187]: I0430 03:34:37.266613 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh5g7\" (UniqueName: \"kubernetes.io/projected/ccd8b844-7a81-4d0f-8402-b2f92a811b6a-kube-api-access-lh5g7\") pod \"kube-proxy-f58f2\" (UID: \"ccd8b844-7a81-4d0f-8402-b2f92a811b6a\") " pod="kube-system/kube-proxy-f58f2" Apr 30 03:34:37.266721 kubelet[3187]: I0430 03:34:37.266645 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ccd8b844-7a81-4d0f-8402-b2f92a811b6a-xtables-lock\") pod \"kube-proxy-f58f2\" (UID: \"ccd8b844-7a81-4d0f-8402-b2f92a811b6a\") " pod="kube-system/kube-proxy-f58f2" Apr 30 03:34:37.328258 systemd[1]: Created slice kubepods-besteffort-pod46573595_c546_48ad_8190_0c3dcbc3160f.slice - libcontainer container kubepods-besteffort-pod46573595_c546_48ad_8190_0c3dcbc3160f.slice. Apr 30 03:34:37.366967 kubelet[3187]: I0430 03:34:37.366922 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/46573595-c546-48ad-8190-0c3dcbc3160f-var-lib-calico\") pod \"tigera-operator-789496d6f5-mgw6w\" (UID: \"46573595-c546-48ad-8190-0c3dcbc3160f\") " pod="tigera-operator/tigera-operator-789496d6f5-mgw6w" Apr 30 03:34:37.367099 kubelet[3187]: I0430 03:34:37.366986 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsx68\" (UniqueName: \"kubernetes.io/projected/46573595-c546-48ad-8190-0c3dcbc3160f-kube-api-access-jsx68\") pod \"tigera-operator-789496d6f5-mgw6w\" (UID: \"46573595-c546-48ad-8190-0c3dcbc3160f\") " pod="tigera-operator/tigera-operator-789496d6f5-mgw6w" Apr 30 03:34:37.504981 containerd[1992]: time="2025-04-30T03:34:37.504863187Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-f58f2,Uid:ccd8b844-7a81-4d0f-8402-b2f92a811b6a,Namespace:kube-system,Attempt:0,}" Apr 30 03:34:37.542443 containerd[1992]: time="2025-04-30T03:34:37.541953761Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:34:37.542443 containerd[1992]: time="2025-04-30T03:34:37.542039333Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:34:37.542443 containerd[1992]: time="2025-04-30T03:34:37.542063310Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:37.542948 containerd[1992]: time="2025-04-30T03:34:37.542179252Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:37.573231 systemd[1]: Started cri-containerd-173cad8a831468524f47d6e1ae29f21f58dc96abb81d6c3675319ac74e01de6f.scope - libcontainer container 173cad8a831468524f47d6e1ae29f21f58dc96abb81d6c3675319ac74e01de6f. Apr 30 03:34:37.602114 containerd[1992]: time="2025-04-30T03:34:37.601729627Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-f58f2,Uid:ccd8b844-7a81-4d0f-8402-b2f92a811b6a,Namespace:kube-system,Attempt:0,} returns sandbox id \"173cad8a831468524f47d6e1ae29f21f58dc96abb81d6c3675319ac74e01de6f\"" Apr 30 03:34:37.606791 containerd[1992]: time="2025-04-30T03:34:37.606598895Z" level=info msg="CreateContainer within sandbox \"173cad8a831468524f47d6e1ae29f21f58dc96abb81d6c3675319ac74e01de6f\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 30 03:34:37.634244 containerd[1992]: time="2025-04-30T03:34:37.634188244Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-mgw6w,Uid:46573595-c546-48ad-8190-0c3dcbc3160f,Namespace:tigera-operator,Attempt:0,}" Apr 30 03:34:37.634811 containerd[1992]: time="2025-04-30T03:34:37.634672084Z" level=info msg="CreateContainer within sandbox \"173cad8a831468524f47d6e1ae29f21f58dc96abb81d6c3675319ac74e01de6f\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4e36ace63690fb8eb48c89617cf5061dd1c5aaa0e22e248356fa67964da122ef\"" Apr 30 03:34:37.635583 containerd[1992]: time="2025-04-30T03:34:37.635551428Z" level=info msg="StartContainer for \"4e36ace63690fb8eb48c89617cf5061dd1c5aaa0e22e248356fa67964da122ef\"" Apr 30 03:34:37.669479 systemd[1]: Started cri-containerd-4e36ace63690fb8eb48c89617cf5061dd1c5aaa0e22e248356fa67964da122ef.scope - libcontainer container 4e36ace63690fb8eb48c89617cf5061dd1c5aaa0e22e248356fa67964da122ef. Apr 30 03:34:37.698813 containerd[1992]: time="2025-04-30T03:34:37.698289851Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:34:37.698813 containerd[1992]: time="2025-04-30T03:34:37.698394346Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:34:37.698813 containerd[1992]: time="2025-04-30T03:34:37.698415523Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:37.698813 containerd[1992]: time="2025-04-30T03:34:37.698560402Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:37.727812 systemd[1]: Started cri-containerd-32d428b1833b4814a7e6947cd9800fcdd9ee500c575445f8f20fb867063e9f30.scope - libcontainer container 32d428b1833b4814a7e6947cd9800fcdd9ee500c575445f8f20fb867063e9f30. Apr 30 03:34:37.728881 containerd[1992]: time="2025-04-30T03:34:37.727976466Z" level=info msg="StartContainer for \"4e36ace63690fb8eb48c89617cf5061dd1c5aaa0e22e248356fa67964da122ef\" returns successfully" Apr 30 03:34:37.781292 containerd[1992]: time="2025-04-30T03:34:37.781234188Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-mgw6w,Uid:46573595-c546-48ad-8190-0c3dcbc3160f,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"32d428b1833b4814a7e6947cd9800fcdd9ee500c575445f8f20fb867063e9f30\"" Apr 30 03:34:37.784258 containerd[1992]: time="2025-04-30T03:34:37.783980292Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" Apr 30 03:34:39.054378 kubelet[3187]: I0430 03:34:39.054300 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-f58f2" podStartSLOduration=2.054262337 podStartE2EDuration="2.054262337s" podCreationTimestamp="2025-04-30 03:34:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 03:34:38.289961457 +0000 UTC m=+9.185249092" watchObservedRunningTime="2025-04-30 03:34:39.054262337 +0000 UTC m=+9.949549968" Apr 30 03:34:39.431400 update_engine[1972]: I20250430 03:34:39.431228 1972 update_attempter.cc:509] Updating boot flags... Apr 30 03:34:39.491111 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (3527) Apr 30 03:34:39.726844 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 38 scanned by (udev-worker) (3531) Apr 30 03:34:39.872351 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3014937905.mount: Deactivated successfully. Apr 30 03:34:41.008629 containerd[1992]: time="2025-04-30T03:34:41.008573151Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:41.009659 containerd[1992]: time="2025-04-30T03:34:41.009600283Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.7: active requests=0, bytes read=22002662" Apr 30 03:34:41.010979 containerd[1992]: time="2025-04-30T03:34:41.010925739Z" level=info msg="ImageCreate event name:\"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:41.013511 containerd[1992]: time="2025-04-30T03:34:41.013481415Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:41.014311 containerd[1992]: time="2025-04-30T03:34:41.014188558Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.7\" with image id \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\", repo tag \"quay.io/tigera/operator:v1.36.7\", repo digest \"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\", size \"21998657\" in 3.230078175s" Apr 30 03:34:41.014311 containerd[1992]: time="2025-04-30T03:34:41.014219597Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\"" Apr 30 03:34:41.071075 containerd[1992]: time="2025-04-30T03:34:41.071038647Z" level=info msg="CreateContainer within sandbox \"32d428b1833b4814a7e6947cd9800fcdd9ee500c575445f8f20fb867063e9f30\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Apr 30 03:34:41.086690 containerd[1992]: time="2025-04-30T03:34:41.086571538Z" level=info msg="CreateContainer within sandbox \"32d428b1833b4814a7e6947cd9800fcdd9ee500c575445f8f20fb867063e9f30\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d\"" Apr 30 03:34:41.087473 containerd[1992]: time="2025-04-30T03:34:41.087437915Z" level=info msg="StartContainer for \"a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d\"" Apr 30 03:34:41.113255 systemd[1]: run-containerd-runc-k8s.io-a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d-runc.AorbXB.mount: Deactivated successfully. Apr 30 03:34:41.123258 systemd[1]: Started cri-containerd-a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d.scope - libcontainer container a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d. Apr 30 03:34:41.160738 containerd[1992]: time="2025-04-30T03:34:41.160691298Z" level=info msg="StartContainer for \"a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d\" returns successfully" Apr 30 03:34:41.317073 kubelet[3187]: I0430 03:34:41.316994 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-789496d6f5-mgw6w" podStartSLOduration=1.057254599 podStartE2EDuration="4.31696093s" podCreationTimestamp="2025-04-30 03:34:37 +0000 UTC" firstStartedPulling="2025-04-30 03:34:37.783173918 +0000 UTC m=+8.678461531" lastFinishedPulling="2025-04-30 03:34:41.042880238 +0000 UTC m=+11.938167862" observedRunningTime="2025-04-30 03:34:41.316944635 +0000 UTC m=+12.212232270" watchObservedRunningTime="2025-04-30 03:34:41.31696093 +0000 UTC m=+12.212248564" Apr 30 03:34:44.322359 systemd[1]: Created slice kubepods-besteffort-pod101b019f_65a8_44d3_bd99_75128481c13d.slice - libcontainer container kubepods-besteffort-pod101b019f_65a8_44d3_bd99_75128481c13d.slice. Apr 30 03:34:44.381649 kubelet[3187]: I0430 03:34:44.381510 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/101b019f-65a8-44d3-bd99-75128481c13d-typha-certs\") pod \"calico-typha-7795dbccc-k6r2f\" (UID: \"101b019f-65a8-44d3-bd99-75128481c13d\") " pod="calico-system/calico-typha-7795dbccc-k6r2f" Apr 30 03:34:44.381649 kubelet[3187]: I0430 03:34:44.381550 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/101b019f-65a8-44d3-bd99-75128481c13d-tigera-ca-bundle\") pod \"calico-typha-7795dbccc-k6r2f\" (UID: \"101b019f-65a8-44d3-bd99-75128481c13d\") " pod="calico-system/calico-typha-7795dbccc-k6r2f" Apr 30 03:34:44.381649 kubelet[3187]: I0430 03:34:44.381571 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp9j9\" (UniqueName: \"kubernetes.io/projected/101b019f-65a8-44d3-bd99-75128481c13d-kube-api-access-wp9j9\") pod \"calico-typha-7795dbccc-k6r2f\" (UID: \"101b019f-65a8-44d3-bd99-75128481c13d\") " pod="calico-system/calico-typha-7795dbccc-k6r2f" Apr 30 03:34:44.455280 systemd[1]: Created slice kubepods-besteffort-pod095b08fe_1393_4535_91ee_a4799ab77084.slice - libcontainer container kubepods-besteffort-pod095b08fe_1393_4535_91ee_a4799ab77084.slice. Apr 30 03:34:44.584147 kubelet[3187]: I0430 03:34:44.583754 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-policysync\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584147 kubelet[3187]: I0430 03:34:44.583797 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-cni-bin-dir\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584147 kubelet[3187]: I0430 03:34:44.583822 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-flexvol-driver-host\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584147 kubelet[3187]: I0430 03:34:44.583852 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-lib-modules\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584147 kubelet[3187]: I0430 03:34:44.583876 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-var-run-calico\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584447 kubelet[3187]: I0430 03:34:44.583902 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hztw9\" (UniqueName: \"kubernetes.io/projected/095b08fe-1393-4535-91ee-a4799ab77084-kube-api-access-hztw9\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584447 kubelet[3187]: I0430 03:34:44.583923 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-xtables-lock\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584447 kubelet[3187]: I0430 03:34:44.583959 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-var-lib-calico\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584447 kubelet[3187]: I0430 03:34:44.583982 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/095b08fe-1393-4535-91ee-a4799ab77084-tigera-ca-bundle\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584447 kubelet[3187]: I0430 03:34:44.584035 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-cni-log-dir\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584680 kubelet[3187]: I0430 03:34:44.584065 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/095b08fe-1393-4535-91ee-a4799ab77084-node-certs\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.584680 kubelet[3187]: I0430 03:34:44.584086 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/095b08fe-1393-4535-91ee-a4799ab77084-cni-net-dir\") pod \"calico-node-nql5l\" (UID: \"095b08fe-1393-4535-91ee-a4799ab77084\") " pod="calico-system/calico-node-nql5l" Apr 30 03:34:44.616608 kubelet[3187]: E0430 03:34:44.615940 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-p2dkq" podUID="5e820db8-1ee7-4b3a-bfb7-17c5a434fca3" Apr 30 03:34:44.633095 containerd[1992]: time="2025-04-30T03:34:44.633039075Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7795dbccc-k6r2f,Uid:101b019f-65a8-44d3-bd99-75128481c13d,Namespace:calico-system,Attempt:0,}" Apr 30 03:34:44.689026 kubelet[3187]: I0430 03:34:44.685241 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/5e820db8-1ee7-4b3a-bfb7-17c5a434fca3-varrun\") pod \"csi-node-driver-p2dkq\" (UID: \"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3\") " pod="calico-system/csi-node-driver-p2dkq" Apr 30 03:34:44.689026 kubelet[3187]: I0430 03:34:44.685302 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e820db8-1ee7-4b3a-bfb7-17c5a434fca3-kubelet-dir\") pod \"csi-node-driver-p2dkq\" (UID: \"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3\") " pod="calico-system/csi-node-driver-p2dkq" Apr 30 03:34:44.689026 kubelet[3187]: I0430 03:34:44.685377 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e820db8-1ee7-4b3a-bfb7-17c5a434fca3-socket-dir\") pod \"csi-node-driver-p2dkq\" (UID: \"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3\") " pod="calico-system/csi-node-driver-p2dkq" Apr 30 03:34:44.689026 kubelet[3187]: I0430 03:34:44.685424 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e820db8-1ee7-4b3a-bfb7-17c5a434fca3-registration-dir\") pod \"csi-node-driver-p2dkq\" (UID: \"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3\") " pod="calico-system/csi-node-driver-p2dkq" Apr 30 03:34:44.689026 kubelet[3187]: I0430 03:34:44.685479 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vgm2\" (UniqueName: \"kubernetes.io/projected/5e820db8-1ee7-4b3a-bfb7-17c5a434fca3-kube-api-access-8vgm2\") pod \"csi-node-driver-p2dkq\" (UID: \"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3\") " pod="calico-system/csi-node-driver-p2dkq" Apr 30 03:34:44.689263 containerd[1992]: time="2025-04-30T03:34:44.684769176Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:34:44.689263 containerd[1992]: time="2025-04-30T03:34:44.684818824Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:34:44.689263 containerd[1992]: time="2025-04-30T03:34:44.684829364Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:44.689263 containerd[1992]: time="2025-04-30T03:34:44.684903886Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:44.705838 kubelet[3187]: E0430 03:34:44.705808 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.706033 kubelet[3187]: W0430 03:34:44.706015 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.706353 kubelet[3187]: E0430 03:34:44.706338 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.717131 kubelet[3187]: E0430 03:34:44.717042 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.717131 kubelet[3187]: W0430 03:34:44.717066 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.717131 kubelet[3187]: E0430 03:34:44.717087 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.733213 systemd[1]: Started cri-containerd-09bd49e9fb92ba648fcd597775784b58fffe4e0a2b4286cd7295f0f1933c1934.scope - libcontainer container 09bd49e9fb92ba648fcd597775784b58fffe4e0a2b4286cd7295f0f1933c1934. Apr 30 03:34:44.764304 containerd[1992]: time="2025-04-30T03:34:44.764170317Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-nql5l,Uid:095b08fe-1393-4535-91ee-a4799ab77084,Namespace:calico-system,Attempt:0,}" Apr 30 03:34:44.787587 kubelet[3187]: E0430 03:34:44.787538 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.787587 kubelet[3187]: W0430 03:34:44.787569 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.787587 kubelet[3187]: E0430 03:34:44.787594 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.788014 kubelet[3187]: E0430 03:34:44.787919 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.788014 kubelet[3187]: W0430 03:34:44.787933 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.788014 kubelet[3187]: E0430 03:34:44.787958 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.788895 kubelet[3187]: E0430 03:34:44.788268 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.788895 kubelet[3187]: W0430 03:34:44.788279 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.788895 kubelet[3187]: E0430 03:34:44.788304 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.788895 kubelet[3187]: E0430 03:34:44.788593 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.788895 kubelet[3187]: W0430 03:34:44.788605 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.788895 kubelet[3187]: E0430 03:34:44.788694 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.789909 kubelet[3187]: E0430 03:34:44.789885 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.789909 kubelet[3187]: W0430 03:34:44.789906 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.790071 kubelet[3187]: E0430 03:34:44.789992 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.790455 kubelet[3187]: E0430 03:34:44.790437 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.790455 kubelet[3187]: W0430 03:34:44.790454 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.790741 kubelet[3187]: E0430 03:34:44.790653 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.791051 kubelet[3187]: E0430 03:34:44.790848 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.791051 kubelet[3187]: W0430 03:34:44.790858 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.791051 kubelet[3187]: E0430 03:34:44.791038 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.791513 kubelet[3187]: E0430 03:34:44.791384 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.792053 kubelet[3187]: W0430 03:34:44.791580 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.792053 kubelet[3187]: E0430 03:34:44.791831 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.794042 kubelet[3187]: E0430 03:34:44.793470 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.794042 kubelet[3187]: W0430 03:34:44.793486 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.794042 kubelet[3187]: E0430 03:34:44.793741 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.794042 kubelet[3187]: E0430 03:34:44.793948 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.794042 kubelet[3187]: W0430 03:34:44.793959 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.794042 kubelet[3187]: E0430 03:34:44.794018 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.796426 kubelet[3187]: E0430 03:34:44.796407 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.796426 kubelet[3187]: W0430 03:34:44.796427 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.796620 kubelet[3187]: E0430 03:34:44.796526 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.796803 kubelet[3187]: E0430 03:34:44.796690 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.796803 kubelet[3187]: W0430 03:34:44.796703 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.796932 kubelet[3187]: E0430 03:34:44.796821 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.797524 kubelet[3187]: E0430 03:34:44.797506 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.797524 kubelet[3187]: W0430 03:34:44.797525 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.797826 kubelet[3187]: E0430 03:34:44.797809 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.799956 kubelet[3187]: E0430 03:34:44.799937 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.800051 kubelet[3187]: W0430 03:34:44.799966 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.800653 kubelet[3187]: E0430 03:34:44.800555 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.800726 kubelet[3187]: W0430 03:34:44.800713 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.802060 kubelet[3187]: E0430 03:34:44.801398 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.802060 kubelet[3187]: W0430 03:34:44.801414 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.802183 kubelet[3187]: E0430 03:34:44.802173 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.802325 kubelet[3187]: W0430 03:34:44.802294 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.802838 kubelet[3187]: E0430 03:34:44.802819 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.802838 kubelet[3187]: W0430 03:34:44.802836 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.802978 kubelet[3187]: E0430 03:34:44.802960 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.803169 kubelet[3187]: E0430 03:34:44.803154 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.803218 kubelet[3187]: E0430 03:34:44.803175 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.803218 kubelet[3187]: E0430 03:34:44.803188 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.803358 kubelet[3187]: E0430 03:34:44.803343 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.803848 kubelet[3187]: E0430 03:34:44.803771 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.803911 kubelet[3187]: W0430 03:34:44.803852 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.804032 kubelet[3187]: E0430 03:34:44.804015 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.804560 kubelet[3187]: E0430 03:34:44.804472 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.804560 kubelet[3187]: W0430 03:34:44.804487 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.804560 kubelet[3187]: E0430 03:34:44.804510 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.805252 kubelet[3187]: E0430 03:34:44.805201 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.805252 kubelet[3187]: W0430 03:34:44.805224 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.805252 kubelet[3187]: E0430 03:34:44.805242 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.806151 kubelet[3187]: E0430 03:34:44.806131 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.806151 kubelet[3187]: W0430 03:34:44.806148 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.806274 kubelet[3187]: E0430 03:34:44.806163 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.806795 kubelet[3187]: E0430 03:34:44.806777 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.806795 kubelet[3187]: W0430 03:34:44.806794 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.806919 kubelet[3187]: E0430 03:34:44.806845 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.807607 kubelet[3187]: E0430 03:34:44.807487 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.807607 kubelet[3187]: W0430 03:34:44.807607 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.807745 kubelet[3187]: E0430 03:34:44.807689 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.809176 kubelet[3187]: E0430 03:34:44.809154 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.812468 kubelet[3187]: W0430 03:34:44.812366 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.812468 kubelet[3187]: E0430 03:34:44.812410 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.823461 containerd[1992]: time="2025-04-30T03:34:44.823250026Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:34:44.823913 containerd[1992]: time="2025-04-30T03:34:44.823387662Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:34:44.823913 containerd[1992]: time="2025-04-30T03:34:44.823408182Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:44.824370 containerd[1992]: time="2025-04-30T03:34:44.824106413Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:34:44.829863 kubelet[3187]: E0430 03:34:44.829652 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:44.829863 kubelet[3187]: W0430 03:34:44.829681 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:44.829863 kubelet[3187]: E0430 03:34:44.829704 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:44.861194 containerd[1992]: time="2025-04-30T03:34:44.860661687Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7795dbccc-k6r2f,Uid:101b019f-65a8-44d3-bd99-75128481c13d,Namespace:calico-system,Attempt:0,} returns sandbox id \"09bd49e9fb92ba648fcd597775784b58fffe4e0a2b4286cd7295f0f1933c1934\"" Apr 30 03:34:44.868687 containerd[1992]: time="2025-04-30T03:34:44.867437534Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" Apr 30 03:34:44.876858 systemd[1]: Started cri-containerd-c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e.scope - libcontainer container c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e. Apr 30 03:34:44.938319 containerd[1992]: time="2025-04-30T03:34:44.938210694Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-nql5l,Uid:095b08fe-1393-4535-91ee-a4799ab77084,Namespace:calico-system,Attempt:0,} returns sandbox id \"c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e\"" Apr 30 03:34:46.218350 kubelet[3187]: E0430 03:34:46.218290 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-p2dkq" podUID="5e820db8-1ee7-4b3a-bfb7-17c5a434fca3" Apr 30 03:34:46.914276 containerd[1992]: time="2025-04-30T03:34:46.914207009Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:46.915494 containerd[1992]: time="2025-04-30T03:34:46.915339009Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.3: active requests=0, bytes read=30426870" Apr 30 03:34:46.917523 containerd[1992]: time="2025-04-30T03:34:46.916499417Z" level=info msg="ImageCreate event name:\"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:46.919032 containerd[1992]: time="2025-04-30T03:34:46.918314535Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:46.919178 containerd[1992]: time="2025-04-30T03:34:46.919140779Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.3\" with image id \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\", size \"31919484\" in 2.050728471s" Apr 30 03:34:46.919247 containerd[1992]: time="2025-04-30T03:34:46.919235996Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\"" Apr 30 03:34:46.920402 containerd[1992]: time="2025-04-30T03:34:46.920385687Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" Apr 30 03:34:46.933600 containerd[1992]: time="2025-04-30T03:34:46.933561609Z" level=info msg="CreateContainer within sandbox \"09bd49e9fb92ba648fcd597775784b58fffe4e0a2b4286cd7295f0f1933c1934\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Apr 30 03:34:46.948936 containerd[1992]: time="2025-04-30T03:34:46.948894889Z" level=info msg="CreateContainer within sandbox \"09bd49e9fb92ba648fcd597775784b58fffe4e0a2b4286cd7295f0f1933c1934\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"196cb4cb71b52d90fe83d94a853611a76a2346629c796882402e00936fcc2556\"" Apr 30 03:34:46.949558 containerd[1992]: time="2025-04-30T03:34:46.949464195Z" level=info msg="StartContainer for \"196cb4cb71b52d90fe83d94a853611a76a2346629c796882402e00936fcc2556\"" Apr 30 03:34:47.013187 systemd[1]: Started cri-containerd-196cb4cb71b52d90fe83d94a853611a76a2346629c796882402e00936fcc2556.scope - libcontainer container 196cb4cb71b52d90fe83d94a853611a76a2346629c796882402e00936fcc2556. Apr 30 03:34:47.056427 containerd[1992]: time="2025-04-30T03:34:47.056365198Z" level=info msg="StartContainer for \"196cb4cb71b52d90fe83d94a853611a76a2346629c796882402e00936fcc2556\" returns successfully" Apr 30 03:34:47.336928 kubelet[3187]: I0430 03:34:47.336837 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7795dbccc-k6r2f" podStartSLOduration=1.281697753 podStartE2EDuration="3.336818407s" podCreationTimestamp="2025-04-30 03:34:44 +0000 UTC" firstStartedPulling="2025-04-30 03:34:44.864870654 +0000 UTC m=+15.760158270" lastFinishedPulling="2025-04-30 03:34:46.919991309 +0000 UTC m=+17.815278924" observedRunningTime="2025-04-30 03:34:47.336282373 +0000 UTC m=+18.231570008" watchObservedRunningTime="2025-04-30 03:34:47.336818407 +0000 UTC m=+18.232106041" Apr 30 03:34:47.351077 kubelet[3187]: E0430 03:34:47.351042 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.351077 kubelet[3187]: W0430 03:34:47.351070 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.351298 kubelet[3187]: E0430 03:34:47.351115 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.351460 kubelet[3187]: E0430 03:34:47.351437 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.351460 kubelet[3187]: W0430 03:34:47.351456 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.351651 kubelet[3187]: E0430 03:34:47.351474 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.351738 kubelet[3187]: E0430 03:34:47.351726 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.351738 kubelet[3187]: W0430 03:34:47.351736 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.351884 kubelet[3187]: E0430 03:34:47.351751 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.352097 kubelet[3187]: E0430 03:34:47.352051 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.352097 kubelet[3187]: W0430 03:34:47.352064 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.352097 kubelet[3187]: E0430 03:34:47.352077 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.352370 kubelet[3187]: E0430 03:34:47.352349 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.352370 kubelet[3187]: W0430 03:34:47.352365 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.352521 kubelet[3187]: E0430 03:34:47.352382 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.352647 kubelet[3187]: E0430 03:34:47.352620 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.352647 kubelet[3187]: W0430 03:34:47.352634 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.352829 kubelet[3187]: E0430 03:34:47.352647 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.352899 kubelet[3187]: E0430 03:34:47.352851 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.352899 kubelet[3187]: W0430 03:34:47.352862 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.352899 kubelet[3187]: E0430 03:34:47.352874 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.353139 kubelet[3187]: E0430 03:34:47.353099 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.353139 kubelet[3187]: W0430 03:34:47.353110 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.353139 kubelet[3187]: E0430 03:34:47.353122 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.353376 kubelet[3187]: E0430 03:34:47.353346 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.353376 kubelet[3187]: W0430 03:34:47.353360 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.353376 kubelet[3187]: E0430 03:34:47.353373 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.353613 kubelet[3187]: E0430 03:34:47.353579 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.353613 kubelet[3187]: W0430 03:34:47.353594 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.353613 kubelet[3187]: E0430 03:34:47.353607 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.353862 kubelet[3187]: E0430 03:34:47.353843 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.353862 kubelet[3187]: W0430 03:34:47.353858 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.353977 kubelet[3187]: E0430 03:34:47.353871 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.354141 kubelet[3187]: E0430 03:34:47.354125 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.354141 kubelet[3187]: W0430 03:34:47.354138 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.354273 kubelet[3187]: E0430 03:34:47.354151 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.354396 kubelet[3187]: E0430 03:34:47.354379 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.354396 kubelet[3187]: W0430 03:34:47.354392 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.354725 kubelet[3187]: E0430 03:34:47.354404 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.354725 kubelet[3187]: E0430 03:34:47.354640 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.354915 kubelet[3187]: W0430 03:34:47.354652 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.354915 kubelet[3187]: E0430 03:34:47.354795 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.355156 kubelet[3187]: E0430 03:34:47.355138 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.355156 kubelet[3187]: W0430 03:34:47.355152 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.355267 kubelet[3187]: E0430 03:34:47.355166 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.408071 kubelet[3187]: E0430 03:34:47.408039 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.408071 kubelet[3187]: W0430 03:34:47.408061 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.408071 kubelet[3187]: E0430 03:34:47.408081 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.408426 kubelet[3187]: E0430 03:34:47.408402 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.408426 kubelet[3187]: W0430 03:34:47.408420 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.408550 kubelet[3187]: E0430 03:34:47.408439 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.408717 kubelet[3187]: E0430 03:34:47.408704 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.408717 kubelet[3187]: W0430 03:34:47.408716 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.408782 kubelet[3187]: E0430 03:34:47.408736 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.408988 kubelet[3187]: E0430 03:34:47.408974 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.408988 kubelet[3187]: W0430 03:34:47.408984 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.409091 kubelet[3187]: E0430 03:34:47.409075 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.409275 kubelet[3187]: E0430 03:34:47.409251 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.409275 kubelet[3187]: W0430 03:34:47.409270 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.409383 kubelet[3187]: E0430 03:34:47.409295 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.409628 kubelet[3187]: E0430 03:34:47.409528 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.409628 kubelet[3187]: W0430 03:34:47.409539 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.409628 kubelet[3187]: E0430 03:34:47.409556 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.409763 kubelet[3187]: E0430 03:34:47.409755 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.409816 kubelet[3187]: W0430 03:34:47.409799 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.409894 kubelet[3187]: E0430 03:34:47.409879 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.410191 kubelet[3187]: E0430 03:34:47.410045 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.410191 kubelet[3187]: W0430 03:34:47.410062 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.410304 kubelet[3187]: E0430 03:34:47.410225 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.410304 kubelet[3187]: E0430 03:34:47.410236 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.410304 kubelet[3187]: W0430 03:34:47.410243 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.410304 kubelet[3187]: E0430 03:34:47.410258 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.410447 kubelet[3187]: E0430 03:34:47.410409 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.410447 kubelet[3187]: W0430 03:34:47.410416 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.410447 kubelet[3187]: E0430 03:34:47.410423 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.410800 kubelet[3187]: E0430 03:34:47.410782 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.410909 kubelet[3187]: W0430 03:34:47.410873 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.410968 kubelet[3187]: E0430 03:34:47.410918 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.411495 kubelet[3187]: E0430 03:34:47.411479 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.411495 kubelet[3187]: W0430 03:34:47.411492 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.411571 kubelet[3187]: E0430 03:34:47.411506 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.411843 kubelet[3187]: E0430 03:34:47.411829 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.411843 kubelet[3187]: W0430 03:34:47.411840 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.411843 kubelet[3187]: E0430 03:34:47.411913 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.412089 kubelet[3187]: E0430 03:34:47.412074 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.412089 kubelet[3187]: W0430 03:34:47.412085 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.412161 kubelet[3187]: E0430 03:34:47.412147 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.412336 kubelet[3187]: E0430 03:34:47.412279 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.412336 kubelet[3187]: W0430 03:34:47.412292 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.412336 kubelet[3187]: E0430 03:34:47.412310 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.412533 kubelet[3187]: E0430 03:34:47.412516 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.412533 kubelet[3187]: W0430 03:34:47.412528 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.412533 kubelet[3187]: E0430 03:34:47.412537 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.412778 kubelet[3187]: E0430 03:34:47.412759 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.412778 kubelet[3187]: W0430 03:34:47.412775 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.412851 kubelet[3187]: E0430 03:34:47.412787 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.413183 kubelet[3187]: E0430 03:34:47.413166 3187 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 03:34:47.413183 kubelet[3187]: W0430 03:34:47.413179 3187 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 03:34:47.413275 kubelet[3187]: E0430 03:34:47.413189 3187 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 03:34:47.927260 systemd[1]: run-containerd-runc-k8s.io-196cb4cb71b52d90fe83d94a853611a76a2346629c796882402e00936fcc2556-runc.mBahI6.mount: Deactivated successfully. Apr 30 03:34:48.086775 containerd[1992]: time="2025-04-30T03:34:48.086601514Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:48.087737 containerd[1992]: time="2025-04-30T03:34:48.087685082Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3: active requests=0, bytes read=5366937" Apr 30 03:34:48.090140 containerd[1992]: time="2025-04-30T03:34:48.088989890Z" level=info msg="ImageCreate event name:\"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:48.091112 containerd[1992]: time="2025-04-30T03:34:48.091074609Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:48.092340 containerd[1992]: time="2025-04-30T03:34:48.091866120Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" with image id \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\", size \"6859519\" in 1.171330983s" Apr 30 03:34:48.092340 containerd[1992]: time="2025-04-30T03:34:48.091901048Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\"" Apr 30 03:34:48.093479 containerd[1992]: time="2025-04-30T03:34:48.093453341Z" level=info msg="CreateContainer within sandbox \"c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Apr 30 03:34:48.109781 containerd[1992]: time="2025-04-30T03:34:48.109736709Z" level=info msg="CreateContainer within sandbox \"c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34\"" Apr 30 03:34:48.110803 containerd[1992]: time="2025-04-30T03:34:48.110608580Z" level=info msg="StartContainer for \"bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34\"" Apr 30 03:34:48.141182 systemd[1]: run-containerd-runc-k8s.io-bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34-runc.Xx4TeI.mount: Deactivated successfully. Apr 30 03:34:48.151244 systemd[1]: Started cri-containerd-bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34.scope - libcontainer container bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34. Apr 30 03:34:48.181394 containerd[1992]: time="2025-04-30T03:34:48.181286446Z" level=info msg="StartContainer for \"bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34\" returns successfully" Apr 30 03:34:48.202979 systemd[1]: cri-containerd-bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34.scope: Deactivated successfully. Apr 30 03:34:48.219040 kubelet[3187]: E0430 03:34:48.218725 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-p2dkq" podUID="5e820db8-1ee7-4b3a-bfb7-17c5a434fca3" Apr 30 03:34:48.331875 kubelet[3187]: I0430 03:34:48.331130 3187 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 03:34:48.385601 containerd[1992]: time="2025-04-30T03:34:48.315795376Z" level=info msg="shim disconnected" id=bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34 namespace=k8s.io Apr 30 03:34:48.385601 containerd[1992]: time="2025-04-30T03:34:48.385600323Z" level=warning msg="cleaning up after shim disconnected" id=bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34 namespace=k8s.io Apr 30 03:34:48.385824 containerd[1992]: time="2025-04-30T03:34:48.385620172Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 03:34:48.928460 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bc4a11c0b32493961a3119320806fb43c0bb1cc0ad653130569c27a0871d2b34-rootfs.mount: Deactivated successfully. Apr 30 03:34:49.336721 containerd[1992]: time="2025-04-30T03:34:49.336457851Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" Apr 30 03:34:50.218473 kubelet[3187]: E0430 03:34:50.218121 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-p2dkq" podUID="5e820db8-1ee7-4b3a-bfb7-17c5a434fca3" Apr 30 03:34:52.218018 kubelet[3187]: E0430 03:34:52.217920 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-p2dkq" podUID="5e820db8-1ee7-4b3a-bfb7-17c5a434fca3" Apr 30 03:34:53.040760 containerd[1992]: time="2025-04-30T03:34:53.040709356Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:53.041826 containerd[1992]: time="2025-04-30T03:34:53.041701915Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.3: active requests=0, bytes read=97793683" Apr 30 03:34:53.044034 containerd[1992]: time="2025-04-30T03:34:53.042712589Z" level=info msg="ImageCreate event name:\"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:53.045690 containerd[1992]: time="2025-04-30T03:34:53.045050904Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:34:53.045690 containerd[1992]: time="2025-04-30T03:34:53.045573401Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.3\" with image id \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\", size \"99286305\" in 3.709074075s" Apr 30 03:34:53.045690 containerd[1992]: time="2025-04-30T03:34:53.045600016Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\"" Apr 30 03:34:53.048293 containerd[1992]: time="2025-04-30T03:34:53.048263300Z" level=info msg="CreateContainer within sandbox \"c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Apr 30 03:34:53.073894 containerd[1992]: time="2025-04-30T03:34:53.073768485Z" level=info msg="CreateContainer within sandbox \"c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd\"" Apr 30 03:34:53.074786 containerd[1992]: time="2025-04-30T03:34:53.074505231Z" level=info msg="StartContainer for \"7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd\"" Apr 30 03:34:53.137530 systemd[1]: run-containerd-runc-k8s.io-7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd-runc.1gK1sM.mount: Deactivated successfully. Apr 30 03:34:53.147279 systemd[1]: Started cri-containerd-7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd.scope - libcontainer container 7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd. Apr 30 03:34:53.179904 containerd[1992]: time="2025-04-30T03:34:53.179854010Z" level=info msg="StartContainer for \"7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd\" returns successfully" Apr 30 03:34:54.049140 systemd[1]: cri-containerd-7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd.scope: Deactivated successfully. Apr 30 03:34:54.092539 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd-rootfs.mount: Deactivated successfully. Apr 30 03:34:54.098597 containerd[1992]: time="2025-04-30T03:34:54.098361701Z" level=info msg="shim disconnected" id=7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd namespace=k8s.io Apr 30 03:34:54.098597 containerd[1992]: time="2025-04-30T03:34:54.098418537Z" level=warning msg="cleaning up after shim disconnected" id=7a006bd341d4cc0a77256b01bb7fbac078f4002d8392dddf27689f96b44938fd namespace=k8s.io Apr 30 03:34:54.098597 containerd[1992]: time="2025-04-30T03:34:54.098432844Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 03:34:54.108561 kubelet[3187]: I0430 03:34:54.108515 3187 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Apr 30 03:34:54.158921 systemd[1]: Created slice kubepods-burstable-pod0e7b462c_a68a_448d_8c44_825a2ecffb4c.slice - libcontainer container kubepods-burstable-pod0e7b462c_a68a_448d_8c44_825a2ecffb4c.slice. Apr 30 03:34:54.176996 systemd[1]: Created slice kubepods-besteffort-pod3ed29fa0_99c1_4495_bb89_f321fae4f034.slice - libcontainer container kubepods-besteffort-pod3ed29fa0_99c1_4495_bb89_f321fae4f034.slice. Apr 30 03:34:54.183719 systemd[1]: Created slice kubepods-besteffort-podb847a159_42eb_41a9_96eb_3e5925320f2d.slice - libcontainer container kubepods-besteffort-podb847a159_42eb_41a9_96eb_3e5925320f2d.slice. Apr 30 03:34:54.194173 systemd[1]: Created slice kubepods-besteffort-podbd94215e_67e1_475c_ae9b_d543ead9090d.slice - libcontainer container kubepods-besteffort-podbd94215e_67e1_475c_ae9b_d543ead9090d.slice. Apr 30 03:34:54.207622 systemd[1]: Created slice kubepods-burstable-pod1e9b51f5_da9d_4ed3_aa3d_02aa185b5484.slice - libcontainer container kubepods-burstable-pod1e9b51f5_da9d_4ed3_aa3d_02aa185b5484.slice. Apr 30 03:34:54.230274 systemd[1]: Created slice kubepods-besteffort-pod5e820db8_1ee7_4b3a_bfb7_17c5a434fca3.slice - libcontainer container kubepods-besteffort-pod5e820db8_1ee7_4b3a_bfb7_17c5a434fca3.slice. Apr 30 03:34:54.238862 containerd[1992]: time="2025-04-30T03:34:54.237914237Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-p2dkq,Uid:5e820db8-1ee7-4b3a-bfb7-17c5a434fca3,Namespace:calico-system,Attempt:0,}" Apr 30 03:34:54.260053 kubelet[3187]: I0430 03:34:54.258655 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lssl\" (UniqueName: \"kubernetes.io/projected/1e9b51f5-da9d-4ed3-aa3d-02aa185b5484-kube-api-access-2lssl\") pod \"coredns-668d6bf9bc-rncmj\" (UID: \"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484\") " pod="kube-system/coredns-668d6bf9bc-rncmj" Apr 30 03:34:54.260053 kubelet[3187]: I0430 03:34:54.258710 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e7b462c-a68a-448d-8c44-825a2ecffb4c-config-volume\") pod \"coredns-668d6bf9bc-xfzg8\" (UID: \"0e7b462c-a68a-448d-8c44-825a2ecffb4c\") " pod="kube-system/coredns-668d6bf9bc-xfzg8" Apr 30 03:34:54.260053 kubelet[3187]: I0430 03:34:54.258744 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qczq\" (UniqueName: \"kubernetes.io/projected/3ed29fa0-99c1-4495-bb89-f321fae4f034-kube-api-access-2qczq\") pod \"calico-kube-controllers-7bf8b64b7f-sxxbz\" (UID: \"3ed29fa0-99c1-4495-bb89-f321fae4f034\") " pod="calico-system/calico-kube-controllers-7bf8b64b7f-sxxbz" Apr 30 03:34:54.260053 kubelet[3187]: I0430 03:34:54.258775 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/bd94215e-67e1-475c-ae9b-d543ead9090d-calico-apiserver-certs\") pod \"calico-apiserver-7ff7f69886-dgpcd\" (UID: \"bd94215e-67e1-475c-ae9b-d543ead9090d\") " pod="calico-apiserver/calico-apiserver-7ff7f69886-dgpcd" Apr 30 03:34:54.260053 kubelet[3187]: I0430 03:34:54.258817 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/b847a159-42eb-41a9-96eb-3e5925320f2d-calico-apiserver-certs\") pod \"calico-apiserver-7ff7f69886-wjp2q\" (UID: \"b847a159-42eb-41a9-96eb-3e5925320f2d\") " pod="calico-apiserver/calico-apiserver-7ff7f69886-wjp2q" Apr 30 03:34:54.260557 kubelet[3187]: I0430 03:34:54.258842 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e9b51f5-da9d-4ed3-aa3d-02aa185b5484-config-volume\") pod \"coredns-668d6bf9bc-rncmj\" (UID: \"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484\") " pod="kube-system/coredns-668d6bf9bc-rncmj" Apr 30 03:34:54.260557 kubelet[3187]: I0430 03:34:54.258869 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjh2q\" (UniqueName: \"kubernetes.io/projected/0e7b462c-a68a-448d-8c44-825a2ecffb4c-kube-api-access-mjh2q\") pod \"coredns-668d6bf9bc-xfzg8\" (UID: \"0e7b462c-a68a-448d-8c44-825a2ecffb4c\") " pod="kube-system/coredns-668d6bf9bc-xfzg8" Apr 30 03:34:54.260557 kubelet[3187]: I0430 03:34:54.258902 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82dcd\" (UniqueName: \"kubernetes.io/projected/bd94215e-67e1-475c-ae9b-d543ead9090d-kube-api-access-82dcd\") pod \"calico-apiserver-7ff7f69886-dgpcd\" (UID: \"bd94215e-67e1-475c-ae9b-d543ead9090d\") " pod="calico-apiserver/calico-apiserver-7ff7f69886-dgpcd" Apr 30 03:34:54.260557 kubelet[3187]: I0430 03:34:54.258971 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66frp\" (UniqueName: \"kubernetes.io/projected/b847a159-42eb-41a9-96eb-3e5925320f2d-kube-api-access-66frp\") pod \"calico-apiserver-7ff7f69886-wjp2q\" (UID: \"b847a159-42eb-41a9-96eb-3e5925320f2d\") " pod="calico-apiserver/calico-apiserver-7ff7f69886-wjp2q" Apr 30 03:34:54.260557 kubelet[3187]: I0430 03:34:54.259049 3187 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ed29fa0-99c1-4495-bb89-f321fae4f034-tigera-ca-bundle\") pod \"calico-kube-controllers-7bf8b64b7f-sxxbz\" (UID: \"3ed29fa0-99c1-4495-bb89-f321fae4f034\") " pod="calico-system/calico-kube-controllers-7bf8b64b7f-sxxbz" Apr 30 03:34:54.352879 containerd[1992]: time="2025-04-30T03:34:54.351756543Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" Apr 30 03:34:54.476758 containerd[1992]: time="2025-04-30T03:34:54.476716199Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-xfzg8,Uid:0e7b462c-a68a-448d-8c44-825a2ecffb4c,Namespace:kube-system,Attempt:0,}" Apr 30 03:34:54.482226 containerd[1992]: time="2025-04-30T03:34:54.482187330Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bf8b64b7f-sxxbz,Uid:3ed29fa0-99c1-4495-bb89-f321fae4f034,Namespace:calico-system,Attempt:0,}" Apr 30 03:34:54.488681 containerd[1992]: time="2025-04-30T03:34:54.488628866Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-wjp2q,Uid:b847a159-42eb-41a9-96eb-3e5925320f2d,Namespace:calico-apiserver,Attempt:0,}" Apr 30 03:34:54.503977 containerd[1992]: time="2025-04-30T03:34:54.503166234Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-dgpcd,Uid:bd94215e-67e1-475c-ae9b-d543ead9090d,Namespace:calico-apiserver,Attempt:0,}" Apr 30 03:34:54.513015 containerd[1992]: time="2025-04-30T03:34:54.512972597Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rncmj,Uid:1e9b51f5-da9d-4ed3-aa3d-02aa185b5484,Namespace:kube-system,Attempt:0,}" Apr 30 03:34:54.775440 containerd[1992]: time="2025-04-30T03:34:54.775376505Z" level=error msg="Failed to destroy network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.786020 containerd[1992]: time="2025-04-30T03:34:54.785946639Z" level=error msg="encountered an error cleaning up failed sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.786276 containerd[1992]: time="2025-04-30T03:34:54.786245405Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-p2dkq,Uid:5e820db8-1ee7-4b3a-bfb7-17c5a434fca3,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.786759 kubelet[3187]: E0430 03:34:54.786711 3187 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.786956 kubelet[3187]: E0430 03:34:54.786932 3187 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-p2dkq" Apr 30 03:34:54.787110 kubelet[3187]: E0430 03:34:54.787088 3187 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-p2dkq" Apr 30 03:34:54.788027 kubelet[3187]: E0430 03:34:54.787651 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-p2dkq_calico-system(5e820db8-1ee7-4b3a-bfb7-17c5a434fca3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-p2dkq_calico-system(5e820db8-1ee7-4b3a-bfb7-17c5a434fca3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-p2dkq" podUID="5e820db8-1ee7-4b3a-bfb7-17c5a434fca3" Apr 30 03:34:54.816137 containerd[1992]: time="2025-04-30T03:34:54.815207822Z" level=error msg="Failed to destroy network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.816137 containerd[1992]: time="2025-04-30T03:34:54.815591305Z" level=error msg="encountered an error cleaning up failed sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.816137 containerd[1992]: time="2025-04-30T03:34:54.815658303Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-wjp2q,Uid:b847a159-42eb-41a9-96eb-3e5925320f2d,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.816493 kubelet[3187]: E0430 03:34:54.815887 3187 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.816493 kubelet[3187]: E0430 03:34:54.815947 3187 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7ff7f69886-wjp2q" Apr 30 03:34:54.816493 kubelet[3187]: E0430 03:34:54.815980 3187 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7ff7f69886-wjp2q" Apr 30 03:34:54.816685 kubelet[3187]: E0430 03:34:54.816094 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7ff7f69886-wjp2q_calico-apiserver(b847a159-42eb-41a9-96eb-3e5925320f2d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7ff7f69886-wjp2q_calico-apiserver(b847a159-42eb-41a9-96eb-3e5925320f2d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7ff7f69886-wjp2q" podUID="b847a159-42eb-41a9-96eb-3e5925320f2d" Apr 30 03:34:54.827932 containerd[1992]: time="2025-04-30T03:34:54.827324530Z" level=error msg="Failed to destroy network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.827932 containerd[1992]: time="2025-04-30T03:34:54.827703422Z" level=error msg="encountered an error cleaning up failed sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.827932 containerd[1992]: time="2025-04-30T03:34:54.827765256Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-xfzg8,Uid:0e7b462c-a68a-448d-8c44-825a2ecffb4c,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.828186 kubelet[3187]: E0430 03:34:54.827991 3187 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.828186 kubelet[3187]: E0430 03:34:54.828081 3187 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-xfzg8" Apr 30 03:34:54.828186 kubelet[3187]: E0430 03:34:54.828111 3187 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-xfzg8" Apr 30 03:34:54.828327 kubelet[3187]: E0430 03:34:54.828163 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-xfzg8_kube-system(0e7b462c-a68a-448d-8c44-825a2ecffb4c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-xfzg8_kube-system(0e7b462c-a68a-448d-8c44-825a2ecffb4c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-xfzg8" podUID="0e7b462c-a68a-448d-8c44-825a2ecffb4c" Apr 30 03:34:54.838447 containerd[1992]: time="2025-04-30T03:34:54.837054975Z" level=error msg="Failed to destroy network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.838447 containerd[1992]: time="2025-04-30T03:34:54.837752518Z" level=error msg="encountered an error cleaning up failed sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.838447 containerd[1992]: time="2025-04-30T03:34:54.837812632Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rncmj,Uid:1e9b51f5-da9d-4ed3-aa3d-02aa185b5484,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.840218 containerd[1992]: time="2025-04-30T03:34:54.839143192Z" level=error msg="Failed to destroy network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.841063 kubelet[3187]: E0430 03:34:54.839394 3187 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.841063 kubelet[3187]: E0430 03:34:54.839469 3187 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-rncmj" Apr 30 03:34:54.841063 kubelet[3187]: E0430 03:34:54.839495 3187 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-rncmj" Apr 30 03:34:54.841262 kubelet[3187]: E0430 03:34:54.839551 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-rncmj_kube-system(1e9b51f5-da9d-4ed3-aa3d-02aa185b5484)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-rncmj_kube-system(1e9b51f5-da9d-4ed3-aa3d-02aa185b5484)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-rncmj" podUID="1e9b51f5-da9d-4ed3-aa3d-02aa185b5484" Apr 30 03:34:54.844026 containerd[1992]: time="2025-04-30T03:34:54.842989117Z" level=error msg="encountered an error cleaning up failed sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.844026 containerd[1992]: time="2025-04-30T03:34:54.843163015Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-dgpcd,Uid:bd94215e-67e1-475c-ae9b-d543ead9090d,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.844530 kubelet[3187]: E0430 03:34:54.843563 3187 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.844530 kubelet[3187]: E0430 03:34:54.843646 3187 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7ff7f69886-dgpcd" Apr 30 03:34:54.844530 kubelet[3187]: E0430 03:34:54.843687 3187 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7ff7f69886-dgpcd" Apr 30 03:34:54.845515 kubelet[3187]: E0430 03:34:54.843765 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7ff7f69886-dgpcd_calico-apiserver(bd94215e-67e1-475c-ae9b-d543ead9090d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7ff7f69886-dgpcd_calico-apiserver(bd94215e-67e1-475c-ae9b-d543ead9090d)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7ff7f69886-dgpcd" podUID="bd94215e-67e1-475c-ae9b-d543ead9090d" Apr 30 03:34:54.854164 containerd[1992]: time="2025-04-30T03:34:54.854111067Z" level=error msg="Failed to destroy network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.854975 containerd[1992]: time="2025-04-30T03:34:54.854762863Z" level=error msg="encountered an error cleaning up failed sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.854975 containerd[1992]: time="2025-04-30T03:34:54.854859938Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bf8b64b7f-sxxbz,Uid:3ed29fa0-99c1-4495-bb89-f321fae4f034,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.856281 kubelet[3187]: E0430 03:34:54.855170 3187 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:54.856281 kubelet[3187]: E0430 03:34:54.855230 3187 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7bf8b64b7f-sxxbz" Apr 30 03:34:54.856281 kubelet[3187]: E0430 03:34:54.855260 3187 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7bf8b64b7f-sxxbz" Apr 30 03:34:54.856436 kubelet[3187]: E0430 03:34:54.855314 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7bf8b64b7f-sxxbz_calico-system(3ed29fa0-99c1-4495-bb89-f321fae4f034)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7bf8b64b7f-sxxbz_calico-system(3ed29fa0-99c1-4495-bb89-f321fae4f034)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7bf8b64b7f-sxxbz" podUID="3ed29fa0-99c1-4495-bb89-f321fae4f034" Apr 30 03:34:55.095153 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529-shm.mount: Deactivated successfully. Apr 30 03:34:55.352775 kubelet[3187]: I0430 03:34:55.352675 3187 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:34:55.356180 kubelet[3187]: I0430 03:34:55.355165 3187 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:34:55.357541 containerd[1992]: time="2025-04-30T03:34:55.357271231Z" level=info msg="StopPodSandbox for \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\"" Apr 30 03:34:55.358469 containerd[1992]: time="2025-04-30T03:34:55.358436436Z" level=info msg="Ensure that sandbox 6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9 in task-service has been cleanup successfully" Apr 30 03:34:55.362065 containerd[1992]: time="2025-04-30T03:34:55.362016505Z" level=info msg="StopPodSandbox for \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\"" Apr 30 03:34:55.362273 containerd[1992]: time="2025-04-30T03:34:55.362251848Z" level=info msg="Ensure that sandbox 365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f in task-service has been cleanup successfully" Apr 30 03:34:55.366953 kubelet[3187]: I0430 03:34:55.366921 3187 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:34:55.368575 containerd[1992]: time="2025-04-30T03:34:55.368547316Z" level=info msg="StopPodSandbox for \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\"" Apr 30 03:34:55.368968 containerd[1992]: time="2025-04-30T03:34:55.368705148Z" level=info msg="Ensure that sandbox fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529 in task-service has been cleanup successfully" Apr 30 03:34:55.370615 kubelet[3187]: I0430 03:34:55.370377 3187 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:34:55.371958 containerd[1992]: time="2025-04-30T03:34:55.371591084Z" level=info msg="StopPodSandbox for \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\"" Apr 30 03:34:55.371958 containerd[1992]: time="2025-04-30T03:34:55.371750661Z" level=info msg="Ensure that sandbox 1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097 in task-service has been cleanup successfully" Apr 30 03:34:55.374551 kubelet[3187]: I0430 03:34:55.374488 3187 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:34:55.376721 containerd[1992]: time="2025-04-30T03:34:55.376660590Z" level=info msg="StopPodSandbox for \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\"" Apr 30 03:34:55.377253 containerd[1992]: time="2025-04-30T03:34:55.377218045Z" level=info msg="Ensure that sandbox 74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af in task-service has been cleanup successfully" Apr 30 03:34:55.382376 kubelet[3187]: I0430 03:34:55.381891 3187 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:34:55.382984 containerd[1992]: time="2025-04-30T03:34:55.382946986Z" level=info msg="StopPodSandbox for \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\"" Apr 30 03:34:55.383560 containerd[1992]: time="2025-04-30T03:34:55.383173035Z" level=info msg="Ensure that sandbox 5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9 in task-service has been cleanup successfully" Apr 30 03:34:55.471352 containerd[1992]: time="2025-04-30T03:34:55.471305483Z" level=error msg="StopPodSandbox for \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\" failed" error="failed to destroy network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:55.471753 containerd[1992]: time="2025-04-30T03:34:55.471681667Z" level=error msg="StopPodSandbox for \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\" failed" error="failed to destroy network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:55.471929 kubelet[3187]: E0430 03:34:55.471870 3187 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:34:55.472073 kubelet[3187]: E0430 03:34:55.471930 3187 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af"} Apr 30 03:34:55.472073 kubelet[3187]: E0430 03:34:55.471991 3187 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"bd94215e-67e1-475c-ae9b-d543ead9090d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 03:34:55.472073 kubelet[3187]: E0430 03:34:55.472025 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"bd94215e-67e1-475c-ae9b-d543ead9090d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7ff7f69886-dgpcd" podUID="bd94215e-67e1-475c-ae9b-d543ead9090d" Apr 30 03:34:55.472073 kubelet[3187]: E0430 03:34:55.472056 3187 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:34:55.472073 kubelet[3187]: E0430 03:34:55.472070 3187 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529"} Apr 30 03:34:55.472269 kubelet[3187]: E0430 03:34:55.472090 3187 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 03:34:55.472269 kubelet[3187]: E0430 03:34:55.472104 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-p2dkq" podUID="5e820db8-1ee7-4b3a-bfb7-17c5a434fca3" Apr 30 03:34:55.474408 containerd[1992]: time="2025-04-30T03:34:55.474248298Z" level=error msg="StopPodSandbox for \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\" failed" error="failed to destroy network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:55.474649 kubelet[3187]: E0430 03:34:55.474505 3187 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:34:55.474713 kubelet[3187]: E0430 03:34:55.474663 3187 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f"} Apr 30 03:34:55.474758 kubelet[3187]: E0430 03:34:55.474706 3187 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"3ed29fa0-99c1-4495-bb89-f321fae4f034\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 03:34:55.474825 kubelet[3187]: E0430 03:34:55.474745 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"3ed29fa0-99c1-4495-bb89-f321fae4f034\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7bf8b64b7f-sxxbz" podUID="3ed29fa0-99c1-4495-bb89-f321fae4f034" Apr 30 03:34:55.474958 containerd[1992]: time="2025-04-30T03:34:55.474863868Z" level=error msg="StopPodSandbox for \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\" failed" error="failed to destroy network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:55.475490 kubelet[3187]: E0430 03:34:55.475358 3187 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:34:55.475490 kubelet[3187]: E0430 03:34:55.475445 3187 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9"} Apr 30 03:34:55.475584 kubelet[3187]: E0430 03:34:55.475489 3187 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"0e7b462c-a68a-448d-8c44-825a2ecffb4c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 03:34:55.475584 kubelet[3187]: E0430 03:34:55.475509 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"0e7b462c-a68a-448d-8c44-825a2ecffb4c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-xfzg8" podUID="0e7b462c-a68a-448d-8c44-825a2ecffb4c" Apr 30 03:34:55.478533 containerd[1992]: time="2025-04-30T03:34:55.478328496Z" level=error msg="StopPodSandbox for \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\" failed" error="failed to destroy network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:55.478962 kubelet[3187]: E0430 03:34:55.478922 3187 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:34:55.479048 kubelet[3187]: E0430 03:34:55.478970 3187 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9"} Apr 30 03:34:55.479048 kubelet[3187]: E0430 03:34:55.479016 3187 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"b847a159-42eb-41a9-96eb-3e5925320f2d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 03:34:55.479048 kubelet[3187]: E0430 03:34:55.479035 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"b847a159-42eb-41a9-96eb-3e5925320f2d\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7ff7f69886-wjp2q" podUID="b847a159-42eb-41a9-96eb-3e5925320f2d" Apr 30 03:34:55.492041 containerd[1992]: time="2025-04-30T03:34:55.491509191Z" level=error msg="StopPodSandbox for \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\" failed" error="failed to destroy network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 03:34:55.492170 kubelet[3187]: E0430 03:34:55.491775 3187 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:34:55.492170 kubelet[3187]: E0430 03:34:55.491838 3187 kuberuntime_manager.go:1546] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097"} Apr 30 03:34:55.492170 kubelet[3187]: E0430 03:34:55.491886 3187 kuberuntime_manager.go:1146] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Apr 30 03:34:55.492170 kubelet[3187]: E0430 03:34:55.491910 3187 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-rncmj" podUID="1e9b51f5-da9d-4ed3-aa3d-02aa185b5484" Apr 30 03:35:00.541389 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1652142888.mount: Deactivated successfully. Apr 30 03:35:00.657787 containerd[1992]: time="2025-04-30T03:35:00.657707386Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.3: active requests=0, bytes read=144068748" Apr 30 03:35:00.664134 containerd[1992]: time="2025-04-30T03:35:00.663367426Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:00.680810 containerd[1992]: time="2025-04-30T03:35:00.680733780Z" level=info msg="ImageCreate event name:\"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:00.682754 containerd[1992]: time="2025-04-30T03:35:00.682687057Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:00.688225 containerd[1992]: time="2025-04-30T03:35:00.688171579Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.3\" with image id \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\", size \"144068610\" in 6.331382231s" Apr 30 03:35:00.688225 containerd[1992]: time="2025-04-30T03:35:00.688223937Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\"" Apr 30 03:35:00.774012 containerd[1992]: time="2025-04-30T03:35:00.773961416Z" level=info msg="CreateContainer within sandbox \"c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Apr 30 03:35:00.832626 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2854111019.mount: Deactivated successfully. Apr 30 03:35:00.850633 containerd[1992]: time="2025-04-30T03:35:00.850428689Z" level=info msg="CreateContainer within sandbox \"c6b88c697aa8efe5a5173c2eca0882e9722c0f64a86ec598b8d70157d4ea023e\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"9709400f2d62de8cdeb9c90dadcf69791774657e58b7c8331fa93adacf9a15d2\"" Apr 30 03:35:00.851736 containerd[1992]: time="2025-04-30T03:35:00.851584039Z" level=info msg="StartContainer for \"9709400f2d62de8cdeb9c90dadcf69791774657e58b7c8331fa93adacf9a15d2\"" Apr 30 03:35:00.976299 systemd[1]: Started cri-containerd-9709400f2d62de8cdeb9c90dadcf69791774657e58b7c8331fa93adacf9a15d2.scope - libcontainer container 9709400f2d62de8cdeb9c90dadcf69791774657e58b7c8331fa93adacf9a15d2. Apr 30 03:35:01.048635 containerd[1992]: time="2025-04-30T03:35:01.048596983Z" level=info msg="StartContainer for \"9709400f2d62de8cdeb9c90dadcf69791774657e58b7c8331fa93adacf9a15d2\" returns successfully" Apr 30 03:35:01.225285 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Apr 30 03:35:01.235652 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Apr 30 03:35:01.502459 kubelet[3187]: I0430 03:35:01.493271 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-nql5l" podStartSLOduration=1.729665792 podStartE2EDuration="17.478963183s" podCreationTimestamp="2025-04-30 03:34:44 +0000 UTC" firstStartedPulling="2025-04-30 03:34:44.939615617 +0000 UTC m=+15.834903243" lastFinishedPulling="2025-04-30 03:35:00.688913007 +0000 UTC m=+31.584200634" observedRunningTime="2025-04-30 03:35:01.471469478 +0000 UTC m=+32.366757128" watchObservedRunningTime="2025-04-30 03:35:01.478963183 +0000 UTC m=+32.374250818" Apr 30 03:35:02.419765 kubelet[3187]: I0430 03:35:02.419713 3187 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 03:35:06.219345 containerd[1992]: time="2025-04-30T03:35:06.219278171Z" level=info msg="StopPodSandbox for \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\"" Apr 30 03:35:06.221024 containerd[1992]: time="2025-04-30T03:35:06.219296297Z" level=info msg="StopPodSandbox for \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\"" Apr 30 03:35:06.532804 kubelet[3187]: I0430 03:35:06.532660 3187 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.327 [INFO][4633] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.328 [INFO][4633] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" iface="eth0" netns="/var/run/netns/cni-3d762cb5-3646-95e4-be3b-7b57e769b952" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.328 [INFO][4633] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" iface="eth0" netns="/var/run/netns/cni-3d762cb5-3646-95e4-be3b-7b57e769b952" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.332 [INFO][4633] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" iface="eth0" netns="/var/run/netns/cni-3d762cb5-3646-95e4-be3b-7b57e769b952" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.332 [INFO][4633] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.332 [INFO][4633] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.706 [INFO][4643] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.708 [INFO][4643] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.709 [INFO][4643] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.725 [WARNING][4643] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.725 [INFO][4643] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.728 [INFO][4643] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:06.739487 containerd[1992]: 2025-04-30 03:35:06.736 [INFO][4633] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:06.746619 systemd[1]: run-netns-cni\x2d3d762cb5\x2d3646\x2d95e4\x2dbe3b\x2d7b57e769b952.mount: Deactivated successfully. Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.332 [INFO][4624] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.332 [INFO][4624] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" iface="eth0" netns="/var/run/netns/cni-03eb430b-fa17-5e6e-2c2c-c16662512c25" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.333 [INFO][4624] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" iface="eth0" netns="/var/run/netns/cni-03eb430b-fa17-5e6e-2c2c-c16662512c25" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.333 [INFO][4624] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" iface="eth0" netns="/var/run/netns/cni-03eb430b-fa17-5e6e-2c2c-c16662512c25" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.333 [INFO][4624] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.333 [INFO][4624] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.705 [INFO][4642] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.710 [INFO][4642] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.727 [INFO][4642] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.742 [WARNING][4642] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.742 [INFO][4642] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.744 [INFO][4642] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:06.760576 containerd[1992]: 2025-04-30 03:35:06.755 [INFO][4624] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:06.760576 containerd[1992]: time="2025-04-30T03:35:06.760363175Z" level=info msg="TearDown network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\" successfully" Apr 30 03:35:06.760576 containerd[1992]: time="2025-04-30T03:35:06.760395349Z" level=info msg="StopPodSandbox for \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\" returns successfully" Apr 30 03:35:06.774275 containerd[1992]: time="2025-04-30T03:35:06.762167265Z" level=info msg="TearDown network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\" successfully" Apr 30 03:35:06.774275 containerd[1992]: time="2025-04-30T03:35:06.762203905Z" level=info msg="StopPodSandbox for \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\" returns successfully" Apr 30 03:35:06.774275 containerd[1992]: time="2025-04-30T03:35:06.764587449Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-xfzg8,Uid:0e7b462c-a68a-448d-8c44-825a2ecffb4c,Namespace:kube-system,Attempt:1,}" Apr 30 03:35:06.774275 containerd[1992]: time="2025-04-30T03:35:06.768364573Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rncmj,Uid:1e9b51f5-da9d-4ed3-aa3d-02aa185b5484,Namespace:kube-system,Attempt:1,}" Apr 30 03:35:06.770135 systemd[1]: run-netns-cni\x2d03eb430b\x2dfa17\x2d5e6e\x2d2c2c\x2dc16662512c25.mount: Deactivated successfully. Apr 30 03:35:07.080531 systemd[1]: Started sshd@7-172.31.20.13:22-147.75.109.163:35606.service - OpenSSH per-connection server daemon (147.75.109.163:35606). Apr 30 03:35:07.222844 containerd[1992]: time="2025-04-30T03:35:07.222462334Z" level=info msg="StopPodSandbox for \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\"" Apr 30 03:35:07.439416 sshd[4711]: Accepted publickey for core from 147.75.109.163 port 35606 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:07.446117 sshd[4711]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:07.460237 systemd-logind[1970]: New session 8 of user core. Apr 30 03:35:07.465753 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 30 03:35:07.490887 systemd-networkd[1829]: calibb075411776: Link UP Apr 30 03:35:07.494834 systemd-networkd[1829]: calibb075411776: Gained carrier Apr 30 03:35:07.495977 (udev-worker)[4771]: Network interface NamePolicy= disabled on kernel command line. Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:06.912 [INFO][4676] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:06.998 [INFO][4676] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0 coredns-668d6bf9bc- kube-system 1e9b51f5-da9d-4ed3-aa3d-02aa185b5484 722 0 2025-04-30 03:34:37 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-20-13 coredns-668d6bf9bc-rncmj eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calibb075411776 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:06.999 [INFO][4676] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.190 [INFO][4702] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" HandleID="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.321 [INFO][4702] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" HandleID="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00039f480), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-20-13", "pod":"coredns-668d6bf9bc-rncmj", "timestamp":"2025-04-30 03:35:07.190712451 +0000 UTC"}, Hostname:"ip-172-31-20-13", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.321 [INFO][4702] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.321 [INFO][4702] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.321 [INFO][4702] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-20-13' Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.331 [INFO][4702] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.347 [INFO][4702] ipam/ipam.go 372: Looking up existing affinities for host host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.369 [INFO][4702] ipam/ipam.go 489: Trying affinity for 192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.379 [INFO][4702] ipam/ipam.go 155: Attempting to load block cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.389 [INFO][4702] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.389 [INFO][4702] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.23.0/26 handle="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.394 [INFO][4702] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.415 [INFO][4702] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.23.0/26 handle="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.434 [INFO][4702] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.23.1/26] block=192.168.23.0/26 handle="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.434 [INFO][4702] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.23.1/26] handle="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" host="ip-172-31-20-13" Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.434 [INFO][4702] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:07.541445 containerd[1992]: 2025-04-30 03:35:07.434 [INFO][4702] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.23.1/26] IPv6=[] ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" HandleID="k8s-pod-network.7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:07.544150 containerd[1992]: 2025-04-30 03:35:07.455 [INFO][4676] cni-plugin/k8s.go 386: Populated endpoint ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484", ResourceVersion:"722", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"", Pod:"coredns-668d6bf9bc-rncmj", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calibb075411776", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:07.544150 containerd[1992]: 2025-04-30 03:35:07.459 [INFO][4676] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.23.1/32] ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:07.544150 containerd[1992]: 2025-04-30 03:35:07.459 [INFO][4676] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calibb075411776 ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:07.544150 containerd[1992]: 2025-04-30 03:35:07.498 [INFO][4676] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:07.544150 containerd[1992]: 2025-04-30 03:35:07.501 [INFO][4676] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484", ResourceVersion:"722", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c", Pod:"coredns-668d6bf9bc-rncmj", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calibb075411776", MAC:"82:61:13:08:36:06", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:07.544150 containerd[1992]: 2025-04-30 03:35:07.533 [INFO][4676] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c" Namespace="kube-system" Pod="coredns-668d6bf9bc-rncmj" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:07.616086 systemd-networkd[1829]: cali22e4b179ea0: Link UP Apr 30 03:35:07.616465 systemd-networkd[1829]: cali22e4b179ea0: Gained carrier Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.359 [INFO][4731] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.360 [INFO][4731] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" iface="eth0" netns="/var/run/netns/cni-69ace9e3-29db-8d73-93e2-07a5dfd28096" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.360 [INFO][4731] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" iface="eth0" netns="/var/run/netns/cni-69ace9e3-29db-8d73-93e2-07a5dfd28096" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.361 [INFO][4731] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" iface="eth0" netns="/var/run/netns/cni-69ace9e3-29db-8d73-93e2-07a5dfd28096" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.361 [INFO][4731] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.362 [INFO][4731] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.486 [INFO][4744] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.498 [INFO][4744] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.583 [INFO][4744] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.616 [WARNING][4744] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.620 [INFO][4744] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.627 [INFO][4744] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:07.640098 containerd[1992]: 2025-04-30 03:35:07.633 [INFO][4731] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:07.645131 containerd[1992]: time="2025-04-30T03:35:07.643244951Z" level=info msg="TearDown network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\" successfully" Apr 30 03:35:07.646602 systemd[1]: run-netns-cni\x2d69ace9e3\x2d29db\x2d8d73\x2d93e2\x2d07a5dfd28096.mount: Deactivated successfully. Apr 30 03:35:07.647695 containerd[1992]: time="2025-04-30T03:35:07.644970588Z" level=info msg="StopPodSandbox for \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\" returns successfully" Apr 30 03:35:07.651472 containerd[1992]: time="2025-04-30T03:35:07.651202634Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-p2dkq,Uid:5e820db8-1ee7-4b3a-bfb7-17c5a434fca3,Namespace:calico-system,Attempt:1,}" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:06.945 [INFO][4674] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:06.993 [INFO][4674] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0 coredns-668d6bf9bc- kube-system 0e7b462c-a68a-448d-8c44-825a2ecffb4c 721 0 2025-04-30 03:34:37 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ip-172-31-20-13 coredns-668d6bf9bc-xfzg8 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali22e4b179ea0 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:06.993 [INFO][4674] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.214 [INFO][4700] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" HandleID="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.330 [INFO][4700] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" HandleID="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ec0c0), Attrs:map[string]string{"namespace":"kube-system", "node":"ip-172-31-20-13", "pod":"coredns-668d6bf9bc-xfzg8", "timestamp":"2025-04-30 03:35:07.214683236 +0000 UTC"}, Hostname:"ip-172-31-20-13", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.330 [INFO][4700] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.438 [INFO][4700] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.438 [INFO][4700] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-20-13' Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.449 [INFO][4700] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.475 [INFO][4700] ipam/ipam.go 372: Looking up existing affinities for host host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.507 [INFO][4700] ipam/ipam.go 489: Trying affinity for 192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.512 [INFO][4700] ipam/ipam.go 155: Attempting to load block cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.534 [INFO][4700] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.534 [INFO][4700] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.23.0/26 handle="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.545 [INFO][4700] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583 Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.563 [INFO][4700] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.23.0/26 handle="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.583 [INFO][4700] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.23.2/26] block=192.168.23.0/26 handle="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.583 [INFO][4700] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.23.2/26] handle="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" host="ip-172-31-20-13" Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.583 [INFO][4700] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:07.693631 containerd[1992]: 2025-04-30 03:35:07.584 [INFO][4700] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.23.2/26] IPv6=[] ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" HandleID="k8s-pod-network.b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:07.694974 containerd[1992]: 2025-04-30 03:35:07.592 [INFO][4674] cni-plugin/k8s.go 386: Populated endpoint ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0e7b462c-a68a-448d-8c44-825a2ecffb4c", ResourceVersion:"721", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"", Pod:"coredns-668d6bf9bc-xfzg8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali22e4b179ea0", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:07.694974 containerd[1992]: 2025-04-30 03:35:07.592 [INFO][4674] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.23.2/32] ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:07.694974 containerd[1992]: 2025-04-30 03:35:07.592 [INFO][4674] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali22e4b179ea0 ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:07.694974 containerd[1992]: 2025-04-30 03:35:07.620 [INFO][4674] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:07.694974 containerd[1992]: 2025-04-30 03:35:07.623 [INFO][4674] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0e7b462c-a68a-448d-8c44-825a2ecffb4c", ResourceVersion:"721", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583", Pod:"coredns-668d6bf9bc-xfzg8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali22e4b179ea0", MAC:"ca:01:a5:49:31:3f", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:07.694974 containerd[1992]: 2025-04-30 03:35:07.678 [INFO][4674] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583" Namespace="kube-system" Pod="coredns-668d6bf9bc-xfzg8" WorkloadEndpoint="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:07.701594 containerd[1992]: time="2025-04-30T03:35:07.696395460Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:35:07.701594 containerd[1992]: time="2025-04-30T03:35:07.696480395Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:35:07.701594 containerd[1992]: time="2025-04-30T03:35:07.696505077Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:07.701594 containerd[1992]: time="2025-04-30T03:35:07.696624922Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:07.768473 systemd[1]: Started cri-containerd-7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c.scope - libcontainer container 7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c. Apr 30 03:35:07.849058 containerd[1992]: time="2025-04-30T03:35:07.848080745Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:35:07.849058 containerd[1992]: time="2025-04-30T03:35:07.848201279Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:35:07.849058 containerd[1992]: time="2025-04-30T03:35:07.848220099Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:07.849058 containerd[1992]: time="2025-04-30T03:35:07.848344461Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:07.910652 containerd[1992]: time="2025-04-30T03:35:07.910607700Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-rncmj,Uid:1e9b51f5-da9d-4ed3-aa3d-02aa185b5484,Namespace:kube-system,Attempt:1,} returns sandbox id \"7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c\"" Apr 30 03:35:07.920852 containerd[1992]: time="2025-04-30T03:35:07.920813921Z" level=info msg="CreateContainer within sandbox \"7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 30 03:35:07.959102 systemd[1]: Started cri-containerd-b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583.scope - libcontainer container b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583. Apr 30 03:35:08.017813 containerd[1992]: time="2025-04-30T03:35:08.017750906Z" level=info msg="CreateContainer within sandbox \"7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c6753c8b6151b0201a7a323ef5b75e3b6d7ae3402ae52840bc68ea9cdf57162d\"" Apr 30 03:35:08.021349 containerd[1992]: time="2025-04-30T03:35:08.021305110Z" level=info msg="StartContainer for \"c6753c8b6151b0201a7a323ef5b75e3b6d7ae3402ae52840bc68ea9cdf57162d\"" Apr 30 03:35:08.155814 systemd[1]: Started cri-containerd-c6753c8b6151b0201a7a323ef5b75e3b6d7ae3402ae52840bc68ea9cdf57162d.scope - libcontainer container c6753c8b6151b0201a7a323ef5b75e3b6d7ae3402ae52840bc68ea9cdf57162d. Apr 30 03:35:08.169128 containerd[1992]: time="2025-04-30T03:35:08.169081678Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-xfzg8,Uid:0e7b462c-a68a-448d-8c44-825a2ecffb4c,Namespace:kube-system,Attempt:1,} returns sandbox id \"b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583\"" Apr 30 03:35:08.181024 containerd[1992]: time="2025-04-30T03:35:08.180019610Z" level=info msg="CreateContainer within sandbox \"b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 30 03:35:08.216200 kernel: bpftool[4942]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Apr 30 03:35:08.223344 containerd[1992]: time="2025-04-30T03:35:08.222439611Z" level=info msg="StopPodSandbox for \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\"" Apr 30 03:35:08.232941 containerd[1992]: time="2025-04-30T03:35:08.229443108Z" level=info msg="CreateContainer within sandbox \"b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b6acbb69e78ee6cb263fac724e62d4829ccd090bb9d551462e3d10c320c41296\"" Apr 30 03:35:08.232941 containerd[1992]: time="2025-04-30T03:35:08.229648791Z" level=info msg="StopPodSandbox for \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\"" Apr 30 03:35:08.236800 systemd-networkd[1829]: calia9b1461fcde: Link UP Apr 30 03:35:08.244272 systemd-networkd[1829]: calia9b1461fcde: Gained carrier Apr 30 03:35:08.248951 containerd[1992]: time="2025-04-30T03:35:08.247781531Z" level=info msg="StartContainer for \"b6acbb69e78ee6cb263fac724e62d4829ccd090bb9d551462e3d10c320c41296\"" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:07.847 [INFO][4840] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:07.928 [INFO][4840] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0 csi-node-driver- calico-system 5e820db8-1ee7-4b3a-bfb7-17c5a434fca3 766 0 2025-04-30 03:34:44 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:5b5cc68cd5 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ip-172-31-20-13 csi-node-driver-p2dkq eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calia9b1461fcde [] []}} ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:07.935 [INFO][4840] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.096 [INFO][4892] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" HandleID="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.128 [INFO][4892] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" HandleID="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000f80d0), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-20-13", "pod":"csi-node-driver-p2dkq", "timestamp":"2025-04-30 03:35:08.094653716 +0000 UTC"}, Hostname:"ip-172-31-20-13", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.129 [INFO][4892] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.129 [INFO][4892] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.129 [INFO][4892] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-20-13' Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.134 [INFO][4892] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.143 [INFO][4892] ipam/ipam.go 372: Looking up existing affinities for host host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.150 [INFO][4892] ipam/ipam.go 489: Trying affinity for 192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.156 [INFO][4892] ipam/ipam.go 155: Attempting to load block cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.161 [INFO][4892] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.161 [INFO][4892] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.23.0/26 handle="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.167 [INFO][4892] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.195 [INFO][4892] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.23.0/26 handle="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.212 [INFO][4892] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.23.3/26] block=192.168.23.0/26 handle="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.213 [INFO][4892] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.23.3/26] handle="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" host="ip-172-31-20-13" Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.213 [INFO][4892] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:08.298211 containerd[1992]: 2025-04-30 03:35:08.213 [INFO][4892] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.23.3/26] IPv6=[] ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" HandleID="k8s-pod-network.f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:08.299338 containerd[1992]: 2025-04-30 03:35:08.222 [INFO][4840] cni-plugin/k8s.go 386: Populated endpoint ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3", ResourceVersion:"766", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"", Pod:"csi-node-driver-p2dkq", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.23.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia9b1461fcde", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:08.299338 containerd[1992]: 2025-04-30 03:35:08.222 [INFO][4840] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.23.3/32] ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:08.299338 containerd[1992]: 2025-04-30 03:35:08.222 [INFO][4840] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia9b1461fcde ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:08.299338 containerd[1992]: 2025-04-30 03:35:08.250 [INFO][4840] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:08.299338 containerd[1992]: 2025-04-30 03:35:08.252 [INFO][4840] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3", ResourceVersion:"766", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af", Pod:"csi-node-driver-p2dkq", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.23.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia9b1461fcde", MAC:"36:25:5a:52:14:98", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:08.299338 containerd[1992]: 2025-04-30 03:35:08.286 [INFO][4840] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af" Namespace="calico-system" Pod="csi-node-driver-p2dkq" WorkloadEndpoint="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:08.382609 containerd[1992]: time="2025-04-30T03:35:08.378081898Z" level=info msg="StartContainer for \"c6753c8b6151b0201a7a323ef5b75e3b6d7ae3402ae52840bc68ea9cdf57162d\" returns successfully" Apr 30 03:35:08.441474 systemd[1]: Started cri-containerd-b6acbb69e78ee6cb263fac724e62d4829ccd090bb9d551462e3d10c320c41296.scope - libcontainer container b6acbb69e78ee6cb263fac724e62d4829ccd090bb9d551462e3d10c320c41296. Apr 30 03:35:08.511336 sshd[4711]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:08.522632 systemd[1]: sshd@7-172.31.20.13:22-147.75.109.163:35606.service: Deactivated successfully. Apr 30 03:35:08.529019 systemd[1]: session-8.scope: Deactivated successfully. Apr 30 03:35:08.540157 systemd-logind[1970]: Session 8 logged out. Waiting for processes to exit. Apr 30 03:35:08.549286 systemd-logind[1970]: Removed session 8. Apr 30 03:35:08.553592 containerd[1992]: time="2025-04-30T03:35:08.553251827Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:35:08.554096 containerd[1992]: time="2025-04-30T03:35:08.553859326Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:35:08.555140 containerd[1992]: time="2025-04-30T03:35:08.555090954Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:08.555442 containerd[1992]: time="2025-04-30T03:35:08.555344386Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:08.610265 systemd[1]: Started cri-containerd-f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af.scope - libcontainer container f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af. Apr 30 03:35:08.616047 kubelet[3187]: I0430 03:35:08.615705 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-rncmj" podStartSLOduration=31.615538695 podStartE2EDuration="31.615538695s" podCreationTimestamp="2025-04-30 03:34:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 03:35:08.520094118 +0000 UTC m=+39.415381774" watchObservedRunningTime="2025-04-30 03:35:08.615538695 +0000 UTC m=+39.510826329" Apr 30 03:35:08.640792 containerd[1992]: time="2025-04-30T03:35:08.640741931Z" level=info msg="StartContainer for \"b6acbb69e78ee6cb263fac724e62d4829ccd090bb9d551462e3d10c320c41296\" returns successfully" Apr 30 03:35:08.752935 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount289993096.mount: Deactivated successfully. Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.617 [INFO][4997] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.617 [INFO][4997] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" iface="eth0" netns="/var/run/netns/cni-c794b0c5-48db-86a8-5c37-8fae1f89263c" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.618 [INFO][4997] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" iface="eth0" netns="/var/run/netns/cni-c794b0c5-48db-86a8-5c37-8fae1f89263c" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.619 [INFO][4997] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" iface="eth0" netns="/var/run/netns/cni-c794b0c5-48db-86a8-5c37-8fae1f89263c" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.620 [INFO][4997] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.620 [INFO][4997] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.730 [INFO][5061] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.733 [INFO][5061] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.733 [INFO][5061] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.754 [WARNING][5061] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.757 [INFO][5061] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.762 [INFO][5061] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:08.782684 containerd[1992]: 2025-04-30 03:35:08.773 [INFO][4997] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:08.787350 containerd[1992]: time="2025-04-30T03:35:08.786707437Z" level=info msg="TearDown network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\" successfully" Apr 30 03:35:08.787350 containerd[1992]: time="2025-04-30T03:35:08.786746885Z" level=info msg="StopPodSandbox for \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\" returns successfully" Apr 30 03:35:08.789453 systemd[1]: run-netns-cni\x2dc794b0c5\x2d48db\x2d86a8\x2d5c37\x2d8fae1f89263c.mount: Deactivated successfully. Apr 30 03:35:08.791189 containerd[1992]: time="2025-04-30T03:35:08.790076710Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-dgpcd,Uid:bd94215e-67e1-475c-ae9b-d543ead9090d,Namespace:calico-apiserver,Attempt:1,}" Apr 30 03:35:08.809177 containerd[1992]: time="2025-04-30T03:35:08.809135496Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-p2dkq,Uid:5e820db8-1ee7-4b3a-bfb7-17c5a434fca3,Namespace:calico-system,Attempt:1,} returns sandbox id \"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af\"" Apr 30 03:35:08.822650 containerd[1992]: time="2025-04-30T03:35:08.822396841Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.671 [INFO][4995] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.675 [INFO][4995] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" iface="eth0" netns="/var/run/netns/cni-393700c0-3b81-f337-a80f-c4cd875ee082" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.677 [INFO][4995] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" iface="eth0" netns="/var/run/netns/cni-393700c0-3b81-f337-a80f-c4cd875ee082" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.679 [INFO][4995] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" iface="eth0" netns="/var/run/netns/cni-393700c0-3b81-f337-a80f-c4cd875ee082" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.679 [INFO][4995] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.680 [INFO][4995] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.802 [INFO][5078] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.803 [INFO][5078] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.806 [INFO][5078] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.825 [WARNING][5078] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.826 [INFO][5078] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.835 [INFO][5078] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:08.860518 containerd[1992]: 2025-04-30 03:35:08.849 [INFO][4995] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:08.862256 containerd[1992]: time="2025-04-30T03:35:08.860909468Z" level=info msg="TearDown network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\" successfully" Apr 30 03:35:08.862256 containerd[1992]: time="2025-04-30T03:35:08.860945125Z" level=info msg="StopPodSandbox for \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\" returns successfully" Apr 30 03:35:08.864255 containerd[1992]: time="2025-04-30T03:35:08.863790139Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bf8b64b7f-sxxbz,Uid:3ed29fa0-99c1-4495-bb89-f321fae4f034,Namespace:calico-system,Attempt:1,}" Apr 30 03:35:08.868959 systemd[1]: run-netns-cni\x2d393700c0\x2d3b81\x2df337\x2da80f\x2dc4cd875ee082.mount: Deactivated successfully. Apr 30 03:35:09.185106 systemd-networkd[1829]: cali22e4b179ea0: Gained IPv6LL Apr 30 03:35:09.200548 systemd-networkd[1829]: calie0f68c29924: Link UP Apr 30 03:35:09.203579 systemd-networkd[1829]: calie0f68c29924: Gained carrier Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:08.998 [INFO][5095] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0 calico-apiserver-7ff7f69886- calico-apiserver bd94215e-67e1-475c-ae9b-d543ead9090d 792 0 2025-04-30 03:34:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7ff7f69886 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-20-13 calico-apiserver-7ff7f69886-dgpcd eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calie0f68c29924 [] []}} ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:08.999 [INFO][5095] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.128 [INFO][5125] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" HandleID="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.145 [INFO][5125] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" HandleID="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000fd510), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-20-13", "pod":"calico-apiserver-7ff7f69886-dgpcd", "timestamp":"2025-04-30 03:35:09.127986103 +0000 UTC"}, Hostname:"ip-172-31-20-13", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.145 [INFO][5125] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.145 [INFO][5125] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.145 [INFO][5125] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-20-13' Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.149 [INFO][5125] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.155 [INFO][5125] ipam/ipam.go 372: Looking up existing affinities for host host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.161 [INFO][5125] ipam/ipam.go 489: Trying affinity for 192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.163 [INFO][5125] ipam/ipam.go 155: Attempting to load block cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.166 [INFO][5125] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.166 [INFO][5125] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.23.0/26 handle="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.168 [INFO][5125] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.174 [INFO][5125] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.23.0/26 handle="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.188 [INFO][5125] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.23.4/26] block=192.168.23.0/26 handle="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.188 [INFO][5125] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.23.4/26] handle="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" host="ip-172-31-20-13" Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.189 [INFO][5125] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:09.246728 containerd[1992]: 2025-04-30 03:35:09.189 [INFO][5125] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.23.4/26] IPv6=[] ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" HandleID="k8s-pod-network.5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:09.259642 containerd[1992]: 2025-04-30 03:35:09.195 [INFO][5095] cni-plugin/k8s.go 386: Populated endpoint ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"bd94215e-67e1-475c-ae9b-d543ead9090d", ResourceVersion:"792", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"", Pod:"calico-apiserver-7ff7f69886-dgpcd", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie0f68c29924", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:09.259642 containerd[1992]: 2025-04-30 03:35:09.195 [INFO][5095] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.23.4/32] ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:09.259642 containerd[1992]: 2025-04-30 03:35:09.195 [INFO][5095] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie0f68c29924 ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:09.259642 containerd[1992]: 2025-04-30 03:35:09.203 [INFO][5095] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:09.259642 containerd[1992]: 2025-04-30 03:35:09.206 [INFO][5095] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"bd94215e-67e1-475c-ae9b-d543ead9090d", ResourceVersion:"792", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d", Pod:"calico-apiserver-7ff7f69886-dgpcd", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie0f68c29924", MAC:"6a:5a:9b:95:97:92", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:09.259642 containerd[1992]: 2025-04-30 03:35:09.238 [INFO][5095] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-dgpcd" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:09.343648 containerd[1992]: time="2025-04-30T03:35:09.343541265Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:35:09.343648 containerd[1992]: time="2025-04-30T03:35:09.343602324Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:35:09.347296 containerd[1992]: time="2025-04-30T03:35:09.347216010Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:09.349400 containerd[1992]: time="2025-04-30T03:35:09.349174731Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:09.350615 systemd-networkd[1829]: cali22c71cb6299: Link UP Apr 30 03:35:09.354298 systemd-networkd[1829]: cali22c71cb6299: Gained carrier Apr 30 03:35:09.396250 systemd[1]: Started cri-containerd-5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d.scope - libcontainer container 5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d. Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.004 [INFO][5104] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0 calico-kube-controllers-7bf8b64b7f- calico-system 3ed29fa0-99c1-4495-bb89-f321fae4f034 797 0 2025-04-30 03:34:44 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:7bf8b64b7f projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ip-172-31-20-13 calico-kube-controllers-7bf8b64b7f-sxxbz eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali22c71cb6299 [] []}} ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.005 [INFO][5104] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.143 [INFO][5130] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" HandleID="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.155 [INFO][5130] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" HandleID="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003b7460), Attrs:map[string]string{"namespace":"calico-system", "node":"ip-172-31-20-13", "pod":"calico-kube-controllers-7bf8b64b7f-sxxbz", "timestamp":"2025-04-30 03:35:09.143030889 +0000 UTC"}, Hostname:"ip-172-31-20-13", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.155 [INFO][5130] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.190 [INFO][5130] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.190 [INFO][5130] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-20-13' Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.258 [INFO][5130] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.277 [INFO][5130] ipam/ipam.go 372: Looking up existing affinities for host host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.284 [INFO][5130] ipam/ipam.go 489: Trying affinity for 192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.288 [INFO][5130] ipam/ipam.go 155: Attempting to load block cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.297 [INFO][5130] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.298 [INFO][5130] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.23.0/26 handle="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.302 [INFO][5130] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.319 [INFO][5130] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.23.0/26 handle="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.335 [INFO][5130] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.23.5/26] block=192.168.23.0/26 handle="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.335 [INFO][5130] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.23.5/26] handle="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" host="ip-172-31-20-13" Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.335 [INFO][5130] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:09.398181 containerd[1992]: 2025-04-30 03:35:09.336 [INFO][5130] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.23.5/26] IPv6=[] ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" HandleID="k8s-pod-network.a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:09.398925 containerd[1992]: 2025-04-30 03:35:09.344 [INFO][5104] cni-plugin/k8s.go 386: Populated endpoint ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0", GenerateName:"calico-kube-controllers-7bf8b64b7f-", Namespace:"calico-system", SelfLink:"", UID:"3ed29fa0-99c1-4495-bb89-f321fae4f034", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7bf8b64b7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"", Pod:"calico-kube-controllers-7bf8b64b7f-sxxbz", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.23.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali22c71cb6299", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:09.398925 containerd[1992]: 2025-04-30 03:35:09.344 [INFO][5104] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.23.5/32] ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:09.398925 containerd[1992]: 2025-04-30 03:35:09.344 [INFO][5104] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali22c71cb6299 ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:09.398925 containerd[1992]: 2025-04-30 03:35:09.355 [INFO][5104] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:09.398925 containerd[1992]: 2025-04-30 03:35:09.356 [INFO][5104] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0", GenerateName:"calico-kube-controllers-7bf8b64b7f-", Namespace:"calico-system", SelfLink:"", UID:"3ed29fa0-99c1-4495-bb89-f321fae4f034", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7bf8b64b7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca", Pod:"calico-kube-controllers-7bf8b64b7f-sxxbz", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.23.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali22c71cb6299", MAC:"1e:67:08:9c:8b:8e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:09.398925 containerd[1992]: 2025-04-30 03:35:09.380 [INFO][5104] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca" Namespace="calico-system" Pod="calico-kube-controllers-7bf8b64b7f-sxxbz" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:09.428579 containerd[1992]: time="2025-04-30T03:35:09.428147438Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:35:09.428579 containerd[1992]: time="2025-04-30T03:35:09.428224251Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:35:09.428579 containerd[1992]: time="2025-04-30T03:35:09.428252678Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:09.428579 containerd[1992]: time="2025-04-30T03:35:09.428438931Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:09.459673 systemd[1]: Started cri-containerd-a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca.scope - libcontainer container a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca. Apr 30 03:35:09.505445 systemd-networkd[1829]: calibb075411776: Gained IPv6LL Apr 30 03:35:09.505719 systemd-networkd[1829]: vxlan.calico: Link UP Apr 30 03:35:09.505725 systemd-networkd[1829]: vxlan.calico: Gained carrier Apr 30 03:35:09.550067 kubelet[3187]: I0430 03:35:09.548989 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-xfzg8" podStartSLOduration=32.548965351 podStartE2EDuration="32.548965351s" podCreationTimestamp="2025-04-30 03:34:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 03:35:09.545978689 +0000 UTC m=+40.441266325" watchObservedRunningTime="2025-04-30 03:35:09.548965351 +0000 UTC m=+40.444252987" Apr 30 03:35:09.575107 containerd[1992]: time="2025-04-30T03:35:09.574868616Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-dgpcd,Uid:bd94215e-67e1-475c-ae9b-d543ead9090d,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d\"" Apr 30 03:35:09.641881 (udev-worker)[4770]: Network interface NamePolicy= disabled on kernel command line. Apr 30 03:35:09.695917 containerd[1992]: time="2025-04-30T03:35:09.695518281Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7bf8b64b7f-sxxbz,Uid:3ed29fa0-99c1-4495-bb89-f321fae4f034,Namespace:calico-system,Attempt:1,} returns sandbox id \"a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca\"" Apr 30 03:35:09.697527 systemd-networkd[1829]: calia9b1461fcde: Gained IPv6LL Apr 30 03:35:10.219513 containerd[1992]: time="2025-04-30T03:35:10.219435964Z" level=info msg="StopPodSandbox for \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\"" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.273 [INFO][5338] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.273 [INFO][5338] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" iface="eth0" netns="/var/run/netns/cni-52402fd0-a058-2c7e-6483-6e237a65adb8" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.274 [INFO][5338] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" iface="eth0" netns="/var/run/netns/cni-52402fd0-a058-2c7e-6483-6e237a65adb8" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.275 [INFO][5338] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" iface="eth0" netns="/var/run/netns/cni-52402fd0-a058-2c7e-6483-6e237a65adb8" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.275 [INFO][5338] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.275 [INFO][5338] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.307 [INFO][5345] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.307 [INFO][5345] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.307 [INFO][5345] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.321 [WARNING][5345] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.321 [INFO][5345] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.323 [INFO][5345] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:10.327237 containerd[1992]: 2025-04-30 03:35:10.325 [INFO][5338] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:10.328296 containerd[1992]: time="2025-04-30T03:35:10.328137377Z" level=info msg="TearDown network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\" successfully" Apr 30 03:35:10.328296 containerd[1992]: time="2025-04-30T03:35:10.328175932Z" level=info msg="StopPodSandbox for \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\" returns successfully" Apr 30 03:35:10.330475 containerd[1992]: time="2025-04-30T03:35:10.329272190Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-wjp2q,Uid:b847a159-42eb-41a9-96eb-3e5925320f2d,Namespace:calico-apiserver,Attempt:1,}" Apr 30 03:35:10.332228 systemd[1]: run-netns-cni\x2d52402fd0\x2da058\x2d2c7e\x2d6483\x2d6e237a65adb8.mount: Deactivated successfully. Apr 30 03:35:10.528310 systemd-networkd[1829]: calie0f68c29924: Gained IPv6LL Apr 30 03:35:10.605398 systemd-networkd[1829]: calie7225df2991: Link UP Apr 30 03:35:10.608397 systemd-networkd[1829]: calie7225df2991: Gained carrier Apr 30 03:35:10.608513 (udev-worker)[5279]: Network interface NamePolicy= disabled on kernel command line. Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.454 [INFO][5353] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0 calico-apiserver-7ff7f69886- calico-apiserver b847a159-42eb-41a9-96eb-3e5925320f2d 827 0 2025-04-30 03:34:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7ff7f69886 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ip-172-31-20-13 calico-apiserver-7ff7f69886-wjp2q eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calie7225df2991 [] []}} ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.454 [INFO][5353] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.504 [INFO][5365] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" HandleID="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.519 [INFO][5365] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" HandleID="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000291760), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ip-172-31-20-13", "pod":"calico-apiserver-7ff7f69886-wjp2q", "timestamp":"2025-04-30 03:35:10.504582714 +0000 UTC"}, Hostname:"ip-172-31-20-13", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.519 [INFO][5365] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.519 [INFO][5365] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.519 [INFO][5365] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ip-172-31-20-13' Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.524 [INFO][5365] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.535 [INFO][5365] ipam/ipam.go 372: Looking up existing affinities for host host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.543 [INFO][5365] ipam/ipam.go 489: Trying affinity for 192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.546 [INFO][5365] ipam/ipam.go 155: Attempting to load block cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.550 [INFO][5365] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.23.0/26 host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.550 [INFO][5365] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.23.0/26 handle="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.552 [INFO][5365] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.561 [INFO][5365] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.23.0/26 handle="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.590 [INFO][5365] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.23.6/26] block=192.168.23.0/26 handle="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.590 [INFO][5365] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.23.6/26] handle="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" host="ip-172-31-20-13" Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.590 [INFO][5365] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:10.643929 containerd[1992]: 2025-04-30 03:35:10.590 [INFO][5365] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.23.6/26] IPv6=[] ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" HandleID="k8s-pod-network.7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.645366 containerd[1992]: 2025-04-30 03:35:10.596 [INFO][5353] cni-plugin/k8s.go 386: Populated endpoint ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"b847a159-42eb-41a9-96eb-3e5925320f2d", ResourceVersion:"827", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"", Pod:"calico-apiserver-7ff7f69886-wjp2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie7225df2991", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:10.645366 containerd[1992]: 2025-04-30 03:35:10.597 [INFO][5353] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.23.6/32] ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.645366 containerd[1992]: 2025-04-30 03:35:10.597 [INFO][5353] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie7225df2991 ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.645366 containerd[1992]: 2025-04-30 03:35:10.609 [INFO][5353] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.645366 containerd[1992]: 2025-04-30 03:35:10.612 [INFO][5353] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"b847a159-42eb-41a9-96eb-3e5925320f2d", ResourceVersion:"827", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc", Pod:"calico-apiserver-7ff7f69886-wjp2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie7225df2991", MAC:"6a:fa:93:e2:b2:4b", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:10.645366 containerd[1992]: 2025-04-30 03:35:10.639 [INFO][5353] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc" Namespace="calico-apiserver" Pod="calico-apiserver-7ff7f69886-wjp2q" WorkloadEndpoint="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:10.697069 containerd[1992]: time="2025-04-30T03:35:10.694705181Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 03:35:10.697069 containerd[1992]: time="2025-04-30T03:35:10.694806481Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 03:35:10.697069 containerd[1992]: time="2025-04-30T03:35:10.694823230Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:10.697069 containerd[1992]: time="2025-04-30T03:35:10.694934318Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 03:35:10.735654 systemd[1]: Started cri-containerd-7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc.scope - libcontainer container 7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc. Apr 30 03:35:10.816619 containerd[1992]: time="2025-04-30T03:35:10.816365197Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7ff7f69886-wjp2q,Uid:b847a159-42eb-41a9-96eb-3e5925320f2d,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc\"" Apr 30 03:35:10.844557 containerd[1992]: time="2025-04-30T03:35:10.844504192Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:10.846379 containerd[1992]: time="2025-04-30T03:35:10.846323481Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.3: active requests=0, bytes read=7912898" Apr 30 03:35:10.848858 containerd[1992]: time="2025-04-30T03:35:10.848787992Z" level=info msg="ImageCreate event name:\"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:10.852977 containerd[1992]: time="2025-04-30T03:35:10.852388739Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:10.852977 containerd[1992]: time="2025-04-30T03:35:10.852853669Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.3\" with image id \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\", size \"9405520\" in 2.030411114s" Apr 30 03:35:10.852977 containerd[1992]: time="2025-04-30T03:35:10.852884462Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\"" Apr 30 03:35:10.854707 containerd[1992]: time="2025-04-30T03:35:10.854661888Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" Apr 30 03:35:10.856966 containerd[1992]: time="2025-04-30T03:35:10.856926725Z" level=info msg="CreateContainer within sandbox \"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Apr 30 03:35:10.881387 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2939761157.mount: Deactivated successfully. Apr 30 03:35:10.887449 containerd[1992]: time="2025-04-30T03:35:10.887327759Z" level=info msg="CreateContainer within sandbox \"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"25425fece5b920d98ae762ad98d4d001ac131a821bd6fa0fa19f4d1ba56538b8\"" Apr 30 03:35:10.890220 containerd[1992]: time="2025-04-30T03:35:10.888118911Z" level=info msg="StartContainer for \"25425fece5b920d98ae762ad98d4d001ac131a821bd6fa0fa19f4d1ba56538b8\"" Apr 30 03:35:10.924199 systemd[1]: Started cri-containerd-25425fece5b920d98ae762ad98d4d001ac131a821bd6fa0fa19f4d1ba56538b8.scope - libcontainer container 25425fece5b920d98ae762ad98d4d001ac131a821bd6fa0fa19f4d1ba56538b8. Apr 30 03:35:10.959862 containerd[1992]: time="2025-04-30T03:35:10.959810710Z" level=info msg="StartContainer for \"25425fece5b920d98ae762ad98d4d001ac131a821bd6fa0fa19f4d1ba56538b8\" returns successfully" Apr 30 03:35:11.040712 systemd-networkd[1829]: cali22c71cb6299: Gained IPv6LL Apr 30 03:35:11.232155 systemd-networkd[1829]: vxlan.calico: Gained IPv6LL Apr 30 03:35:12.002442 systemd-networkd[1829]: calie7225df2991: Gained IPv6LL Apr 30 03:35:13.529548 kubelet[3187]: I0430 03:35:13.529045 3187 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 03:35:13.569288 systemd[1]: Started sshd@8-172.31.20.13:22-147.75.109.163:35616.service - OpenSSH per-connection server daemon (147.75.109.163:35616). Apr 30 03:35:13.848543 systemd[1]: run-containerd-runc-k8s.io-9709400f2d62de8cdeb9c90dadcf69791774657e58b7c8331fa93adacf9a15d2-runc.rC1MZA.mount: Deactivated successfully. Apr 30 03:35:13.903426 sshd[5469]: Accepted publickey for core from 147.75.109.163 port 35616 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:13.909272 sshd[5469]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:13.929043 systemd-logind[1970]: New session 9 of user core. Apr 30 03:35:13.935607 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 30 03:35:13.971314 containerd[1992]: time="2025-04-30T03:35:13.971258872Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:13.976436 containerd[1992]: time="2025-04-30T03:35:13.976276699Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=43021437" Apr 30 03:35:13.978918 containerd[1992]: time="2025-04-30T03:35:13.978858902Z" level=info msg="ImageCreate event name:\"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:13.983824 containerd[1992]: time="2025-04-30T03:35:13.983713853Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:13.986020 containerd[1992]: time="2025-04-30T03:35:13.985286796Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 3.130078274s" Apr 30 03:35:13.986020 containerd[1992]: time="2025-04-30T03:35:13.985349828Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" Apr 30 03:35:13.987531 containerd[1992]: time="2025-04-30T03:35:13.987283595Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" Apr 30 03:35:14.021510 containerd[1992]: time="2025-04-30T03:35:14.021411690Z" level=info msg="CreateContainer within sandbox \"5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 30 03:35:14.040992 containerd[1992]: time="2025-04-30T03:35:14.040946651Z" level=info msg="CreateContainer within sandbox \"5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"e247dd0b17296c3fe8947e9474afaacadc80ebdeeadf66e4c63bb49e8f1d1af6\"" Apr 30 03:35:14.042839 containerd[1992]: time="2025-04-30T03:35:14.041593672Z" level=info msg="StartContainer for \"e247dd0b17296c3fe8947e9474afaacadc80ebdeeadf66e4c63bb49e8f1d1af6\"" Apr 30 03:35:14.087260 systemd[1]: Started cri-containerd-e247dd0b17296c3fe8947e9474afaacadc80ebdeeadf66e4c63bb49e8f1d1af6.scope - libcontainer container e247dd0b17296c3fe8947e9474afaacadc80ebdeeadf66e4c63bb49e8f1d1af6. Apr 30 03:35:14.161564 containerd[1992]: time="2025-04-30T03:35:14.161387259Z" level=info msg="StartContainer for \"e247dd0b17296c3fe8947e9474afaacadc80ebdeeadf66e4c63bb49e8f1d1af6\" returns successfully" Apr 30 03:35:14.565316 sshd[5469]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:14.570031 systemd-logind[1970]: Session 9 logged out. Waiting for processes to exit. Apr 30 03:35:14.573429 systemd[1]: sshd@8-172.31.20.13:22-147.75.109.163:35616.service: Deactivated successfully. Apr 30 03:35:14.577639 systemd[1]: session-9.scope: Deactivated successfully. Apr 30 03:35:14.579491 systemd-logind[1970]: Removed session 9. Apr 30 03:35:14.629416 kubelet[3187]: I0430 03:35:14.629345 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7ff7f69886-dgpcd" podStartSLOduration=26.221848316 podStartE2EDuration="30.629320864s" podCreationTimestamp="2025-04-30 03:34:44 +0000 UTC" firstStartedPulling="2025-04-30 03:35:09.57966308 +0000 UTC m=+40.474950706" lastFinishedPulling="2025-04-30 03:35:13.987135627 +0000 UTC m=+44.882423254" observedRunningTime="2025-04-30 03:35:14.626982167 +0000 UTC m=+45.522269802" watchObservedRunningTime="2025-04-30 03:35:14.629320864 +0000 UTC m=+45.524608499" Apr 30 03:35:14.987384 ntpd[1964]: Listen normally on 7 vxlan.calico 192.168.23.0:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 7 vxlan.calico 192.168.23.0:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 8 calibb075411776 [fe80::ecee:eeff:feee:eeee%4]:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 9 cali22e4b179ea0 [fe80::ecee:eeff:feee:eeee%5]:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 10 calia9b1461fcde [fe80::ecee:eeff:feee:eeee%6]:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 11 calie0f68c29924 [fe80::ecee:eeff:feee:eeee%7]:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 12 cali22c71cb6299 [fe80::ecee:eeff:feee:eeee%8]:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 13 vxlan.calico [fe80::64d3:43ff:fea2:e926%9]:123 Apr 30 03:35:14.988528 ntpd[1964]: 30 Apr 03:35:14 ntpd[1964]: Listen normally on 14 calie7225df2991 [fe80::ecee:eeff:feee:eeee%12]:123 Apr 30 03:35:14.987480 ntpd[1964]: Listen normally on 8 calibb075411776 [fe80::ecee:eeff:feee:eeee%4]:123 Apr 30 03:35:14.987538 ntpd[1964]: Listen normally on 9 cali22e4b179ea0 [fe80::ecee:eeff:feee:eeee%5]:123 Apr 30 03:35:14.987582 ntpd[1964]: Listen normally on 10 calia9b1461fcde [fe80::ecee:eeff:feee:eeee%6]:123 Apr 30 03:35:14.987619 ntpd[1964]: Listen normally on 11 calie0f68c29924 [fe80::ecee:eeff:feee:eeee%7]:123 Apr 30 03:35:14.987652 ntpd[1964]: Listen normally on 12 cali22c71cb6299 [fe80::ecee:eeff:feee:eeee%8]:123 Apr 30 03:35:14.987686 ntpd[1964]: Listen normally on 13 vxlan.calico [fe80::64d3:43ff:fea2:e926%9]:123 Apr 30 03:35:14.987723 ntpd[1964]: Listen normally on 14 calie7225df2991 [fe80::ecee:eeff:feee:eeee%12]:123 Apr 30 03:35:15.638267 kubelet[3187]: I0430 03:35:15.638114 3187 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 03:35:16.529246 containerd[1992]: time="2025-04-30T03:35:16.529191002Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:16.530873 containerd[1992]: time="2025-04-30T03:35:16.530822385Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.3: active requests=0, bytes read=34789138" Apr 30 03:35:16.532565 containerd[1992]: time="2025-04-30T03:35:16.531329720Z" level=info msg="ImageCreate event name:\"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:16.533687 containerd[1992]: time="2025-04-30T03:35:16.533651401Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:16.534398 containerd[1992]: time="2025-04-30T03:35:16.534358545Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" with image id \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\", size \"36281728\" in 2.547020085s" Apr 30 03:35:16.534579 containerd[1992]: time="2025-04-30T03:35:16.534558869Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\"" Apr 30 03:35:16.535772 containerd[1992]: time="2025-04-30T03:35:16.535753122Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" Apr 30 03:35:16.603259 containerd[1992]: time="2025-04-30T03:35:16.602166101Z" level=info msg="CreateContainer within sandbox \"a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Apr 30 03:35:16.624691 containerd[1992]: time="2025-04-30T03:35:16.624644503Z" level=info msg="CreateContainer within sandbox \"a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2\"" Apr 30 03:35:16.625466 containerd[1992]: time="2025-04-30T03:35:16.625424779Z" level=info msg="StartContainer for \"b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2\"" Apr 30 03:35:16.688206 systemd[1]: Started cri-containerd-b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2.scope - libcontainer container b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2. Apr 30 03:35:16.762436 containerd[1992]: time="2025-04-30T03:35:16.761440622Z" level=info msg="StartContainer for \"b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2\" returns successfully" Apr 30 03:35:16.965997 containerd[1992]: time="2025-04-30T03:35:16.965858748Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:16.968043 containerd[1992]: time="2025-04-30T03:35:16.966938204Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=77" Apr 30 03:35:16.969289 containerd[1992]: time="2025-04-30T03:35:16.969221521Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 433.087379ms" Apr 30 03:35:16.969289 containerd[1992]: time="2025-04-30T03:35:16.969276094Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" Apr 30 03:35:16.972331 containerd[1992]: time="2025-04-30T03:35:16.972039629Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" Apr 30 03:35:16.973203 containerd[1992]: time="2025-04-30T03:35:16.973130260Z" level=info msg="CreateContainer within sandbox \"7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 30 03:35:16.998657 containerd[1992]: time="2025-04-30T03:35:16.998553935Z" level=info msg="CreateContainer within sandbox \"7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"f2cfe7aa0692422658f1ad81ac270abd7ef7167ec6f712db8b1247e63be360aa\"" Apr 30 03:35:16.999571 containerd[1992]: time="2025-04-30T03:35:16.999525192Z" level=info msg="StartContainer for \"f2cfe7aa0692422658f1ad81ac270abd7ef7167ec6f712db8b1247e63be360aa\"" Apr 30 03:35:17.038227 systemd[1]: Started cri-containerd-f2cfe7aa0692422658f1ad81ac270abd7ef7167ec6f712db8b1247e63be360aa.scope - libcontainer container f2cfe7aa0692422658f1ad81ac270abd7ef7167ec6f712db8b1247e63be360aa. Apr 30 03:35:17.090441 containerd[1992]: time="2025-04-30T03:35:17.090367596Z" level=info msg="StartContainer for \"f2cfe7aa0692422658f1ad81ac270abd7ef7167ec6f712db8b1247e63be360aa\" returns successfully" Apr 30 03:35:17.706740 kubelet[3187]: I0430 03:35:17.706214 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7ff7f69886-wjp2q" podStartSLOduration=27.555983824 podStartE2EDuration="33.706191016s" podCreationTimestamp="2025-04-30 03:34:44 +0000 UTC" firstStartedPulling="2025-04-30 03:35:10.819809906 +0000 UTC m=+41.715097520" lastFinishedPulling="2025-04-30 03:35:16.970017087 +0000 UTC m=+47.865304712" observedRunningTime="2025-04-30 03:35:17.702539509 +0000 UTC m=+48.597827144" watchObservedRunningTime="2025-04-30 03:35:17.706191016 +0000 UTC m=+48.601478651" Apr 30 03:35:17.805966 kubelet[3187]: I0430 03:35:17.805912 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-7bf8b64b7f-sxxbz" podStartSLOduration=26.969683896 podStartE2EDuration="33.805898518s" podCreationTimestamp="2025-04-30 03:34:44 +0000 UTC" firstStartedPulling="2025-04-30 03:35:09.699335382 +0000 UTC m=+40.594622995" lastFinishedPulling="2025-04-30 03:35:16.53554997 +0000 UTC m=+47.430837617" observedRunningTime="2025-04-30 03:35:17.735049655 +0000 UTC m=+48.630337290" watchObservedRunningTime="2025-04-30 03:35:17.805898518 +0000 UTC m=+48.701186153" Apr 30 03:35:18.824039 containerd[1992]: time="2025-04-30T03:35:18.823079837Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:18.827447 containerd[1992]: time="2025-04-30T03:35:18.827377950Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3: active requests=0, bytes read=13991773" Apr 30 03:35:18.831696 containerd[1992]: time="2025-04-30T03:35:18.831615971Z" level=info msg="ImageCreate event name:\"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:18.838215 containerd[1992]: time="2025-04-30T03:35:18.837669885Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 03:35:18.840455 containerd[1992]: time="2025-04-30T03:35:18.840196922Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" with image id \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\", size \"15484347\" in 1.868014017s" Apr 30 03:35:18.841881 containerd[1992]: time="2025-04-30T03:35:18.840327502Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\"" Apr 30 03:35:18.845222 containerd[1992]: time="2025-04-30T03:35:18.845180559Z" level=info msg="CreateContainer within sandbox \"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Apr 30 03:35:18.905343 containerd[1992]: time="2025-04-30T03:35:18.905214883Z" level=info msg="CreateContainer within sandbox \"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"12558729a2e4cbad01463dfef1bcc52544077d2b8faee8ac1fb58c15d877c0d9\"" Apr 30 03:35:18.913024 containerd[1992]: time="2025-04-30T03:35:18.908236398Z" level=info msg="StartContainer for \"12558729a2e4cbad01463dfef1bcc52544077d2b8faee8ac1fb58c15d877c0d9\"" Apr 30 03:35:18.980711 systemd[1]: Started cri-containerd-12558729a2e4cbad01463dfef1bcc52544077d2b8faee8ac1fb58c15d877c0d9.scope - libcontainer container 12558729a2e4cbad01463dfef1bcc52544077d2b8faee8ac1fb58c15d877c0d9. Apr 30 03:35:19.032956 containerd[1992]: time="2025-04-30T03:35:19.032912860Z" level=info msg="StartContainer for \"12558729a2e4cbad01463dfef1bcc52544077d2b8faee8ac1fb58c15d877c0d9\" returns successfully" Apr 30 03:35:19.624390 systemd[1]: Started sshd@9-172.31.20.13:22-147.75.109.163:51430.service - OpenSSH per-connection server daemon (147.75.109.163:51430). Apr 30 03:35:19.692747 kubelet[3187]: I0430 03:35:19.688420 3187 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Apr 30 03:35:19.693276 kubelet[3187]: I0430 03:35:19.693232 3187 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Apr 30 03:35:19.912203 sshd[5723]: Accepted publickey for core from 147.75.109.163 port 51430 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:19.915697 sshd[5723]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:19.921934 systemd-logind[1970]: New session 10 of user core. Apr 30 03:35:19.928269 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 30 03:35:20.777711 sshd[5723]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:20.781387 systemd-logind[1970]: Session 10 logged out. Waiting for processes to exit. Apr 30 03:35:20.781746 systemd[1]: sshd@9-172.31.20.13:22-147.75.109.163:51430.service: Deactivated successfully. Apr 30 03:35:20.784352 systemd[1]: session-10.scope: Deactivated successfully. Apr 30 03:35:20.785398 systemd-logind[1970]: Removed session 10. Apr 30 03:35:20.827426 systemd[1]: Started sshd@10-172.31.20.13:22-147.75.109.163:51440.service - OpenSSH per-connection server daemon (147.75.109.163:51440). Apr 30 03:35:21.085740 sshd[5751]: Accepted publickey for core from 147.75.109.163 port 51440 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:21.087313 sshd[5751]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:21.092576 systemd-logind[1970]: New session 11 of user core. Apr 30 03:35:21.095245 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 30 03:35:21.489010 sshd[5751]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:21.493269 systemd[1]: sshd@10-172.31.20.13:22-147.75.109.163:51440.service: Deactivated successfully. Apr 30 03:35:21.497180 systemd[1]: session-11.scope: Deactivated successfully. Apr 30 03:35:21.499289 systemd-logind[1970]: Session 11 logged out. Waiting for processes to exit. Apr 30 03:35:21.502028 systemd-logind[1970]: Removed session 11. Apr 30 03:35:21.533476 systemd[1]: Started sshd@11-172.31.20.13:22-147.75.109.163:51442.service - OpenSSH per-connection server daemon (147.75.109.163:51442). Apr 30 03:35:21.808549 sshd[5762]: Accepted publickey for core from 147.75.109.163 port 51442 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:21.810033 sshd[5762]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:21.815637 systemd-logind[1970]: New session 12 of user core. Apr 30 03:35:21.818248 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 30 03:35:22.083230 sshd[5762]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:22.087224 systemd[1]: sshd@11-172.31.20.13:22-147.75.109.163:51442.service: Deactivated successfully. Apr 30 03:35:22.089244 systemd[1]: session-12.scope: Deactivated successfully. Apr 30 03:35:22.090102 systemd-logind[1970]: Session 12 logged out. Waiting for processes to exit. Apr 30 03:35:22.091231 systemd-logind[1970]: Removed session 12. Apr 30 03:35:27.131110 systemd[1]: Started sshd@12-172.31.20.13:22-147.75.109.163:34452.service - OpenSSH per-connection server daemon (147.75.109.163:34452). Apr 30 03:35:27.393050 sshd[5775]: Accepted publickey for core from 147.75.109.163 port 34452 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:27.394612 sshd[5775]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:27.399475 systemd-logind[1970]: New session 13 of user core. Apr 30 03:35:27.403206 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 30 03:35:27.686681 sshd[5775]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:27.692126 systemd[1]: sshd@12-172.31.20.13:22-147.75.109.163:34452.service: Deactivated successfully. Apr 30 03:35:27.694734 systemd[1]: session-13.scope: Deactivated successfully. Apr 30 03:35:27.697364 systemd-logind[1970]: Session 13 logged out. Waiting for processes to exit. Apr 30 03:35:27.698827 systemd-logind[1970]: Removed session 13. Apr 30 03:35:29.260675 containerd[1992]: time="2025-04-30T03:35:29.260366920Z" level=info msg="StopPodSandbox for \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\"" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.496 [WARNING][5806] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3", ResourceVersion:"914", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af", Pod:"csi-node-driver-p2dkq", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.23.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia9b1461fcde", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.501 [INFO][5806] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.501 [INFO][5806] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" iface="eth0" netns="" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.501 [INFO][5806] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.501 [INFO][5806] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.528 [INFO][5813] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.528 [INFO][5813] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.528 [INFO][5813] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.534 [WARNING][5813] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.534 [INFO][5813] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.536 [INFO][5813] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:29.542096 containerd[1992]: 2025-04-30 03:35:29.539 [INFO][5806] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.542897 containerd[1992]: time="2025-04-30T03:35:29.542152114Z" level=info msg="TearDown network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\" successfully" Apr 30 03:35:29.542897 containerd[1992]: time="2025-04-30T03:35:29.542185892Z" level=info msg="StopPodSandbox for \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\" returns successfully" Apr 30 03:35:29.554724 containerd[1992]: time="2025-04-30T03:35:29.554659816Z" level=info msg="RemovePodSandbox for \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\"" Apr 30 03:35:29.554724 containerd[1992]: time="2025-04-30T03:35:29.554724201Z" level=info msg="Forcibly stopping sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\"" Apr 30 03:35:29.593424 kubelet[3187]: I0430 03:35:29.593055 3187 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 03:35:29.626671 kubelet[3187]: I0430 03:35:29.625519 3187 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-p2dkq" podStartSLOduration=35.594471734 podStartE2EDuration="45.625496258s" podCreationTimestamp="2025-04-30 03:34:44 +0000 UTC" firstStartedPulling="2025-04-30 03:35:08.811591336 +0000 UTC m=+39.706878949" lastFinishedPulling="2025-04-30 03:35:18.842615855 +0000 UTC m=+49.737903473" observedRunningTime="2025-04-30 03:35:19.703494898 +0000 UTC m=+50.598782532" watchObservedRunningTime="2025-04-30 03:35:29.625496258 +0000 UTC m=+60.520783896" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.597 [WARNING][5831] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"5e820db8-1ee7-4b3a-bfb7-17c5a434fca3", ResourceVersion:"914", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"f5f492494154c43c63076b8c88e12b704db40a7611f6b6e619370ded298170af", Pod:"csi-node-driver-p2dkq", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.23.3/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calia9b1461fcde", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.597 [INFO][5831] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.597 [INFO][5831] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" iface="eth0" netns="" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.598 [INFO][5831] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.598 [INFO][5831] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.666 [INFO][5838] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.667 [INFO][5838] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.667 [INFO][5838] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.683 [WARNING][5838] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.683 [INFO][5838] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" HandleID="k8s-pod-network.fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Workload="ip--172--31--20--13-k8s-csi--node--driver--p2dkq-eth0" Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.686 [INFO][5838] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:29.695707 containerd[1992]: 2025-04-30 03:35:29.692 [INFO][5831] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529" Apr 30 03:35:29.696509 containerd[1992]: time="2025-04-30T03:35:29.695856247Z" level=info msg="TearDown network for sandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\" successfully" Apr 30 03:35:29.707468 containerd[1992]: time="2025-04-30T03:35:29.707415593Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 03:35:29.737232 containerd[1992]: time="2025-04-30T03:35:29.737166270Z" level=info msg="RemovePodSandbox \"fbb1ba2f5e68e19e3e201eee348cae118e507ed5c8658e068e4bea7c431be529\" returns successfully" Apr 30 03:35:29.778918 containerd[1992]: time="2025-04-30T03:35:29.778882632Z" level=info msg="StopPodSandbox for \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\"" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.824 [WARNING][5858] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"bd94215e-67e1-475c-ae9b-d543ead9090d", ResourceVersion:"970", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d", Pod:"calico-apiserver-7ff7f69886-dgpcd", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie0f68c29924", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.825 [INFO][5858] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.825 [INFO][5858] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" iface="eth0" netns="" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.825 [INFO][5858] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.825 [INFO][5858] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.883 [INFO][5865] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.884 [INFO][5865] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.884 [INFO][5865] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.905 [WARNING][5865] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.906 [INFO][5865] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.910 [INFO][5865] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:29.916840 containerd[1992]: 2025-04-30 03:35:29.913 [INFO][5858] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:29.916840 containerd[1992]: time="2025-04-30T03:35:29.916203816Z" level=info msg="TearDown network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\" successfully" Apr 30 03:35:29.916840 containerd[1992]: time="2025-04-30T03:35:29.916235476Z" level=info msg="StopPodSandbox for \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\" returns successfully" Apr 30 03:35:29.923730 containerd[1992]: time="2025-04-30T03:35:29.921582951Z" level=info msg="RemovePodSandbox for \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\"" Apr 30 03:35:29.923730 containerd[1992]: time="2025-04-30T03:35:29.921633237Z" level=info msg="Forcibly stopping sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\"" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:29.987 [WARNING][5884] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"bd94215e-67e1-475c-ae9b-d543ead9090d", ResourceVersion:"970", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"5bc3c5ed1252ff135b88f9d051b16cf67308fd76c0b646ad160e54d33345f47d", Pod:"calico-apiserver-7ff7f69886-dgpcd", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.4/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie0f68c29924", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:29.987 [INFO][5884] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:29.987 [INFO][5884] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" iface="eth0" netns="" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:29.987 [INFO][5884] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:29.987 [INFO][5884] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:30.029 [INFO][5891] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:30.030 [INFO][5891] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:30.030 [INFO][5891] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:30.043 [WARNING][5891] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:30.043 [INFO][5891] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" HandleID="k8s-pod-network.74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--dgpcd-eth0" Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:30.045 [INFO][5891] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.050961 containerd[1992]: 2025-04-30 03:35:30.049 [INFO][5884] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af" Apr 30 03:35:30.051685 containerd[1992]: time="2025-04-30T03:35:30.051162878Z" level=info msg="TearDown network for sandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\" successfully" Apr 30 03:35:30.060181 containerd[1992]: time="2025-04-30T03:35:30.060110505Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 03:35:30.060336 containerd[1992]: time="2025-04-30T03:35:30.060226192Z" level=info msg="RemovePodSandbox \"74e94f289858fa90269cdd6b2382d9d196fa2185a9d63db006dc1562fe47b8af\" returns successfully" Apr 30 03:35:30.060865 containerd[1992]: time="2025-04-30T03:35:30.060836156Z" level=info msg="StopPodSandbox for \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\"" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.102 [WARNING][5915] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c", Pod:"coredns-668d6bf9bc-rncmj", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calibb075411776", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.102 [INFO][5915] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.102 [INFO][5915] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" iface="eth0" netns="" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.102 [INFO][5915] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.102 [INFO][5915] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.136 [INFO][5922] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.136 [INFO][5922] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.136 [INFO][5922] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.142 [WARNING][5922] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.142 [INFO][5922] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.144 [INFO][5922] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.147772 containerd[1992]: 2025-04-30 03:35:30.146 [INFO][5915] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.148563 containerd[1992]: time="2025-04-30T03:35:30.147785294Z" level=info msg="TearDown network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\" successfully" Apr 30 03:35:30.148563 containerd[1992]: time="2025-04-30T03:35:30.147809421Z" level=info msg="StopPodSandbox for \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\" returns successfully" Apr 30 03:35:30.148563 containerd[1992]: time="2025-04-30T03:35:30.148292905Z" level=info msg="RemovePodSandbox for \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\"" Apr 30 03:35:30.148563 containerd[1992]: time="2025-04-30T03:35:30.148317734Z" level=info msg="Forcibly stopping sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\"" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.192 [WARNING][5940] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"1e9b51f5-da9d-4ed3-aa3d-02aa185b5484", ResourceVersion:"809", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"7c6dbdffa614c5b9b6362bec304e5caa4649b691509915a5bc4520e67649e80c", Pod:"coredns-668d6bf9bc-rncmj", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.1/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calibb075411776", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.192 [INFO][5940] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.193 [INFO][5940] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" iface="eth0" netns="" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.193 [INFO][5940] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.193 [INFO][5940] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.223 [INFO][5947] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.223 [INFO][5947] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.223 [INFO][5947] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.230 [WARNING][5947] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.230 [INFO][5947] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" HandleID="k8s-pod-network.1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--rncmj-eth0" Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.232 [INFO][5947] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.236382 containerd[1992]: 2025-04-30 03:35:30.234 [INFO][5940] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097" Apr 30 03:35:30.237778 containerd[1992]: time="2025-04-30T03:35:30.236408957Z" level=info msg="TearDown network for sandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\" successfully" Apr 30 03:35:30.250225 containerd[1992]: time="2025-04-30T03:35:30.250167493Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 03:35:30.251282 containerd[1992]: time="2025-04-30T03:35:30.250260356Z" level=info msg="RemovePodSandbox \"1bffd9729a9f8a93ac9d4485fb7b975adbafc58fd12bdf4c3454c0a9bad85097\" returns successfully" Apr 30 03:35:30.251282 containerd[1992]: time="2025-04-30T03:35:30.250918806Z" level=info msg="StopPodSandbox for \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\"" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.296 [WARNING][5965] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0", GenerateName:"calico-kube-controllers-7bf8b64b7f-", Namespace:"calico-system", SelfLink:"", UID:"3ed29fa0-99c1-4495-bb89-f321fae4f034", ResourceVersion:"889", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7bf8b64b7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca", Pod:"calico-kube-controllers-7bf8b64b7f-sxxbz", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.23.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali22c71cb6299", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.296 [INFO][5965] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.296 [INFO][5965] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" iface="eth0" netns="" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.296 [INFO][5965] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.296 [INFO][5965] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.323 [INFO][5972] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.323 [INFO][5972] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.323 [INFO][5972] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.330 [WARNING][5972] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.330 [INFO][5972] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.332 [INFO][5972] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.335556 containerd[1992]: 2025-04-30 03:35:30.333 [INFO][5965] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.336271 containerd[1992]: time="2025-04-30T03:35:30.335599112Z" level=info msg="TearDown network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\" successfully" Apr 30 03:35:30.336271 containerd[1992]: time="2025-04-30T03:35:30.335621294Z" level=info msg="StopPodSandbox for \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\" returns successfully" Apr 30 03:35:30.337425 containerd[1992]: time="2025-04-30T03:35:30.336373182Z" level=info msg="RemovePodSandbox for \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\"" Apr 30 03:35:30.337425 containerd[1992]: time="2025-04-30T03:35:30.336397939Z" level=info msg="Forcibly stopping sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\"" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.377 [WARNING][5990] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0", GenerateName:"calico-kube-controllers-7bf8b64b7f-", Namespace:"calico-system", SelfLink:"", UID:"3ed29fa0-99c1-4495-bb89-f321fae4f034", ResourceVersion:"889", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7bf8b64b7f", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"a51bd7b761dfdc8434e52757d762ff4733e0a4ac9c9a78e220c0c302b39a1eca", Pod:"calico-kube-controllers-7bf8b64b7f-sxxbz", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.23.5/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali22c71cb6299", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.377 [INFO][5990] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.377 [INFO][5990] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" iface="eth0" netns="" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.377 [INFO][5990] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.377 [INFO][5990] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.402 [INFO][5997] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.402 [INFO][5997] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.402 [INFO][5997] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.411 [WARNING][5997] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.411 [INFO][5997] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" HandleID="k8s-pod-network.365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Workload="ip--172--31--20--13-k8s-calico--kube--controllers--7bf8b64b7f--sxxbz-eth0" Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.413 [INFO][5997] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.417639 containerd[1992]: 2025-04-30 03:35:30.415 [INFO][5990] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f" Apr 30 03:35:30.418736 containerd[1992]: time="2025-04-30T03:35:30.417697901Z" level=info msg="TearDown network for sandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\" successfully" Apr 30 03:35:30.423113 containerd[1992]: time="2025-04-30T03:35:30.423062199Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 03:35:30.423229 containerd[1992]: time="2025-04-30T03:35:30.423132779Z" level=info msg="RemovePodSandbox \"365b857ded15adc2a1dd1d9b21e66d77eb486d3af360a1cf467b9f49038bd11f\" returns successfully" Apr 30 03:35:30.423953 containerd[1992]: time="2025-04-30T03:35:30.423664635Z" level=info msg="StopPodSandbox for \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\"" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.464 [WARNING][6015] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0e7b462c-a68a-448d-8c44-825a2ecffb4c", ResourceVersion:"812", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583", Pod:"coredns-668d6bf9bc-xfzg8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali22e4b179ea0", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.465 [INFO][6015] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.465 [INFO][6015] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" iface="eth0" netns="" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.465 [INFO][6015] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.465 [INFO][6015] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.491 [INFO][6022] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.491 [INFO][6022] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.491 [INFO][6022] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.499 [WARNING][6022] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.499 [INFO][6022] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.500 [INFO][6022] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.504164 containerd[1992]: 2025-04-30 03:35:30.502 [INFO][6015] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.504164 containerd[1992]: time="2025-04-30T03:35:30.504137060Z" level=info msg="TearDown network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\" successfully" Apr 30 03:35:30.507078 containerd[1992]: time="2025-04-30T03:35:30.504167804Z" level=info msg="StopPodSandbox for \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\" returns successfully" Apr 30 03:35:30.507078 containerd[1992]: time="2025-04-30T03:35:30.506238620Z" level=info msg="RemovePodSandbox for \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\"" Apr 30 03:35:30.507078 containerd[1992]: time="2025-04-30T03:35:30.506276450Z" level=info msg="Forcibly stopping sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\"" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.551 [WARNING][6040] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"0e7b462c-a68a-448d-8c44-825a2ecffb4c", ResourceVersion:"812", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 37, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"b4a2ef62d03c07a92b28a6a46d610d5d092b3c584d5074bdfa1ef327d4dcc583", Pod:"coredns-668d6bf9bc-xfzg8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.23.2/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali22e4b179ea0", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.551 [INFO][6040] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.551 [INFO][6040] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" iface="eth0" netns="" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.551 [INFO][6040] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.551 [INFO][6040] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.578 [INFO][6047] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.579 [INFO][6047] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.579 [INFO][6047] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.585 [WARNING][6047] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.585 [INFO][6047] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" HandleID="k8s-pod-network.6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Workload="ip--172--31--20--13-k8s-coredns--668d6bf9bc--xfzg8-eth0" Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.587 [INFO][6047] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.591521 containerd[1992]: 2025-04-30 03:35:30.589 [INFO][6040] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9" Apr 30 03:35:30.591998 containerd[1992]: time="2025-04-30T03:35:30.591557625Z" level=info msg="TearDown network for sandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\" successfully" Apr 30 03:35:30.597430 containerd[1992]: time="2025-04-30T03:35:30.597368499Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 03:35:30.597588 containerd[1992]: time="2025-04-30T03:35:30.597443762Z" level=info msg="RemovePodSandbox \"6d7ac1a808115c9d5758a7b3ffe3a950abc7b2b499e3d31ffa83d45f7c93a6b9\" returns successfully" Apr 30 03:35:30.597990 containerd[1992]: time="2025-04-30T03:35:30.597951940Z" level=info msg="StopPodSandbox for \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\"" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.636 [WARNING][6066] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"b847a159-42eb-41a9-96eb-3e5925320f2d", ResourceVersion:"893", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc", Pod:"calico-apiserver-7ff7f69886-wjp2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie7225df2991", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.637 [INFO][6066] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.637 [INFO][6066] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" iface="eth0" netns="" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.637 [INFO][6066] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.637 [INFO][6066] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.663 [INFO][6073] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.663 [INFO][6073] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.663 [INFO][6073] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.670 [WARNING][6073] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.670 [INFO][6073] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.672 [INFO][6073] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.677334 containerd[1992]: 2025-04-30 03:35:30.675 [INFO][6066] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.678056 containerd[1992]: time="2025-04-30T03:35:30.677395603Z" level=info msg="TearDown network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\" successfully" Apr 30 03:35:30.678056 containerd[1992]: time="2025-04-30T03:35:30.677426178Z" level=info msg="StopPodSandbox for \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\" returns successfully" Apr 30 03:35:30.678551 containerd[1992]: time="2025-04-30T03:35:30.678520688Z" level=info msg="RemovePodSandbox for \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\"" Apr 30 03:35:30.678645 containerd[1992]: time="2025-04-30T03:35:30.678554676Z" level=info msg="Forcibly stopping sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\"" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.727 [WARNING][6092] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0", GenerateName:"calico-apiserver-7ff7f69886-", Namespace:"calico-apiserver", SelfLink:"", UID:"b847a159-42eb-41a9-96eb-3e5925320f2d", ResourceVersion:"893", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 3, 34, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7ff7f69886", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ip-172-31-20-13", ContainerID:"7d5efb2096769b0089d2b9758a125a54112059443b106c5fcba315c10a9c51bc", Pod:"calico-apiserver-7ff7f69886-wjp2q", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.23.6/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calie7225df2991", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.728 [INFO][6092] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.728 [INFO][6092] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" iface="eth0" netns="" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.728 [INFO][6092] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.728 [INFO][6092] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.753 [INFO][6101] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.753 [INFO][6101] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.753 [INFO][6101] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.763 [WARNING][6101] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.763 [INFO][6101] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" HandleID="k8s-pod-network.5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Workload="ip--172--31--20--13-k8s-calico--apiserver--7ff7f69886--wjp2q-eth0" Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.766 [INFO][6101] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 03:35:30.775827 containerd[1992]: 2025-04-30 03:35:30.772 [INFO][6092] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9" Apr 30 03:35:30.776708 containerd[1992]: time="2025-04-30T03:35:30.775902595Z" level=info msg="TearDown network for sandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\" successfully" Apr 30 03:35:30.782668 containerd[1992]: time="2025-04-30T03:35:30.782613408Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 03:35:30.782796 containerd[1992]: time="2025-04-30T03:35:30.782684765Z" level=info msg="RemovePodSandbox \"5371809dffee0b61da8635f5f9602160c66c2b7b2aea58308507da2686bc75e9\" returns successfully" Apr 30 03:35:31.616682 systemd[1]: run-containerd-runc-k8s.io-b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2-runc.ZEaDjN.mount: Deactivated successfully. Apr 30 03:35:32.734328 systemd[1]: Started sshd@13-172.31.20.13:22-147.75.109.163:34462.service - OpenSSH per-connection server daemon (147.75.109.163:34462). Apr 30 03:35:33.016766 sshd[6127]: Accepted publickey for core from 147.75.109.163 port 34462 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:33.020028 sshd[6127]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:33.025691 systemd-logind[1970]: New session 14 of user core. Apr 30 03:35:33.030236 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 30 03:35:33.481734 sshd[6127]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:33.486573 systemd-logind[1970]: Session 14 logged out. Waiting for processes to exit. Apr 30 03:35:33.487475 systemd[1]: sshd@13-172.31.20.13:22-147.75.109.163:34462.service: Deactivated successfully. Apr 30 03:35:33.490927 systemd[1]: session-14.scope: Deactivated successfully. Apr 30 03:35:33.492134 systemd-logind[1970]: Removed session 14. Apr 30 03:35:38.529051 systemd[1]: Started sshd@14-172.31.20.13:22-147.75.109.163:42032.service - OpenSSH per-connection server daemon (147.75.109.163:42032). Apr 30 03:35:38.778977 sshd[6140]: Accepted publickey for core from 147.75.109.163 port 42032 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:38.780424 sshd[6140]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:38.785098 systemd-logind[1970]: New session 15 of user core. Apr 30 03:35:38.792210 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 30 03:35:39.071090 sshd[6140]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:39.075488 systemd[1]: sshd@14-172.31.20.13:22-147.75.109.163:42032.service: Deactivated successfully. Apr 30 03:35:39.077500 systemd[1]: session-15.scope: Deactivated successfully. Apr 30 03:35:39.078260 systemd-logind[1970]: Session 15 logged out. Waiting for processes to exit. Apr 30 03:35:39.079510 systemd-logind[1970]: Removed session 15. Apr 30 03:35:39.119698 systemd[1]: Started sshd@15-172.31.20.13:22-147.75.109.163:42044.service - OpenSSH per-connection server daemon (147.75.109.163:42044). Apr 30 03:35:39.377812 sshd[6155]: Accepted publickey for core from 147.75.109.163 port 42044 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:39.379352 sshd[6155]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:39.384207 systemd-logind[1970]: New session 16 of user core. Apr 30 03:35:39.390190 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 30 03:35:40.146887 sshd[6155]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:40.154831 systemd[1]: sshd@15-172.31.20.13:22-147.75.109.163:42044.service: Deactivated successfully. Apr 30 03:35:40.156753 systemd[1]: session-16.scope: Deactivated successfully. Apr 30 03:35:40.157652 systemd-logind[1970]: Session 16 logged out. Waiting for processes to exit. Apr 30 03:35:40.158898 systemd-logind[1970]: Removed session 16. Apr 30 03:35:40.200443 systemd[1]: Started sshd@16-172.31.20.13:22-147.75.109.163:42060.service - OpenSSH per-connection server daemon (147.75.109.163:42060). Apr 30 03:35:40.459211 sshd[6166]: Accepted publickey for core from 147.75.109.163 port 42060 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:40.461427 sshd[6166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:40.470173 systemd-logind[1970]: New session 17 of user core. Apr 30 03:35:40.481236 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 30 03:35:41.588139 sshd[6166]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:41.592770 systemd[1]: sshd@16-172.31.20.13:22-147.75.109.163:42060.service: Deactivated successfully. Apr 30 03:35:41.597132 systemd[1]: session-17.scope: Deactivated successfully. Apr 30 03:35:41.600036 systemd-logind[1970]: Session 17 logged out. Waiting for processes to exit. Apr 30 03:35:41.603972 systemd-logind[1970]: Removed session 17. Apr 30 03:35:41.636431 systemd[1]: Started sshd@17-172.31.20.13:22-147.75.109.163:42068.service - OpenSSH per-connection server daemon (147.75.109.163:42068). Apr 30 03:35:41.891575 sshd[6191]: Accepted publickey for core from 147.75.109.163 port 42068 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:41.893149 sshd[6191]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:41.897909 systemd-logind[1970]: New session 18 of user core. Apr 30 03:35:41.902210 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 30 03:35:42.471953 sshd[6191]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:42.475931 systemd[1]: sshd@17-172.31.20.13:22-147.75.109.163:42068.service: Deactivated successfully. Apr 30 03:35:42.477734 systemd[1]: session-18.scope: Deactivated successfully. Apr 30 03:35:42.478883 systemd-logind[1970]: Session 18 logged out. Waiting for processes to exit. Apr 30 03:35:42.480231 systemd-logind[1970]: Removed session 18. Apr 30 03:35:42.525383 systemd[1]: Started sshd@18-172.31.20.13:22-147.75.109.163:42082.service - OpenSSH per-connection server daemon (147.75.109.163:42082). Apr 30 03:35:42.788460 sshd[6202]: Accepted publickey for core from 147.75.109.163 port 42082 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:42.790678 sshd[6202]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:42.796801 systemd-logind[1970]: New session 19 of user core. Apr 30 03:35:42.801211 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 30 03:35:43.071061 sshd[6202]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:43.075033 systemd[1]: sshd@18-172.31.20.13:22-147.75.109.163:42082.service: Deactivated successfully. Apr 30 03:35:43.076848 systemd[1]: session-19.scope: Deactivated successfully. Apr 30 03:35:43.077773 systemd-logind[1970]: Session 19 logged out. Waiting for processes to exit. Apr 30 03:35:43.079037 systemd-logind[1970]: Removed session 19. Apr 30 03:35:48.121397 systemd[1]: Started sshd@19-172.31.20.13:22-147.75.109.163:43196.service - OpenSSH per-connection server daemon (147.75.109.163:43196). Apr 30 03:35:48.377583 sshd[6258]: Accepted publickey for core from 147.75.109.163 port 43196 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:48.380180 sshd[6258]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:48.386375 systemd-logind[1970]: New session 20 of user core. Apr 30 03:35:48.391200 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 30 03:35:48.690778 sshd[6258]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:48.694446 systemd[1]: sshd@19-172.31.20.13:22-147.75.109.163:43196.service: Deactivated successfully. Apr 30 03:35:48.696335 systemd[1]: session-20.scope: Deactivated successfully. Apr 30 03:35:48.697688 systemd-logind[1970]: Session 20 logged out. Waiting for processes to exit. Apr 30 03:35:48.698792 systemd-logind[1970]: Removed session 20. Apr 30 03:35:53.737157 systemd[1]: Started sshd@20-172.31.20.13:22-147.75.109.163:43208.service - OpenSSH per-connection server daemon (147.75.109.163:43208). Apr 30 03:35:53.982967 sshd[6279]: Accepted publickey for core from 147.75.109.163 port 43208 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:53.984387 sshd[6279]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:53.989147 systemd-logind[1970]: New session 21 of user core. Apr 30 03:35:53.993186 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 30 03:35:54.285416 sshd[6279]: pam_unix(sshd:session): session closed for user core Apr 30 03:35:54.289205 systemd-logind[1970]: Session 21 logged out. Waiting for processes to exit. Apr 30 03:35:54.289809 systemd[1]: sshd@20-172.31.20.13:22-147.75.109.163:43208.service: Deactivated successfully. Apr 30 03:35:54.292639 systemd[1]: session-21.scope: Deactivated successfully. Apr 30 03:35:54.294986 systemd-logind[1970]: Removed session 21. Apr 30 03:35:59.336284 systemd[1]: Started sshd@21-172.31.20.13:22-147.75.109.163:34296.service - OpenSSH per-connection server daemon (147.75.109.163:34296). Apr 30 03:35:59.608033 sshd[6293]: Accepted publickey for core from 147.75.109.163 port 34296 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:35:59.610904 sshd[6293]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:35:59.616566 systemd-logind[1970]: New session 22 of user core. Apr 30 03:35:59.622361 systemd[1]: Started session-22.scope - Session 22 of User core. Apr 30 03:36:00.058132 sshd[6293]: pam_unix(sshd:session): session closed for user core Apr 30 03:36:00.084766 systemd[1]: sshd@21-172.31.20.13:22-147.75.109.163:34296.service: Deactivated successfully. Apr 30 03:36:00.089764 systemd[1]: session-22.scope: Deactivated successfully. Apr 30 03:36:00.091099 systemd-logind[1970]: Session 22 logged out. Waiting for processes to exit. Apr 30 03:36:00.092271 systemd-logind[1970]: Removed session 22. Apr 30 03:36:05.105231 systemd[1]: Started sshd@22-172.31.20.13:22-147.75.109.163:34300.service - OpenSSH per-connection server daemon (147.75.109.163:34300). Apr 30 03:36:05.393135 sshd[6306]: Accepted publickey for core from 147.75.109.163 port 34300 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:36:05.399105 sshd[6306]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:36:05.408991 systemd-logind[1970]: New session 23 of user core. Apr 30 03:36:05.413478 systemd[1]: Started session-23.scope - Session 23 of User core. Apr 30 03:36:06.173597 sshd[6306]: pam_unix(sshd:session): session closed for user core Apr 30 03:36:06.179011 systemd[1]: sshd@22-172.31.20.13:22-147.75.109.163:34300.service: Deactivated successfully. Apr 30 03:36:06.183783 systemd[1]: session-23.scope: Deactivated successfully. Apr 30 03:36:06.188057 systemd-logind[1970]: Session 23 logged out. Waiting for processes to exit. Apr 30 03:36:06.191754 systemd-logind[1970]: Removed session 23. Apr 30 03:36:11.225703 systemd[1]: Started sshd@23-172.31.20.13:22-147.75.109.163:46252.service - OpenSSH per-connection server daemon (147.75.109.163:46252). Apr 30 03:36:11.484758 sshd[6321]: Accepted publickey for core from 147.75.109.163 port 46252 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:36:11.488089 sshd[6321]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:36:11.493400 systemd-logind[1970]: New session 24 of user core. Apr 30 03:36:11.502275 systemd[1]: Started session-24.scope - Session 24 of User core. Apr 30 03:36:11.870487 sshd[6321]: pam_unix(sshd:session): session closed for user core Apr 30 03:36:11.875860 systemd[1]: sshd@23-172.31.20.13:22-147.75.109.163:46252.service: Deactivated successfully. Apr 30 03:36:11.878886 systemd[1]: session-24.scope: Deactivated successfully. Apr 30 03:36:11.880763 systemd-logind[1970]: Session 24 logged out. Waiting for processes to exit. Apr 30 03:36:11.883144 systemd-logind[1970]: Removed session 24. Apr 30 03:36:16.922627 systemd[1]: Started sshd@24-172.31.20.13:22-147.75.109.163:37636.service - OpenSSH per-connection server daemon (147.75.109.163:37636). Apr 30 03:36:17.201755 sshd[6355]: Accepted publickey for core from 147.75.109.163 port 37636 ssh2: RSA SHA256:7ZQea3lKZeIY1pq8546y2SpcWopo7i1peiZKBcYFJ3g Apr 30 03:36:17.204836 sshd[6355]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 03:36:17.209926 systemd-logind[1970]: New session 25 of user core. Apr 30 03:36:17.216368 systemd[1]: Started session-25.scope - Session 25 of User core. Apr 30 03:36:17.637553 sshd[6355]: pam_unix(sshd:session): session closed for user core Apr 30 03:36:17.640671 systemd[1]: sshd@24-172.31.20.13:22-147.75.109.163:37636.service: Deactivated successfully. Apr 30 03:36:17.642729 systemd[1]: session-25.scope: Deactivated successfully. Apr 30 03:36:17.644348 systemd-logind[1970]: Session 25 logged out. Waiting for processes to exit. Apr 30 03:36:17.645692 systemd-logind[1970]: Removed session 25. Apr 30 03:36:31.619873 systemd[1]: run-containerd-runc-k8s.io-b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2-runc.bHwjY4.mount: Deactivated successfully. Apr 30 03:36:31.907524 systemd[1]: cri-containerd-9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514.scope: Deactivated successfully. Apr 30 03:36:31.907902 systemd[1]: cri-containerd-9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514.scope: Consumed 2.823s CPU time, 20.7M memory peak, 0B memory swap peak. Apr 30 03:36:31.940231 systemd[1]: cri-containerd-a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d.scope: Deactivated successfully. Apr 30 03:36:31.940522 systemd[1]: cri-containerd-a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d.scope: Consumed 3.022s CPU time. Apr 30 03:36:32.085829 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d-rootfs.mount: Deactivated successfully. Apr 30 03:36:32.104695 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514-rootfs.mount: Deactivated successfully. Apr 30 03:36:32.118098 kubelet[3187]: E0430 03:36:32.106804 3187 controller.go:195] "Failed to update lease" err="Put \"https://172.31.20.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-13?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 30 03:36:32.154203 containerd[1992]: time="2025-04-30T03:36:32.124447808Z" level=info msg="shim disconnected" id=a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d namespace=k8s.io Apr 30 03:36:32.156106 containerd[1992]: time="2025-04-30T03:36:32.115914328Z" level=info msg="shim disconnected" id=9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514 namespace=k8s.io Apr 30 03:36:32.162460 containerd[1992]: time="2025-04-30T03:36:32.162324379Z" level=warning msg="cleaning up after shim disconnected" id=9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514 namespace=k8s.io Apr 30 03:36:32.162460 containerd[1992]: time="2025-04-30T03:36:32.162375739Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 03:36:32.164876 containerd[1992]: time="2025-04-30T03:36:32.164694172Z" level=warning msg="cleaning up after shim disconnected" id=a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d namespace=k8s.io Apr 30 03:36:32.164876 containerd[1992]: time="2025-04-30T03:36:32.164729608Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 03:36:33.089512 kubelet[3187]: I0430 03:36:33.089467 3187 scope.go:117] "RemoveContainer" containerID="a506c7b88b7613fb94eb0631a6e397e40fd573b52a94feb674287077fa47305d" Apr 30 03:36:33.091408 kubelet[3187]: I0430 03:36:33.091264 3187 scope.go:117] "RemoveContainer" containerID="9c739b4e5e4133025372bf41043b6d1890350f0f719dcab02df9514f1d81b514" Apr 30 03:36:33.116406 containerd[1992]: time="2025-04-30T03:36:33.116177000Z" level=info msg="CreateContainer within sandbox \"32d428b1833b4814a7e6947cd9800fcdd9ee500c575445f8f20fb867063e9f30\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Apr 30 03:36:33.117368 containerd[1992]: time="2025-04-30T03:36:33.117330605Z" level=info msg="CreateContainer within sandbox \"d1b18b6167552f4eb7dbe629997ba1d2d6a2845cc6896e2772332d0042ada763\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Apr 30 03:36:33.191920 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2665016611.mount: Deactivated successfully. Apr 30 03:36:33.196276 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2002935302.mount: Deactivated successfully. Apr 30 03:36:33.202596 containerd[1992]: time="2025-04-30T03:36:33.202547110Z" level=info msg="CreateContainer within sandbox \"d1b18b6167552f4eb7dbe629997ba1d2d6a2845cc6896e2772332d0042ada763\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"a07701c28b77fc1b75940bff46701ddad3ed530fb998ca27593d7cf7e47c7a69\"" Apr 30 03:36:33.203136 containerd[1992]: time="2025-04-30T03:36:33.203106987Z" level=info msg="StartContainer for \"a07701c28b77fc1b75940bff46701ddad3ed530fb998ca27593d7cf7e47c7a69\"" Apr 30 03:36:33.205425 containerd[1992]: time="2025-04-30T03:36:33.205293265Z" level=info msg="CreateContainer within sandbox \"32d428b1833b4814a7e6947cd9800fcdd9ee500c575445f8f20fb867063e9f30\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"44c6b0592a11ee80acd98b6ba290e9e9cc0e600c2366b2e405946168a6166aaa\"" Apr 30 03:36:33.206295 containerd[1992]: time="2025-04-30T03:36:33.205845135Z" level=info msg="StartContainer for \"44c6b0592a11ee80acd98b6ba290e9e9cc0e600c2366b2e405946168a6166aaa\"" Apr 30 03:36:33.250246 systemd[1]: Started cri-containerd-44c6b0592a11ee80acd98b6ba290e9e9cc0e600c2366b2e405946168a6166aaa.scope - libcontainer container 44c6b0592a11ee80acd98b6ba290e9e9cc0e600c2366b2e405946168a6166aaa. Apr 30 03:36:33.269270 systemd[1]: Started cri-containerd-a07701c28b77fc1b75940bff46701ddad3ed530fb998ca27593d7cf7e47c7a69.scope - libcontainer container a07701c28b77fc1b75940bff46701ddad3ed530fb998ca27593d7cf7e47c7a69. Apr 30 03:36:33.314798 containerd[1992]: time="2025-04-30T03:36:33.314675786Z" level=info msg="StartContainer for \"44c6b0592a11ee80acd98b6ba290e9e9cc0e600c2366b2e405946168a6166aaa\" returns successfully" Apr 30 03:36:33.334415 containerd[1992]: time="2025-04-30T03:36:33.333788595Z" level=info msg="StartContainer for \"a07701c28b77fc1b75940bff46701ddad3ed530fb998ca27593d7cf7e47c7a69\" returns successfully" Apr 30 03:36:37.857762 systemd[1]: cri-containerd-ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a.scope: Deactivated successfully. Apr 30 03:36:37.858859 systemd[1]: cri-containerd-ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a.scope: Consumed 2.519s CPU time, 20.5M memory peak, 0B memory swap peak. Apr 30 03:36:37.889679 containerd[1992]: time="2025-04-30T03:36:37.887369839Z" level=info msg="shim disconnected" id=ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a namespace=k8s.io Apr 30 03:36:37.889679 containerd[1992]: time="2025-04-30T03:36:37.887462130Z" level=warning msg="cleaning up after shim disconnected" id=ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a namespace=k8s.io Apr 30 03:36:37.889679 containerd[1992]: time="2025-04-30T03:36:37.887489784Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 03:36:37.897450 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a-rootfs.mount: Deactivated successfully. Apr 30 03:36:38.115212 kubelet[3187]: I0430 03:36:38.115111 3187 scope.go:117] "RemoveContainer" containerID="ce360925d59a3a06d12477b833f0539350be446e2f5460eb2e6790d43e30710a" Apr 30 03:36:38.117676 containerd[1992]: time="2025-04-30T03:36:38.117642774Z" level=info msg="CreateContainer within sandbox \"fd706056d2aea998d870b25264e075c95063aea9b4239b51e33bd4a0a773e582\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Apr 30 03:36:38.140084 containerd[1992]: time="2025-04-30T03:36:38.140040548Z" level=info msg="CreateContainer within sandbox \"fd706056d2aea998d870b25264e075c95063aea9b4239b51e33bd4a0a773e582\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"c5218729b7aa81b7cf5fa04f41443f85eabc7a8cfeab3acaf544f5e08d6ba852\"" Apr 30 03:36:38.140541 containerd[1992]: time="2025-04-30T03:36:38.140512868Z" level=info msg="StartContainer for \"c5218729b7aa81b7cf5fa04f41443f85eabc7a8cfeab3acaf544f5e08d6ba852\"" Apr 30 03:36:38.182219 systemd[1]: Started cri-containerd-c5218729b7aa81b7cf5fa04f41443f85eabc7a8cfeab3acaf544f5e08d6ba852.scope - libcontainer container c5218729b7aa81b7cf5fa04f41443f85eabc7a8cfeab3acaf544f5e08d6ba852. Apr 30 03:36:38.231177 containerd[1992]: time="2025-04-30T03:36:38.231045015Z" level=info msg="StartContainer for \"c5218729b7aa81b7cf5fa04f41443f85eabc7a8cfeab3acaf544f5e08d6ba852\" returns successfully" Apr 30 03:36:38.895155 systemd[1]: run-containerd-runc-k8s.io-c5218729b7aa81b7cf5fa04f41443f85eabc7a8cfeab3acaf544f5e08d6ba852-runc.PU9Xr6.mount: Deactivated successfully. Apr 30 03:36:42.118038 kubelet[3187]: E0430 03:36:42.117834 3187 controller.go:195] "Failed to update lease" err="Put \"https://172.31.20.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-13?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 30 03:36:47.700836 systemd[1]: run-containerd-runc-k8s.io-b90ffd3c3764684f5eb245749e77a613e9926e45c21d4453dc95cebe83eb7de2-runc.QKZycl.mount: Deactivated successfully. Apr 30 03:36:52.119326 kubelet[3187]: E0430 03:36:52.119195 3187 controller.go:195] "Failed to update lease" err="Put \"https://172.31.20.13:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ip-172-31-20-13?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)"